Daily arXiv Papers - 2025-10-21

AI-enhanced summaries of 0 research papers from arXiv

Today’s Research Highlights

AI-enhanced summaries of the latest research papers from arXiv.

Table of Contents

cs.CL

[1] Quantum NLP models on Natural Language Inference

Ling Sun, Peter Sullivan, Michael Martin, Yun Zhou

Main category: cs.CL

TL;DR: Quantum NLP models achieve comparable performance to classical transformers with far fewer parameters, showing higher per-parameter learning efficiency in few-shot Natural Language Inference tasks.

DetailsMotivation: To investigate the application of quantum natural language processing models for semantic modeling and Natural Language Inference under constrained few-shot settings, comparing quantum, hybrid, and classical approaches.

Method: Used lambeq library and DisCoCat framework to construct parameterized quantum circuits for sentence pairs, trained for semantic relatedness and inference classification. Introduced Information Gain per Parameter (IGPP) metric and proposed cluster-based architecture for parameter sharing.

Result: Quantum models achieved performance comparable to classical baselines with dramatically fewer parameters, outperformed randomly initialized transformers in inference, and showed up to five orders of magnitude higher per-parameter learning efficiency than classical counterparts.

Conclusion: QNLP shows promise for low-resource, structure-sensitive settings, with quantum models demonstrating superior parameter efficiency and learning dynamics.

Abstract: Quantum natural language processing (QNLP) offers a novel approach to semantic modeling by embedding compositional structure directly into quantum circuits. This paper investigates the application of QNLP models to the task of Natural Language Inference (NLI), comparing quantum, hybrid, and classical transformer-based models under a constrained few-shot setting. Using the lambeq library and the DisCoCat framework, we construct parameterized quantum circuits for sentence pairs and train them for both semantic relatedness and inference classification. To assess efficiency, we introduce a novel information-theoretic metric, Information Gain per Parameter (IGPP), which quantifies learning dynamics independent of model size. Our results demonstrate that quantum models achieve performance comparable to classical baselines while operating with dramatically fewer parameters. The Quantum-based models outperform randomly initialized transformers in inference and achieve lower test error on relatedness tasks. Moreover, quantum models exhibit significantly higher per-parameter learning efficiency (up to five orders of magnitude more than classical counterparts), highlighting the promise of QNLP in low-resource, structure-sensitive settings. To address circuit-level isolation and promote parameter sharing, we also propose a novel cluster-based architecture that improves generalization by tying gate parameters to learned word clusters rather than individual tokens.

[2] Fusion-Augmented Large Language Models: Boosting Diagnostic Trustworthiness via Model Consensus

Md Kamrul Siam, Md Jobair Hossain Faruk, Jerry Q. Cheng, Huanying Gu

Main category: cs.CL

TL;DR: Multi-model fusion framework using ChatGPT and Claude improves chest X-ray interpretation accuracy from 62.8-76.9% (individual models) to 91.3% (consensus) on CheXpert dataset.

DetailsMotivation: To enhance reliability of chest X-ray interpretation and reduce diagnostic errors in AI-assisted radiological diagnosis using complementary models and modalities.

Method: Used two LLMs (ChatGPT and Claude) with image-only and multimodal (image + synthetic clinical notes) inputs, employing similarity-based consensus approach with 95% output similarity threshold.

Result: Consensus accuracy improved from 77.6% (image-only) to 91.3% (multimodal). Individual models: ChatGPT 62.8%→84%, Claude 76.9%→76% with multimodal inputs.

Conclusion: Multi-model fusion with output-level consensus significantly improves diagnostic accuracy and trustworthiness of AI-assisted radiology with minimal computational overhead.

Abstract: This study presents a novel multi-model fusion framework leveraging two state-of-the-art large language models (LLMs), ChatGPT and Claude, to enhance the reliability of chest X-ray interpretation on the CheXpert dataset. From the full CheXpert corpus of 224,316 chest radiographs, we randomly selected 234 radiologist-annotated studies to evaluate unimodal performance using image-only prompts. In this setting, ChatGPT and Claude achieved diagnostic accuracies of 62.8% and 76.9%, respectively. A similarity-based consensus approach, using a 95% output similarity threshold, improved accuracy to 77.6%. To assess the impact of multimodal inputs, we then generated synthetic clinical notes following the MIMIC-CXR template and evaluated a separate subset of 50 randomly selected cases paired with both images and synthetic text. On this multimodal cohort, performance improved to 84% for ChatGPT and 76% for Claude, while consensus accuracy reached 91.3%. Across both experimental conditions, agreement-based fusion consistently outperformed individual models. These findings highlight the utility of integrating complementary modalities and using output-level consensus to improve the trustworthiness and clinical utility of AI-assisted radiological diagnosis, offering a practical path to reduce diagnostic errors with minimal computational overhead.

[3] Can LLMs Correct Themselves? A Benchmark of Self-Correction in LLMs

Guiyao Tie, Zenghui Yuan, Zeli Zhao, Chaoran Hu, Tianhe Gu, Ruihang Zhang, Sizhe Zhang, Junran Wu, Xiaoyue Tu, Ming Jin, Qingsong Wen, Lixing Chen, Pan Zhou, Lichao Sun

Main category: cs.CL

TL;DR: CorrectBench benchmark evaluates self-correction methods for LLMs across reasoning tasks, finding they improve accuracy but reduce efficiency, with CoT baseline showing competitive performance.

DetailsMotivation: To comprehensively evaluate self-correction methods for LLMs and determine whether they can truly correct themselves, as current evaluation remains largely unexplored.

Method: Developed CorrectBench benchmark to test intrinsic, external, and fine-tuned self-correction approaches across commonsense reasoning, mathematical reasoning, and code generation tasks.

Result: Self-correction improves accuracy especially for complex reasoning; mixing strategies yields further improvements but reduces efficiency; reasoning LLMs have limited optimization under additional self-correction; CoT baseline shows competitive accuracy and efficiency.

Conclusion: Self-correction has potential to enhance LLM reasoning but efficiency remains a challenge, advocating for research to optimize balance between reasoning capabilities and operational efficiency.

Abstract: Self-correction of large language models (LLMs) emerges as a critical component for enhancing their reasoning performance. Although various self-correction methods have been proposed, a comprehensive evaluation of these methods remains largely unexplored, and the question of whether LLMs can truly correct themselves is a matter of significant interest and concern. In this study, we introduce CorrectBench, a benchmark developed to evaluate the effectiveness of self-correction strategies, including intrinsic, external, and fine-tuned approaches, across three tasks: commonsense reasoning, mathematical reasoning, and code generation. Our findings reveal that: 1) Self-correction methods can improve accuracy, especially for complex reasoning tasks; 2) Mixing different self-correction strategies yields further improvements, though it reduces efficiency; 3) Reasoning LLMs (e.g., DeepSeek-R1) have limited optimization under additional self-correction methods and have high time costs. Interestingly, a comparatively simple chain-of-thought (CoT) baseline demonstrates competitive accuracy and efficiency. These results underscore the potential of self-correction to enhance LLM’s reasoning performance while highlighting the ongoing challenge of improving their efficiency. Consequently, we advocate for further research focused on optimizing the balance between reasoning capabilities and operational efficiency. Project Page: https://correctbench.github.io/

[4] EvolveR: Self-Evolving LLM Agents through an Experience-Driven Lifecycle

Rong Wu, Xiaoman Wang, Jianbiao Mei, Pinlong Cai, Daocheng Fu, Cheng Yang, Licheng Wen, Xuemeng Yang, Yufan Shen, Yuxin Wang, Botian Shi

Main category: cs.CL

TL;DR: EvolveR is a framework that enables LLM agents to self-improve through a closed-loop experience lifecycle with offline self-distillation and online interaction stages, achieving superior performance on complex multi-hop question-answering benchmarks.

DetailsMotivation: Current LLM agents lack systematic learning from their own experiences and cannot iteratively refine problem-solving strategies, focusing only on mitigating external knowledge gaps rather than fundamental improvement capabilities.

Method: Two-stage closed-loop lifecycle: (1) Offline Self-Distillation synthesizes interaction trajectories into abstract, reusable strategic principles; (2) Online Interaction retrieves distilled principles to guide decision-making while accumulating behavioral trajectories, with policy reinforcement for iterative updates.

Result: EvolveR achieves superior performance over strong agentic baselines on complex multi-hop question-answering benchmarks, demonstrating effective self-improvement capabilities.

Conclusion: The framework provides a comprehensive blueprint for agents that learn from both external data and the consequences of their own actions, enabling more autonomous and continuously improving systems.

Abstract: Current Large Language Model (LLM) agents show strong performance in tool use, but lack the crucial capability to systematically learn from their own experiences. While existing frameworks mainly focus on mitigating external knowledge gaps, they fail to address a more fundamental limitation: the inability to iteratively refine problem-solving strategies. In this work, we introduce EvolveR, a framework designed to enable agent to self-improve through a complete, closed-loop experience lifecycle. This lifecycle comprises two key stages: (1) Offline Self-Distillation, where the agent’s interaction trajectories are synthesized into a structured repository of abstract, reusable strategic principles; (2) Online Interaction, where the agent interacts with tasks and actively retrieves distilled principles to guide its decision-making, accumulating a diverse set of behavioral trajectories. This loop employs a policy reinforcement mechanism to iteratively update the agent based on its performance. We demonstrate the effectiveness of EvolveR on complex multi-hop question-answering benchmarks, where it achieves superior performance over strong agentic baselines. Our work presents a comprehensive blueprint for agents that learn not only from external data but also from the consequences of their own actions, paving the way for more autonomous and continuously improving systems. Code is available at https://github.com/Edaizi/EvolveR.

[5] BenCao: An Instruction-Tuned Large Language Model for Traditional Chinese Medicine

Jiacheng Xie, Yang Yu, Yibo Chen, Hanyao Zhang, Lening Zhao, Jiaxuan He, Lei Jiang, Xiaoting Tang, Guanghui An, Dong Xu

Main category: cs.CL

TL;DR: BenCao is a ChatGPT-based multimodal assistant for Traditional Chinese Medicine that integrates structured knowledge, diagnostic data, and expert feedback to address limitations in existing TCM-domain LLMs.

DetailsMotivation: Existing TCM-domain LLMs lack multimodal integration, interpretability, and clinical applicability despite TCM's reliance on holistic reasoning, implicit logic, and multimodal diagnostic cues.

Method: Developed through natural language instruction tuning (not parameter retraining), integrating comprehensive knowledge base, scenario-based instruction framework, chain-of-thought simulation, expert feedback refinement, and external APIs for tongue-image classification and multimodal database retrieval.

Result: Achieved superior accuracy to general-domain and TCM-domain models in diagnostics, herb recognition, and constitution classification. Deployed as interactive application on OpenAI GPTs Store with nearly 1,000 global users by October 2025.

Conclusion: Demonstrates feasibility of developing TCM-domain LLM through natural language-based instruction tuning and multimodal integration, providing practical framework for aligning generative AI with traditional medical reasoning and scalable real-world deployment.

Abstract: Traditional Chinese Medicine (TCM), with a history spanning over two millennia, plays a role in global healthcare. However, applying large language models (LLMs) to TCM remains challenging due to its reliance on holistic reasoning, implicit logic, and multimodal diagnostic cues. Existing TCM-domain LLMs have made progress in text-based understanding but lack multimodal integration, interpretability, and clinical applicability. To address these limitations, we developed BenCao, a ChatGPT-based multimodal assistant for TCM, integrating structured knowledge bases, diagnostic data, and expert feedback refinement. BenCao was trained through natural language instruction tuning rather than parameter retraining, aligning with expert-level reasoning and ethical norms specific to TCM. The system incorporates a comprehensive knowledge base of over 1,000 classical and modern texts, a scenario-based instruction framework for diverse interactions, a chain-of-thought simulation mechanism for interpretable reasoning, and a feedback refinement process involving licensed TCM practitioners. BenCao connects to external APIs for tongue-image classification and multimodal database retrieval, enabling dynamic access to diagnostic resources. In evaluations across single-choice question benchmarks and multimodal classification tasks, BenCao achieved superior accuracy to general-domain and TCM-domain models, particularly in diagnostics, herb recognition, and constitution classification. The model was deployed as an interactive application on the OpenAI GPTs Store, accessed by nearly 1,000 users globally as of October 2025. This study demonstrates the feasibility of developing a TCM-domain LLM through natural language-based instruction tuning and multimodal integration, offering a practical framework for aligning generative AI with traditional medical reasoning and a scalable pathway for real-world deployment.

[6] Evaluating Prompting Strategies and Large Language Models in Systematic Literature Review Screening: Relevance and Task-Stage Classification

Binglan Han, Anuradha Mathrani, Teo Susnjak

Main category: cs.CL

TL;DR: This study evaluates how different prompting strategies interact with LLMs to automate systematic literature review screening, finding that CoT-few-shot provides the best precision-recall balance and recommending a cost-effective staged workflow.

DetailsMotivation: To quantify how prompting strategies interact with large language models to automate the screening stage of systematic literature reviews, addressing the need for efficient and cost-effective literature screening automation.

Method: Evaluated six LLMs (GPT-4o, GPT-4o-mini, DeepSeek-Chat-V3, Gemini-2.5-Flash, Claude-3.5-Haiku, Llama-4-Maverick) under five prompt types (zero-shot, few-shot, chain-of-thought, CoT-few-shot, self-reflection) across relevance classification and six Level-2 tasks using accuracy, precision, recall, and F1 metrics.

Result: CoT-few-shot yielded the most reliable precision-recall balance; zero-shot maximized recall for high-sensitivity passes; self-reflection underperformed due to over-inclusivity. GPT-4o and DeepSeek provided robust overall performance, while GPT-4o-mini performed competitively at substantially lower cost. Cost-performance analysis revealed large differences among model-prompt pairings.

Conclusion: Recommends a staged workflow using low-cost models with structured prompts for first-pass screening and escalating borderline cases to higher-capacity models. Highlights LLMs’ uneven but promising potential for literature screening automation and provides practical guidance for task-adaptive deployment.

Abstract: This study quantifies how prompting strategies interact with large language models (LLMs) to automate the screening stage of systematic literature reviews (SLRs). We evaluate six LLMs (GPT-4o, GPT-4o-mini, DeepSeek-Chat-V3, Gemini-2.5-Flash, Claude-3.5-Haiku, Llama-4-Maverick) under five prompt types (zero-shot, few-shot, chain-of-thought (CoT), CoT-few-shot, self-reflection) across relevance classification and six Level-2 tasks, using accuracy, precision, recall, and F1. Results show pronounced model-prompt interaction effects: CoT-few-shot yields the most reliable precision-recall balance; zero-shot maximizes recall for high-sensitivity passes; and self-reflection underperforms due to over-inclusivity and instability across models. GPT-4o and DeepSeek provide robust overall performance, while GPT-4o-mini performs competitively at a substantially lower dollar cost. A cost-performance analysis for relevance classification (per 1,000 abstracts) reveals large absolute differences among model-prompt pairings; GPT-4o-mini remains low-cost across prompts, and structured prompts (CoT/CoT-few-shot) on GPT-4o-mini offer attractive F1 at a small incremental cost. We recommend a staged workflow that (1) deploys low-cost models with structured prompts for first-pass screening and (2) escalates only borderline cases to higher-capacity models. These findings highlight LLMs’ uneven but promising potential to automate literature screening. By systematically analyzing prompt-model interactions, we provide a comparative benchmark and practical guidance for task-adaptive LLM deployment.

[7] Facts in Stats: Impacts of Pretraining Diversity on Language Model Generalization

Tina Behnia, Puneesh Deora, Christos Thrampoulidis

Main category: cs.CL

TL;DR: A synthetic testbed is introduced to study how contextual diversity and structure affect language models’ factual recall and generalization, revealing that optimal diversity depends on training duration and that failures stem from embedding/unembedding bottlenecks.

DetailsMotivation: To systematically analyze how the interaction between statistical regularities and factual associations in language models affects generalization, which lacks controlled investigation despite its critical importance.

Method: Develop a flexible synthetic testbed combining statistical token streams with abstract factual source-target pairs, enabling independent control over contextual structure and diversity levels through stream composition manipulation.

Result: Higher contextual diversity delays in-distribution factual accuracy but has varying effects on out-of-distribution generalization depending on contextual structure. Some cases show similar ID/OOD trends, while others require diversity for non-trivial recall. Optimal diversity levels depend on training duration, and failures can be traced to embedding/unembedding layer bottlenecks.

Conclusion: The interplay between contextual design and diversity level significantly impacts different generalization aspects, with the synthetic framework enabling isolation of effects that would be confounded in large-scale studies, providing a controlled testbed for future research.

Abstract: Language models are pretrained on sequences that blend statistical regularities (making text fluent) with factual associations between specific tokens (knowledge of facts). While recent work suggests that the variability of their interaction, such as paraphrases of factual associations, critically determines generalization ability, we lack a systematic analysis of these impacts. This paper introduces a flexible synthetic testbed that combines a statistical stream of generic tokens with an abstract factual stream of source-target token pairs, enabling fine-grained control over their interaction. The design enables the independent control of diversity nature by manipulating stream composition (contextual structure) and the diversity level by varying which statistical streams each fact appears in. Through controlled experiments, we find that while higher contextual diversity delays in-distribution (ID) factual accuracy, its impact on out-of-distribution (OOD) factual generalization depends critically on contextual structure. In some cases, OOD performance follows the same trend as ID, but in others, diversity becomes essential for non-trivial factual recall. Even when low diversity prohibits factual recall, optimal diversity levels depend on training duration. Beyond factual recall failures, we identify structures where statistical generalization fails independently, and others where both capabilities degrade. This shows how the interplay between contextual design and diversity level impacts different generalization aspects. Further, through a series of controlled interventions on the model components, we trace the OOD failures to distinct optimization bottlenecks, highlighting the importance of the embedding and unembedding layers. Our synthetic framework allows us to isolate effects that would be confounded in large-scale studies, offering a controlled testbed for future investigations.

[8] In Generative AI We (Dis)Trust? Computational Analysis of Trust and Distrust in Reddit Discussions

Aria Pessianzadeh, Naima Sultana, Hildegarde Van den Bulck, David Gefen, Shahin Jabari, Rezvaneh Rezapour

Main category: cs.CL

TL;DR: First computational study of trust and distrust in GenAI using Reddit data from 2022-2025, finding balanced attitudes with shifts around model releases, dominated by technical performance and usability concerns.

DetailsMotivation: Understanding public trust in GenAI is essential for responsible adoption and governance, but prior work lacks computational, large-scale, and longitudinal approaches to measuring trust in GenAI and LLMs.

Method: Used multi-year Reddit dataset (2022-2025) spanning 39 subreddits and 197,618 posts, combining crowd-sourced annotations with classification models to scale analysis of trust and distrust patterns.

Result: Trust and distrust are nearly balanced over time with shifts around major model releases; technical performance and usability dominate as dimensions; personal experience is most frequent reason shaping attitudes; distinct patterns emerge across different user types.

Conclusion: Provides a methodological framework for large-scale trust analysis and insights into evolving public perceptions of GenAI, showing dynamic attitudes influenced by technical factors and personal experiences.

Abstract: The rise of generative AI (GenAI) has impacted many aspects of human life. As these systems become embedded in everyday practices, understanding public trust in them also becomes essential for responsible adoption and governance. Prior work on trust in AI has largely drawn from psychology and human-computer interaction, but there is a lack of computational, large-scale, and longitudinal approaches to measuring trust and distrust in GenAI and large language models (LLMs). This paper presents the first computational study of Trust and Distrust in GenAI, using a multi-year Reddit dataset (2022–2025) spanning 39 subreddits and 197,618 posts. Crowd-sourced annotations of a representative sample were combined with classification models to scale analysis. We find that Trust and Distrust are nearly balanced over time, with shifts around major model releases. Technical performance and usability dominate as dimensions, while personal experience is the most frequent reason shaping attitudes. Distinct patterns also emerge across trustors (e.g., experts, ethicists, general users). Our results provide a methodological framework for large-scale Trust analysis and insights into evolving public perceptions of GenAI.

[9] Probing the Hidden Talent of ASR Foundation Models for L2 English Oral Assessment

Fu-An Chao, Bi-Cheng Yan, Berlin Chen

Main category: cs.CL

TL;DR: This paper explores using Whisper ASR model for L2 spoken language assessment by extracting acoustic and linguistic features from hidden representations, achieving state-of-the-art performance without task-specific fine-tuning.

DetailsMotivation: To investigate the untapped potential of Whisper ASR foundation model for L2 spoken language assessment, going beyond just using its transcriptions to probe its latent capabilities in encoding proficiency patterns.

Method: Extract acoustic and linguistic features from Whisper’s hidden representations and train only a lightweight classifier on intermediate and final outputs, incorporating image and text-prompt information as auxiliary cues.

Result: Achieves strong performance on GEPT picture-description dataset, outperforming existing cutting-edge baselines including multimodal approaches, with additional gains from auxiliary relevance cues.

Conclusion: Whisper intrinsically encodes both ordinal proficiency patterns and semantic aspects of speech without task-specific fine-tuning, highlighting its potential as a powerful foundation for SLA and spoken language understanding tasks.

Abstract: In this paper, we explore the untapped potential of Whisper, a well-established automatic speech recognition (ASR) foundation model, in the context of L2 spoken language assessment (SLA). Unlike prior studies that extrinsically analyze transcriptions produced by Whisper, our approach goes a step further to probe its latent capabilities by extracting acoustic and linguistic features from hidden representations. With only a lightweight classifier being trained on top of Whisper’s intermediate and final outputs, our method achieves strong performance on the GEPT picture-description dataset, outperforming existing cutting-edge baselines, including a multimodal approach. Furthermore, by incorporating image and text-prompt information as auxiliary relevance cues, we demonstrate additional performance gains. Finally, we conduct an in-depth analysis of Whisper’s embeddings, which reveals that, even without task-specific fine-tuning, the model intrinsically encodes both ordinal proficiency patterns and semantic aspects of speech, highlighting its potential as a powerful foundation for SLA and other spoken language understanding tasks.

[10] EgMM-Corpus: A Multimodal Vision-Language Dataset for Egyptian Culture

Mohamed Gamil, Abdelrahman Elsayed, Abdelrahman Lila, Ahmed Gad, Hesham Abdelgawad, Mohamed Aref, Ahmed Fares

Main category: cs.CL

TL;DR: EgMM-Corpus is a multimodal dataset for Egyptian culture with 3,000+ images across 313 concepts, addressing cultural bias in AI models.

DetailsMotivation: Limited multimodal culturally diverse datasets for Middle East and Africa regions, particularly for Egyptian culture.

Method: Designed new data collection pipeline to collect images covering landmarks, food, and folklore, with manual validation for cultural authenticity and multimodal coherence.

Result: CLIP achieves 21.2% Top-1 and 36.4% Top-5 accuracy on EgMM-Corpus, revealing cultural bias in vision-language models.

Conclusion: EgMM-Corpus serves as a valuable benchmark for developing culturally aware AI models and highlights existing cultural biases in current systems.

Abstract: Despite recent advances in AI, multimodal culturally diverse datasets are still limited, particularly for regions in the Middle East and Africa. In this paper, we introduce EgMM-Corpus, a multimodal dataset dedicated to Egyptian culture. By designing and running a new data collection pipeline, we collected over 3,000 images, covering 313 concepts across landmarks, food, and folklore. Each entry in the dataset is manually validated for cultural authenticity and multimodal coherence. EgMM-Corpus aims to provide a reliable resource for evaluating and training vision-language models in an Egyptian cultural context. We further evaluate the zero-shot performance of Contrastive Language-Image Pre-training CLIP on EgMM-Corpus, on which it achieves 21.2% Top-1 accuracy and 36.4% Top-5 accuracy in classification. These results underscore the existing cultural bias in large-scale vision-language models and demonstrate the importance of EgMM-Corpus as a benchmark for developing culturally aware models.

[11] Hallucination Benchmark for Speech Foundation Models

Alkis Koudounas, Moreno La Quatra, Manuel Giollo, Sabato Marco Siniscalchi, Elena Baralis

Main category: cs.CL

TL;DR: SHALLOW is the first benchmark framework that systematically categorizes and quantifies hallucination phenomena in ASR along four axes: lexical, phonetic, morphological, and semantic, providing targeted metrics to assess model behavior beyond conventional error rates.

DetailsMotivation: ASR hallucinations - fluent but completely unrelated transcriptions - are more detrimental than conventional errors due to their preservation of syntactically and semantically plausible structure, which can mislead downstream applications and pose serious risks in critical domains. Conventional metrics fail to distinguish between phonetic inaccuracies and hallucinations.

Method: Introduced SHALLOW benchmark framework with four complementary axes (lexical, phonetic, morphological, semantic) and defined targeted metrics within each category to produce interpretable profiles of model behavior.

Result: SHALLOW metrics correlate strongly with WER when recognition quality is high (low WER), but this correlation weakens substantially as WER increases. SHALLOW captures fine-grained error patterns that WER fails to distinguish under degraded and challenging conditions.

Conclusion: SHALLOW supports specific diagnosis of model weaknesses and provides feedback for model improvement beyond what aggregate error rates can offer, addressing the critical need for evaluation frameworks that can effectively identify and assess models with heightened propensity for generating hallucinated content.

Abstract: Hallucinations in automatic speech recognition (ASR) systems refer to fluent and coherent transcriptions produced by neural ASR models that are completely unrelated to the underlying acoustic input (i.e., the speech signal). While similar to conventional decoding errors in potentially compromising the usability of transcriptions for downstream applications, hallucinations can be more detrimental due to their preservation of syntactically and semantically plausible structure. This apparent coherence can mislead subsequent processing stages and introduce serious risks, particularly in critical domains such as healthcare and law. Conventional evaluation metrics are primarily centered on error-based metrics and fail to distinguish between phonetic inaccuracies and hallucinations. Consequently, there is a critical need for new evaluation frameworks that can effectively identify and assess models with a heightened propensity for generating hallucinated content. To this end, we introduce SHALLOW, the first benchmark framework that systematically categorizes and quantifies hallucination phenomena in ASR along four complementary axes: lexical, phonetic, morphological, and semantic. We define targeted metrics within each category to produce interpretable profiles of model behavior. Through evaluation across various architectures and speech domains, we have found that SHALLOW metrics correlate strongly with word error rate (WER) when recognition quality is high (i.e., low WER). Still, this correlation weakens substantially as WER increases. SHALLOW, therefore, captures fine-grained error patterns that WER fails to distinguish under degraded and challenging conditions. Our framework supports specific diagnosis of model weaknesses and provides feedback for model improvement beyond what aggregate error rates can offer.

[12] What Can String Probability Tell Us About Grammaticality?

Jennifer Hu, Ethan Gotlieb Wilcox, Siyuan Song, Kyle Mahowald, Roger P. Levy

Main category: cs.CL

TL;DR: The paper analyzes what language models learn about grammar by examining the relationship between string probability and grammatical knowledge, validating predictions through empirical tests with 280K sentence pairs in English and Chinese.

DetailsMotivation: To understand what language models have learned about grammar, given that probability and grammaticality are distinct concepts in linguistics, and to establish theoretical grounding for using probability to assess LMs' structural knowledge.

Method: Theoretical analysis of the relationship between grammar, meaning, and string probability based on assumptions about corpus data generation, validated empirically using 280K sentence pairs in English and Chinese to test three predictions about minimal pairs and probability correlations.

Result: Empirical validation of three predictions: (1) correlation between probabilities of strings within minimal pairs, (2) correlation between models’ and humans’ deltas within minimal pairs, and (3) poor separation in probability space between unpaired grammatical and ungrammatical strings.

Conclusion: The analyses provide theoretical grounding for using probability to learn about LMs’ structural knowledge and suggest directions for future work in LM grammatical evaluation.

Abstract: What have language models (LMs) learned about grammar? This question remains hotly debated, with major ramifications for linguistic theory. However, since probability and grammaticality are distinct notions in linguistics, it is not obvious what string probabilities can reveal about an LM’s underlying grammatical knowledge. We present a theoretical analysis of the relationship between grammar, meaning, and string probability, based on simple assumptions about the generative process of corpus data. Our framework makes three predictions, which we validate empirically using 280K sentence pairs in English and Chinese: (1) correlation between the probability of strings within minimal pairs, i.e., string pairs with minimal semantic differences; (2) correlation between models’ and humans’ deltas within minimal pairs; and (3) poor separation in probability space between unpaired grammatical and ungrammatical strings. Our analyses give theoretical grounding for using probability to learn about LMs’ structural knowledge, and suggest directions for future work in LM grammatical evaluation.

[13] Towards Low-Resource Alignment to Diverse Perspectives with Sparse Feedback

Chu Fei Luo, Samuel Dahan, Xiaodan Zhu

Main category: cs.CL

TL;DR: This paper proposes two methods (pluralistic decoding and model steering) to enhance pluralistic alignment of language models in low-resource settings, improving alignment with diverse human perspectives using only 50 annotated samples.

DetailsMotivation: Current language model training assumes one optimal answer per query, leading to generic responses and poor alignment with diverse human values and perspectives. There's a need to ensure models reflect nuance and diversity in human values.

Method: Two proposed methods: pluralistic decoding and model steering, designed to work in low-resource settings with only 50 annotated samples.

Result: Model steering consistently improves over zero-shot and few-shot baselines, reduces false positives in high-stakes tasks (hate speech detection, misinformation detection), and improves distributional alignment to human values in GlobalOpinionQA.

Conclusion: The work highlights the importance of diversity and demonstrates how language models can be adapted to consider nuanced perspectives, with model steering showing particular promise for pluralistic alignment.

Abstract: As language models have a greater impact on society, it is important to ensure they are aligned to a diverse range of perspectives and are able to reflect nuance in human values. However, the most popular training paradigms for modern language models often assume there is one optimal answer for every query, leading to generic responses and poor alignment. In this work, we aim to enhance pluralistic alignment of language models in a low-resource setting with two methods: pluralistic decoding and model steering. We empirically demonstrate that model steering offers consistent improvement over zero-shot and few-shot baselines with only 50 annotated samples. Our proposed methods decrease false positives in several high-stakes tasks such as hate speech detection and misinformation detection, and improves the distributional alignment to human values in GlobalOpinionQA. We hope our work highlights the importance of diversity and how language models can be adapted to consider nuanced perspectives.

[14] Instant Personalized Large Language Model Adaptation via Hypernetwork

Zhaoxuan Tan, Zixuan Zhang, Haoyang Wen, Zheng Li, Rongzhi Zhang, Pei Chen, Fengran Mo, Zheyuan Liu, Qingkai Zeng, Qingyu Yin, Meng Jiang

Main category: cs.CL

TL;DR: Profile-to-PEFT is a scalable framework that uses a hypernetwork to map user profiles directly to adapter parameters, eliminating per-user training and enabling instant personalization of LLMs.

DetailsMotivation: Existing PEFT methods require training separate adapters for each user (One-PEFT-Per-User), which is computationally expensive and impractical for real-time updates.

Method: A hypernetwork trained end-to-end maps encoded user profiles directly to full sets of adapter parameters (e.g., LoRA), eliminating per-user training at deployment.

Result: Outperforms both prompt-based personalization and OPPU while using substantially fewer computational resources, with strong generalization to unseen users and robustness across varying conditions.

Conclusion: Profile-to-PEFT enables efficient, scalable, and adaptive LLM personalization suitable for large-scale applications with instant adaptation and privacy-preserving local deployment.

Abstract: Personalized large language models (LLMs) tailor content to individual preferences using user profiles or histories. However, existing parameter-efficient fine-tuning (PEFT) methods, such as the ``One-PEFT-Per-User’’ (OPPU) paradigm, require training a separate adapter for each user, making them computationally expensive and impractical for real-time updates. We introduce Profile-to-PEFT, a scalable framework that employs a hypernetwork, trained end-to-end, to map a user’s encoded profile directly to a full set of adapter parameters (e.g., LoRA), eliminating per-user training at deployment. This design enables instant adaptation, generalization to unseen users, and privacy-preserving local deployment. Experimental results demonstrate that our method outperforms both prompt-based personalization and OPPU while using substantially fewer computational resources at deployment. The framework exhibits strong generalization to out-of-distribution users and maintains robustness across varying user activity levels and different embedding backbones. The proposed Profile-to-PEFT framework enables efficient, scalable, and adaptive LLM personalization suitable for large-scale applications.

[15] Thinking About Thinking: Evaluating Reasoning in Post-Trained Language Models

Pratham Singla, Shivank Garg, Ayush Singh, Ishan Garg, Ketan Suhaas Saichandran

Main category: cs.CL

TL;DR: RL-trained LLMs show better awareness of learned policies and generalization than SFT models, but weaker alignment between reasoning traces and final outputs, especially in GRPO-trained models.

DetailsMotivation: To investigate whether LLMs are aware of what they learn and think after post-training techniques that enhance their capabilities for logic-intensive tasks.

Method: Defined three core competencies (awareness, generalization, alignment) and empirically evaluated them on tasks requiring distinct policies, comparing SFT, DPO, and GRPO training methods.

Result: RL-trained models demonstrated greater awareness of learned behaviors and stronger generalization to novel tasks than SFT models, but showed weak alignment between reasoning traces and final outputs, particularly in GRPO-trained models.

Conclusion: RL training enhances policy awareness and generalization but compromises alignment between internal reasoning and final outputs, with GRPO showing the most pronounced misalignment.

Abstract: Recent advances in post-training techniques have endowed Large Language Models (LLMs) with enhanced capabilities for tackling complex, logic-intensive tasks through the generation of supplementary planning tokens. This development raises a fundamental question: Are these models aware of what they “learn” and “think”? To address this, we define three core competencies: (1) awareness of learned latent policies, (2) generalization of these policies across domains, and (3) alignment between internal reasoning traces and final outputs. We empirically evaluate these abilities on several tasks, each designed to require learning a distinct policy. Furthermore, we contrast the profiles of models post-trained via Supervised Fine-Tuning (SFT), Direct Policy Optimization (DPO), and Group Relative Policy Optimization (GRPO). Our findings indicate that RL-trained models not only demonstrate greater awareness of their learned behaviors and stronger generalizability to novel, structurally similar tasks than SFT models but also often exhibit weak alignment between their reasoning traces and final outputs, an effect most pronounced in GRPO-trained models.

[16] Utilising Large Language Models for Generating Effective Counter Arguments to Anti-Vaccine Tweets

Utsav Dhanuka, Soham Poddar, Saptarshi Ghosh

Main category: cs.CL

TL;DR: LLMs can generate effective counter-arguments against vaccine misinformation using optimized prompting and fine-tuning approaches, with classification of anti-vaccine tweets enabling context-aware rebuttals.

DetailsMotivation: Combat vaccine skepticism and misinformation on social media, which creates barriers to high immunization rates and undermines trust in health recommendations.

Method: Experiment with various prompting strategies and fine-tuning approaches for LLMs, plus train classifiers to categorize anti-vaccine tweets into multi-labeled categories (efficacy, side effects, political influences) for context-aware rebuttals.

Result: Strong alignment across human judgment, LLM-based assessments, and automatic metrics. Integration of label descriptions and structured fine-tuning enhances counter-argument effectiveness.

Conclusion: LLMs offer a promising approach for mitigating vaccine misinformation at scale through optimized counter-argument generation.

Abstract: In an era where public health is increasingly influenced by information shared on social media, combatting vaccine skepticism and misinformation has become a critical societal goal. Misleading narratives around vaccination have spread widely, creating barriers to achieving high immunisation rates and undermining trust in health recommendations. While efforts to detect misinformation have made significant progress, the generation of real time counter-arguments tailored to debunk such claims remains an insufficiently explored area. In this work, we explore the capabilities of LLMs to generate sound counter-argument rebuttals to vaccine misinformation. Building on prior research in misinformation debunking, we experiment with various prompting strategies and fine-tuning approaches to optimise counter-argument generation. Additionally, we train classifiers to categorise anti-vaccine tweets into multi-labeled categories such as concerns about vaccine efficacy, side effects, and political influences allowing for more context aware rebuttals. Our evaluation, conducted through human judgment, LLM based assessments, and automatic metrics, reveals strong alignment across these methods. Our findings demonstrate that integrating label descriptions and structured fine-tuning enhances counter-argument effectiveness, offering a promising approach for mitigating vaccine misinformation at scale.

[17] Unleashing Diverse Thinking Modes in LLMs through Multi-Agent Collaboration

Zhixuan He, Yue Feng

Main category: cs.CL

TL;DR: DiMo is a multi-agent collaboration framework that uses four specialized LLM agents with distinct reasoning paradigms to enhance performance and interpretability through structured debate.

DetailsMotivation: LLMs demonstrate strong performance but often lack interpretable reasoning, creating a need for frameworks that provide both improved accuracy and transparent reasoning processes.

Method: The framework simulates structured debate among four specialized LLM agents, each embodying a distinct reasoning paradigm, allowing collaborative exploration of diverse cognitive approaches through iterative debate.

Result: Across six benchmarks under unified open-source setup, DiMo improves accuracy over single-model and debate baselines, with largest gains on math tasks, while providing explicit, auditable reasoning chains.

Conclusion: DiMo positions itself as a semantics-aware, Web-native multi-agent framework that models human-machine intelligence with LLM agents producing semantically typed, URL-annotated evidence chains for explanations and user-friendly interactions.

Abstract: Large Language Models (LLMs) demonstrate strong performance but often lack interpretable reasoning. This paper introduces the Multi-Agent Collaboration Framework for Diverse Thinking Modes (DiMo), which enhances both performance and interpretability by simulating a structured debate among four specialized LLM agents. Each agent embodies a distinct reasoning paradigm, allowing the framework to collaboratively explore diverse cognitive approaches. Through iterative debate, agents challenge and refine initial responses, yielding more robust conclusions and an explicit, auditable reasoning chain. Across six benchmarks and under a unified open-source setup, DiMo improves accuracy over widely used single-model and debate baselines, with the largest gains on math. We position DiMo as a semantics-aware, Web-native multi-agent framework: it models human-machine intelligence with LLM agents that produce semantically typed, URL-annotated evidence chains for explanations and user-friendly interactions. Although our experiments use standard reasoning benchmarks, the framework is designed to be instantiated over Web corpora and knowledge graphs, combining retrieval-augmented reasoning with structured justifications that downstream systems can inspect and reuse.

[18] End-to-End Argument Mining through Autoregressive Argumentative Structure Prediction

Nilmadhab Das, Vishal Vaibhav, Yash Sunil Choudhary, V. Vijaya Saradhi, Ashish Anand

Main category: cs.CL

TL;DR: Proposes AASP framework for joint argument mining using autoregressive structure prediction with constrained actions, achieving state-of-the-art results.

DetailsMotivation: Existing approaches flatten argumentative structures, failing to capture dependencies between argument components and relations. Joint modeling is needed for better reasoning flow.

Method: Autoregressive Argumentative Structure Prediction (AASP) framework using pre-trained language models with constrained action sets to build structures step-by-step.

Result: Achieved state-of-the-art results on two benchmarks and strong results on third benchmark across all argument mining tasks.

Conclusion: AASP effectively captures argumentative reasoning flow through joint formulation and autoregressive structure prediction.

Abstract: Argument Mining (AM) helps in automating the extraction of complex argumentative structures such as Argument Components (ACs) like Premise, Claim etc. and Argumentative Relations (ARs) like Support, Attack etc. in an argumentative text. Due to the inherent complexity of reasoning involved with this task, modelling dependencies between ACs and ARs is challenging. Most of the recent approaches formulate this task through a generative paradigm by flattening the argumentative structures. In contrast to that, this study jointly formulates the key tasks of AM in an end-to-end fashion using Autoregressive Argumentative Structure Prediction (AASP) framework. The proposed AASP framework is based on the autoregressive structure prediction framework that has given good performance for several NLP tasks. AASP framework models the argumentative structures as constrained pre-defined sets of actions with the help of a conditional pre-trained language model. These actions build the argumentative structures step-by-step in an autoregressive manner to capture the flow of argumentative reasoning in an efficient way. Extensive experiments conducted on three standard AM benchmarks demonstrate that AASP achieves state-of-theart (SoTA) results across all AM tasks in two benchmarks and delivers strong results in one benchmark.

[19] Navigating through the hidden embedding space: steering LLMs to improve mental health assessment

Federico Ravenda, Seyed Ali Bahrainian, Andrea Raballo, Antonietta Mira

Main category: cs.CL

TL;DR: A lightweight method using linear transformation and steering vectors improves mental health assessment capabilities of LLMs without intensive computation, achieving better performance on depression detection tasks.

DetailsMotivation: Smaller LLMs struggle with domain-specific mental health applications despite advancements, creating need for cost-efficient adaptation methods.

Method: Linear transformation applied to specific layer activations using steering vectors to guide model output, without computationally intensive techniques.

Result: Improved performance on two tasks: identifying depression-relevant Reddit posts and completing depression screening questionnaires based on user post history.

Conclusion: Steering mechanisms show untapped potential as computationally efficient tools for LLM domain adaptation in mental health applications.

Abstract: The rapid evolution of Large Language Models (LLMs) is transforming AI, opening new opportunities in sensitive and high-impact areas such as Mental Health (MH). Yet, despite these advancements, recent evidence reveals that smaller-scale models still struggle to deliver optimal performance in domain-specific applications. In this study, we present a cost-efficient yet powerful approach to improve MH assessment capabilities of an LLM, without relying on any computationally intensive techniques. Our lightweight method consists of a linear transformation applied to a specific layer’s activations, leveraging steering vectors to guide the model’s output. Remarkably, this intervention enables the model to achieve improved results across two distinct tasks: (1) identifying whether a Reddit post is useful for detecting the presence or absence of depressive symptoms (relevance prediction task), and (2) completing a standardized psychological screening questionnaire for depression based on users’ Reddit post history (questionnaire completion task). Results highlight the untapped potential of steering mechanisms as computationally efficient tools for LLMs’ MH domain adaptation.

[20] Verification-Aware Planning for Multi-Agent Systems

Tianyang Xu, Dan Zhang, Kushan Mitra, Estevam Hruschka

Main category: cs.CL

TL;DR: VeriMAP is a verification-aware planning framework for multi-agent LLM collaboration that addresses coordination failures through automated verification functions and improves system robustness.

DetailsMotivation: Multi-agent LLM collaboration faces challenges in planning, coordination, and verification, with failures often arising from subtle misalignments in task interpretation, output format, or inter-agent handoffs rather than flawed reasoning alone.

Method: VeriMAP decomposes tasks, models subtask dependencies, and encodes planner-defined passing criteria as subtask verification functions (VFs) in both Python and natural language for automated verification.

Result: VeriMAP outperforms both single- and multi-agent baselines on diverse datasets while enhancing system robustness and interpretability.

Conclusion: Verification-aware planning enables reliable coordination and iterative refinement in multi-agent systems without relying on external labels or annotations.

Abstract: Large language model (LLM) agents are increasingly deployed to tackle complex tasks, often necessitating collaboration among multiple specialized agents. However, multi-agent collaboration introduces new challenges in planning, coordination, and verification. Execution failures frequently arise not from flawed reasoning alone, but from subtle misalignments in task interpretation, output format, or inter-agent handoffs. To address these challenges, we present VeriMAP, a framework for multi-agent collaboration with verification-aware planning. The VeriMAP planner decomposes tasks, models subtask dependencies, and encodes planner-defined passing criteria as subtask verification functions (VFs) in Python and natural language. We evaluate VeriMAP on diverse datasets, demonstrating that it outperforms both single- and multi-agent baselines while enhancing system robustness and interpretability. Our analysis highlights how verification-aware planning enables reliable coordination and iterative refinement in multi-agent systems, without relying on external labels or annotations.

[21] Late Fusion and Multi-Level Fission Amplify Cross-Modal Transfer in Text-Speech LMs

Santiago Cuervo, Adel Moumen, Yanis Labrak, Sameer Khurana, Antoine Laurent, Mickael Rouvier, Phil Woodland, Ricard Marxer

Main category: cs.CL

TL;DR: Text-Speech Language Models using late fusion/fission with access to both high- and low-level features outperform early fusion approaches and rival state-of-the-art models with much less compute.

DetailsMotivation: Early modality fusion/fission in Text-Speech Language Models limits cross-modal transfer by neglecting feature compositionality and the finer-grained nature of speech representations compared to text.

Method: Proposed late fusion and fission approach with fission process that accesses both high- and low-level features for speech generation, implemented in SmolTolk models.

Result: SmolTolk models rival or surpass state-of-the-art TSLMs trained with orders of magnitude more compute, achieving significantly improved cross-modal performance. Representation analyses show enhanced ability to abstract higher-level semantic features from speech and more shared representation spaces across layers.

Conclusion: Late fusion/fission with multi-level feature access addresses limitations of early fusion approaches in TSLMs, enabling better cross-modal transfer and more efficient training while achieving competitive performance.

Abstract: Text-Speech Language Models (TSLMs) – language models trained to jointly process and generate text and speech – are commonly trained through an early modality fusion/fission approach, in which both modalities are fed and predicted from a shared backbone via linear layers. We hypothesize that this approach limits cross-modal transfer by neglecting feature compositionality – specifically, the finer-grained nature of speech representations compared to text – preventing the emergence of a shared feature hierarchy within model layers. In this paper, we argue that this limitation can be addressed through late fusion and fission, with a fission process that accesses both high- and low-level features for speech generation. Our models implementing these principles, SmolTolk, rival or surpass state-of-the-art TSLMs trained with orders of magnitude more compute, and achieve significantly improved cross-modal performance relative to early fusion/fission baselines. Representation analyses further suggest that our method enhances the model’s ability to abstract higher-level, more semantic features from speech, and leads to increasingly shared representation spaces across layers.

[22] MoReBench: Evaluating Procedural and Pluralistic Moral Reasoning in Language Models, More than Outcomes

Yu Ying Chiu, Michael S. Lee, Rachel Calcott, Brandon Handoko, Paul de Font-Reaulx, Paula Rodriguez, Chen Bo Calvin Zhang, Ziwen Han, Udari Madhushani Sehwag, Yash Maurya, Christina Q Knight, Harry R. Lloyd, Florence Bacus, Mantas Mazeika, Bing Liu, Yejin Choi, Mitchell L Gordon, Sydney Levine

Main category: cs.CL

TL;DR: MoReBench is a benchmark for evaluating AI moral reasoning processes through 1,000 moral scenarios with expert-defined rubric criteria, showing that scaling laws and existing benchmarks don’t predict moral reasoning abilities.

DetailsMotivation: To understand how AI systems make decisions, particularly in moral dilemmas where multiple defensible conclusions exist, enabling process-focused evaluation of AI procedural reasoning.

Method: Created MoReBench with 1,000 moral scenarios paired with 23k+ expert-defined rubric criteria covering moral considerations, trade-offs, and recommendations. Also developed MoReBench-Theory with 150 examples testing reasoning under five major normative ethics frameworks.

Result: Scaling laws and existing benchmarks on math, code, and scientific reasoning fail to predict models’ moral reasoning abilities. Models show bias toward specific moral frameworks (Benthamite Act Utilitarianism and Kantian Deontology), possibly due to training paradigms.

Conclusion: These benchmarks advance process-focused reasoning evaluation toward safer and more transparent AI by enabling assessment of how AI systems reason about moral dilemmas.

Abstract: As AI systems progress, we rely more on them to make decisions with us and for us. To ensure that such decisions are aligned with human values, it is imperative for us to understand not only what decisions they make but also how they come to those decisions. Reasoning language models, which provide both final responses and (partially transparent) intermediate thinking traces, present a timely opportunity to study AI procedural reasoning. Unlike math and code problems which often have objectively correct answers, moral dilemmas are an excellent testbed for process-focused evaluation because they allow for multiple defensible conclusions. To do so, we present MoReBench: 1,000 moral scenarios, each paired with a set of rubric criteria that experts consider essential to include (or avoid) when reasoning about the scenarios. MoReBench contains over 23 thousand criteria including identifying moral considerations, weighing trade-offs, and giving actionable recommendations to cover cases on AI advising humans moral decisions as well as making moral decisions autonomously. Separately, we curate MoReBench-Theory: 150 examples to test whether AI can reason under five major frameworks in normative ethics. Our results show that scaling laws and existing benchmarks on math, code, and scientific reasoning tasks fail to predict models’ abilities to perform moral reasoning. Models also show partiality towards specific moral frameworks (e.g., Benthamite Act Utilitarianism and Kantian Deontology), which might be side effects of popular training paradigms. Together, these benchmarks advance process-focused reasoning evaluation towards safer and more transparent AI.

[23] ATA: A Neuro-Symbolic Approach to Implement Autonomous and Trustworthy Agents

David Peer, Sebastian Stabinger

Main category: cs.CL

TL;DR: ATA is a neuro-symbolic framework that decouples LLM tasks into offline knowledge formalization and online symbolic reasoning, achieving competitive performance while ensuring trustworthiness through verifiable symbolic knowledge bases.

DetailsMotivation: Address limitations in LLM trustworthiness (hallucinations, instability, lack of transparency) that hinder deployment in high-stakes domains.

Method: Two-phase approach: 1) Offline knowledge ingestion where LLM translates informal specs into formal symbolic knowledge base, 2) Online task processing where symbolic decision engine uses formal input encoding and knowledge base for reliable reasoning.

Result: Competitive with state-of-the-art end-to-end models in automated setup; with human-verified knowledge base, significantly outperforms larger models while achieving perfect determinism, enhanced stability, and immunity to prompt injection attacks.

Conclusion: ATA provides a practical and controllable architecture for building transparent, auditable, and reliable autonomous agents through symbolic reasoning.

Abstract: Large Language Models (LLMs) have demonstrated impressive capabilities, yet their deployment in high-stakes domains is hindered by inherent limitations in trustworthiness, including hallucinations, instability, and a lack of transparency. To address these challenges, we introduce a generic neuro-symbolic approach, which we call Autonomous Trustworthy Agents (ATA). The core of our approach lies in decoupling tasks into two distinct phases: Offline knowledge ingestion and online task processing. During knowledge ingestion, an LLM translates an informal problem specification into a formal, symbolic knowledge base. This formal representation is crucial as it can be verified and refined by human experts, ensuring its correctness and alignment with domain requirements. In the subsequent task processing phase, each incoming input is encoded into the same formal language. A symbolic decision engine then utilizes this encoded input in conjunction with the formal knowledge base to derive a reliable result. Through an extensive evaluation on a complex reasoning task, we demonstrate that a concrete implementation of ATA is competitive with state-of-the-art end-to-end reasoning models in a fully automated setup while maintaining trustworthiness. Crucially, with a human-verified and corrected knowledge base, our approach significantly outperforms even larger models, while exhibiting perfect determinism, enhanced stability against input perturbations, and inherent immunity to prompt injection attacks. By generating decisions grounded in symbolic reasoning, ATA offers a practical and controllable architecture for building the next generation of transparent, auditable, and reliable autonomous agents.

[24] SHANKS: Simultaneous Hearing and Thinking for Spoken Language Models

Cheng-Han Chiang, Xiaofei Wang, Linjie Li, Chung-Ching Lin, Kevin Lin, Shujie Liu, Zhendong Wang, Zhengyuan Yang, Hung-yi Lee, Lijuan Wang

Main category: cs.CL

TL;DR: SHANKS is a framework that enables spoken language models to generate unspoken reasoning while listening to user input, allowing real-time interaction and interruption during speech rather than waiting for turn completion.

DetailsMotivation: Current LLMs/SLMs only think after users finish speaking, causing high latency that's unsuitable for speech-to-speech interaction where real-time exchange is important. Humans naturally "think while listening," which inspired this approach.

Method: SHANKS streams input speech in fixed-duration chunks and generates unspoken chain-of-thought reasoning as each chunk arrives, using all previous speech and reasoning to decide whether to interrupt users or make tool calls while they continue speaking.

Result: SHANKS achieved 37.1% higher interruption accuracy than baseline when detecting math mistakes, and completed 56.9% of tool calls before users finished speaking in tool-augmented dialogues.

Conclusion: SHANKS enables models to think throughout conversations rather than only after turns end, moving toward more natural, real-time human-model interaction.

Abstract: Current large language models (LLMs) and spoken language models (SLMs) begin thinking and taking actions only after the user has finished their turn. This prevents the model from interacting during the user’s turn and can lead to high response latency while it waits to think. Consequently, thinking after receiving the full input is not suitable for speech-to-speech interaction, where real-time, low-latency exchange is important. We address this by noting that humans naturally “think while listening.” In this paper, we propose SHANKS, a general inference framework that enables SLMs to generate unspoken chain-of-thought reasoning while listening to the user input. SHANKS streams the input speech in fixed-duration chunks and, as soon as a chunk is received, generates unspoken reasoning based on all previous speech and reasoning, while the user continues speaking. SHANKS uses this unspoken reasoning to decide whether to interrupt the user and to make tool calls to complete the task. We demonstrate that SHANKS enhances real-time user-SLM interaction in two scenarios: (1) when the user is presenting a step-by-step solution to a math problem, SHANKS can listen, reason, and interrupt when the user makes a mistake, achieving 37.1% higher interruption accuracy than a baseline that interrupts without thinking; and (2) in a tool-augmented dialogue, SHANKS can complete 56.9% of the tool calls before the user finishes their turn. Overall, SHANKS moves toward models that keep thinking throughout the conversation, not only after a turn ends. Animated illustrations of Shanks can be found at https://d223302.github.io/SHANKS/

[25] FrugalPrompt: Reducing Contextual Overhead in Large Language Models via Token Attribution

Syed Rifat Raiyan, Md Farhan Ishmam, Abdullah Al Imran, Mohammad Ali Moni

Main category: cs.CL

TL;DR: FrugalPrompt is a novel prompt compression framework that retains only the most semantically significant tokens using token attribution methods, achieving 20% prompt reduction with minimal performance loss on most NLP tasks except mathematical reasoning.

DetailsMotivation: Large language models suffer from high costs and latency due to redundant low-utility tokens in prompts, with only a fraction of tokens carrying most semantic weight.

Method: Uses GlobEnc and DecompX token attribution methods to assign salience scores, rank tokens, and preserve top-k% tokens in original order to create sparse frugalized prompts.

Result: 20% prompt reduction causes only marginal performance loss on sentiment analysis, commonsense QA, and summarization, but sharp deterioration on mathematical reasoning. Bottom-k% and random-k% tokens show asymmetric patterns suggesting potential task contamination.

Conclusion: The work provides nuanced understanding of LLM behavior in performance-efficiency trade-offs, delineating boundaries between tasks tolerant to contextual sparsity versus those requiring exhaustive context.

Abstract: Large language models (LLMs) owe much of their stellar performance to expansive input contexts, yet such verbosity inflates monetary costs, carbon footprint, and inference-time latency. Much of this overhead manifests from the redundant low-utility tokens present in typical prompts, as only a fraction of tokens typically carries the majority of the semantic weight. We address this inefficiency by introducing FrugalPrompt, a novel prompt compression framework for LLMs, which retains only the most semantically significant tokens. Leveraging two state-of-the-art token attribution methods, GlobEnc and DecompX, we assign salience scores to every token in an input sequence, rank them to preserve the top-k% tokens in their original order, and obtain a sparse frugalized prompt. We evaluate the approach across four NLP tasks: Sentiment Analysis, Commonsense QA, Summarization, and Mathematical Reasoning, using a suite of frontier LLMs. For the first three tasks, a 20% prompt reduction incurs only a marginal loss in task performance, demonstrating that contemporary LLMs can reconstruct elided context from high-salience cues. In contrast, performance on mathematical reasoning deteriorates sharply, reflecting a stronger dependence on complete token continuity. Further analysis with bottom-k% and random-k% tokens reveals asymmetric performance patterns that may suggest potential task contamination effects, wherein models may resort to shallow memorized patterns from pretraining exposure for conventional NLP tasks. We posit that our work contributes to a more nuanced understanding of LLM behavior in performance-efficiency trade-offs, and delineate the boundary between tasks tolerant to contextual sparsity and those requiring exhaustive context. Our source code and models are available at: https://github.com/Starscream-11813/Frugal-ICL

[26] TrajSelector: Harnessing Latent Representations for Efficient and Effective Best-of-N in Large Reasoning Model

Bin Yu, Xinming Wang, Shijie Lian, Haotian Li, Changti Wu, Ruina Hu, Bailing Wang, Yuliang Wei, Kai Chen

Main category: cs.CL

TL;DR: TrajSelector is an efficient Best-of-N framework that uses hidden states from LLMs for process-level scoring with a lightweight verifier, achieving better performance than majority voting and process reward models while reducing computational costs.

DetailsMotivation: Address limitations of external test-time scaling approaches: high computational overhead of process reward models and underutilization of LLMs' intrinsic latent representations.

Method: Uses hidden states from sampler LLM for process-level scoring with a 0.6B parameter lightweight verifier that evaluates step-wise trajectory quality and aggregates scores to select optimal reasoning path. Employs end-to-end training without step-level annotations.

Result: Outperforms majority voting by 4.61% accuracy and existing process reward models by 4.31% to 12.21% in Best-of-32 settings across five benchmarks, with lower inference costs.

Conclusion: TrajSelector provides an efficient and effective Best-of-N framework that leverages LLM latent representations to achieve scalable performance improvements with reduced computational overhead.

Abstract: Large language models (LLMs) have shown remarkable progress in complex reasoning tasks, largely enabled by test-time scaling (TTS) paradigms that allocate additional compute during inference. Among these, external TTS (particularly the Best-of-N selection paradigm) yields scalable performance improvements by selecting from multiple independently generated reasoning trajectories. However, this approach faces key limitations: (i) the high computational overhead of deploying process reward models, (ii) the underutilization of the LLM’s intrinsic latent representations. We introduce TrajSelector, an efficient and effective Best-of-N framework that exploit the hidden states in the sampler LLM for process-level scoring. A lightweight verifier (with only 0.6B parameters) evaluates the quality of step-wise trajectory, and then aggregates these scores to identify the optimal reasoning trajectory. Our framework employs a fully data-driven, end-to-end training recipe that eliminates reliance on massive step-level annotations. Experiential results across five benchmarks demonstrate that TrajSelector delivers consistent performance gains. In Best-of-32 settings, it surpasses majority voting by 4.61% accuracy and outperforms existing process reward models by 4.31% to 12.21%, all while maintaining lower inference costs.

[27] RAVEN: Robust Advertisement Video Violation Temporal Grounding via Reinforcement Reasoning

Deyi Ji, Yuekui Yang, Haiyang Wu, Shaoping Ma, Tianrun Chen, Lanyun Zhu

Main category: cs.CL

TL;DR: RAVEN is a framework that combines curriculum reinforcement learning with multimodal LLMs for ad video violation detection, addressing temporal grounding, noisy annotations, and generalization issues.

DetailsMotivation: Existing ad video violation detection methods struggle with precise temporal grounding, noisy annotations, and limited generalization capabilities.

Method: Integrates curriculum reinforcement learning with MLLMs, uses progressive training with mixed precision annotations, employs Group Relative Policy Optimization (GRPO) for emergent reasoning, and implements hierarchical reward mechanisms.

Result: Achieves superior performance in violation category accuracy and temporal interval localization on industrial datasets and public benchmarks, with significant improvements in precision and recall in online A/B testing.

Conclusion: RAVEN demonstrates strong generalization, mitigates catastrophic forgetting, and shows practical applicability for online ad services deployment.

Abstract: Advertisement (Ad) video violation detection is critical for ensuring platform compliance, but existing methods struggle with precise temporal grounding, noisy annotations, and limited generalization. We propose RAVEN, a novel framework that integrates curriculum reinforcement learning with multimodal large language models (MLLMs) to enhance reasoning and cognitive capabilities for violation detection. RAVEN employs a progressive training strategy, combining precisely and coarsely annotated data, and leverages Group Relative Policy Optimization (GRPO) to develop emergent reasoning abilities without explicit reasoning annotations. Multiple hierarchical sophisticated reward mechanism ensures precise temporal grounding and consistent category prediction. Experiments on industrial datasets and public benchmarks show that RAVEN achieves superior performances in violation category accuracy and temporal interval localization. We also design a pipeline to deploy the RAVEN on the online Ad services, and online A/B testing further validates its practical applicability, with significant improvements in precision and recall. RAVEN also demonstrates strong generalization, mitigating the catastrophic forgetting issue associated with supervised fine-tuning.

[28] Agree, Disagree, Explain: Decomposing Human Label Variation in NLI through the Lens of Explanations

Pingjun Hong, Beiduo Chen, Siyao Peng, Marie-Catherine de Marneffe, Benjamin Roth, Barbara Plank

Main category: cs.CL

TL;DR: This paper extends the LiTEx taxonomy to analyze NLI annotation variation beyond just within-label differences, examining how annotators diverge in both reasoning types and labeling decisions.

DetailsMotivation: To better understand human label variation in NLI datasets by using explanations as a lens to decompose the reasoning process and analyze individual differences, moving beyond previous focus on within-label variation.

Method: Applied LiTEx taxonomy to two English NLI datasets, aligning annotation variation through multiple aspects: NLI label agreement, explanation similarity, and taxonomy agreement, while considering annotators’ selection bias.

Result: Found instances where annotators disagree on labels but provide highly similar explanations, suggesting surface-level disagreement may mask underlying agreement. Also revealed individual preferences in explanation strategies and label choices.

Conclusion: Agreement in reasoning types better reflects semantic similarity of free-text explanations than label agreement alone, highlighting the richness of reasoning-based explanations and the need for caution in treating labels as ground truth.

Abstract: Natural Language Inference datasets often exhibit human label variation. To better understand these variations, explanation-based approaches analyze the underlying reasoning behind annotators’ decisions. One such approach is the LiTEx taxonomy, which categorizes free-text explanations in English into reasoning types. However, previous work applying such taxonomies has focused on within-label variation: cases where annotators agree on the final NLI label but provide different explanations. In contrast, this paper broadens the scope by examining how annotators may diverge not only in the reasoning type but also in the labeling step. We use explanations as a lens to decompose the reasoning process underlying NLI annotation and to analyze individual differences. We apply LiTEx to two NLI English datasets and align annotation variation from multiple aspects: NLI label agreement, explanation similarity, and taxonomy agreement, with an additional compounding factor of annotators’ selection bias. We observe instances where annotators disagree on the label but provide highly similar explanations, suggesting that surface-level disagreement may mask underlying agreement in interpretation. Moreover, our analysis reveals individual preferences in explanation strategies and label choices. These findings highlight that agreement in reasoning types better reflects the semantic similarity of free-text explanations than label agreement alone. Our findings underscore the richness of reasoning-based explanations and the need for caution in treating labels as ground truth.

[29] Executable Knowledge Graphs for Replicating AI Research

Yujie Luo, Zhuoyun Yu, Xuehai Wang, Yuqi Zhu, Ningyu Zhang, Lanning Wei, Lun Du, Da Zheng, Huajun Chen

Main category: cs.CL

TL;DR: Proposes Executable Knowledge Graphs (xKG) to improve AI research replication by integrating technical insights, code snippets, and domain knowledge from scientific papers, achieving 10.9% performance gains.

DetailsMotivation: Existing approaches struggle with generating executable code due to insufficient background knowledge and limitations of RAG methods that miss latent technical details in referenced papers.

Method: xKG is a modular knowledge base that automatically extracts and integrates technical insights, code snippets, and domain-specific knowledge from scientific literature to support multi-granular retrieval and reuse.

Result: When integrated into three agent frameworks with two LLMs, xKG shows substantial performance gains (10.9% with o3-mini) on PaperBench.

Conclusion: xKG is an effective and extensible solution for automated AI research replication, demonstrating general applicability across different agent frameworks and LLMs.

Abstract: Replicating AI research is a crucial yet challenging task for large language model (LLM) agents. Existing approaches often struggle to generate executable code, primarily due to insufficient background knowledge and the limitations of retrieval-augmented generation (RAG) methods, which fail to capture latent technical details hidden in referenced papers. Furthermore, previous approaches tend to overlook valuable implementation-level code signals and lack structured knowledge representations that support multi-granular retrieval and reuse. To overcome these challenges, we propose Executable Knowledge Graphs (xKG), a modular and pluggable knowledge base that automatically integrates technical insights, code snippets, and domain-specific knowledge extracted from scientific literature. When integrated into three agent frameworks with two different LLMs, xKG shows substantial performance gains (10.9% with o3-mini) on PaperBench, demonstrating its effectiveness as a general and extensible solution for automated AI research replication. Code will released at https://github.com/zjunlp/xKG.

[30] Check Yourself Before You Wreck Yourself: Selectively Quitting Improves LLM Agent Safety

Vamshi Krishna Bonagiri, Ponnurangam Kumaragurum, Khanh Nguyen, Benjamin Plaut

Main category: cs.CL

TL;DR: Using ‘quitting’ as a safety mechanism for LLM agents to withdraw from uncertain situations improves safety significantly with minimal impact on helpfulness.

DetailsMotivation: LLM agents operating in complex real-world environments face compounding uncertainties that can lead to catastrophic risks beyond traditional text generation failures.

Method: Proposed using explicit quit instructions within the ToolEmu framework, systematically evaluating quitting behavior across 12 state-of-the-art LLMs.

Result: Agents with quit instructions improved safety by +0.39 on 0-3 scale (+0.64 for proprietary models) with only -0.03 decrease in helpfulness.

Conclusion: Quitting serves as an effective first-line defense mechanism that can be immediately deployed in existing agent systems for high-stakes applications.

Abstract: As Large Language Model (LLM) agents increasingly operate in complex environments with real-world consequences, their safety becomes critical. While uncertainty quantification is well-studied for single-turn tasks, multi-turn agentic scenarios with real-world tool access present unique challenges where uncertainties and ambiguities compound, leading to severe or catastrophic risks beyond traditional text generation failures. We propose using “quitting” as a simple yet effective behavioral mechanism for LLM agents to recognize and withdraw from situations where they lack confidence. Leveraging the ToolEmu framework, we conduct a systematic evaluation of quitting behavior across 12 state-of-the-art LLMs. Our results demonstrate a highly favorable safety-helpfulness trade-off: agents prompted to quit with explicit instructions improve safety by an average of +0.39 on a 0-3 scale across all models (+0.64 for proprietary models), while maintaining a negligible average decrease of -0.03 in helpfulness. Our analysis demonstrates that simply adding explicit quit instructions proves to be a highly effective safety mechanism that can immediately be deployed in existing agent systems, and establishes quitting as an effective first-line defense mechanism for autonomous agents in high-stakes applications.

[31] Automated Composition of Agents: A Knapsack Approach for Agentic Component Selection

Michelle Yuan, Khushbu Pahwa, Shuaichen Chang, Mustafa Kaba, Jiarong Jiang, Xiaofei Ma, Yi Zhang, Monica Sunkara

Main category: cs.CL

TL;DR: A framework for automated agentic system composition using online knapsack optimization to select optimal agent components based on performance, budget, and compatibility.

DetailsMotivation: Existing methods for agentic system composition rely on static semantic retrieval, which struggles with incomplete capability descriptions and fails to consider real-time utility, cost, and performance trade-offs.

Method: Introduces an online knapsack-based framework where a composer agent dynamically tests candidate components and models their utility in real-time to assemble optimal agent sets under budget constraints.

Result: Empirical evaluation shows the online knapsack composer achieves up to 31.6% higher success rates than retrieval baselines in single-agent setups, and increases success rate from 37% to 87% in multi-agent systems with 100+ agents.

Conclusion: The framework enables scalable reuse of agentic components and consistently achieves Pareto-optimal performance across diverse domains and budget constraints.

Abstract: Designing effective agentic systems requires the seamless composition and integration of agents, tools, and models within dynamic and uncertain environments. Most existing methods rely on static, semantic retrieval approaches for tool or agent discovery. However, effective reuse and composition of existing components remain challenging due to incomplete capability descriptions and the limitations of retrieval methods. Component selection suffers because the decisions are not based on capability, cost, and real-time utility. To address these challenges, we introduce a structured, automated framework for agentic system composition that is inspired by the knapsack problem. Our framework enables a composer agent to systematically identify, select, and assemble an optimal set of agentic components by jointly considering performance, budget constraints, and compatibility. By dynamically testing candidate components and modeling their utility in real-time, our approach streamlines the assembly of agentic systems and facilitates scalable reuse of resources. Empirical evaluation with Claude 3.5 Sonnet across five benchmarking datasets shows that our online-knapsack-based composer consistently lies on the Pareto frontier, achieving higher success rates at significantly lower component costs compared to our baselines. In the single-agent setup, the online knapsack composer shows a success rate improvement of up to 31.6% in comparison to the retrieval baselines. In multi-agent systems, the online knapsack composer increases success rate from 37% to 87% when agents are selected from an agent inventory of 100+ agents. The substantial performance gap confirms the robust adaptability of our method across diverse domains and budget constraints.

[32] ReviewGuard: Enhancing Deficient Peer Review Detection via LLM-Driven Data Augmentation

Haoxuan Zhang, Ruochi Li, Sarthak Shrestha, Shree Harshini Mamidala, Revanth Putta, Arka Krishan Aggarwal, Ting Xiao, Junhua Ding, Haihua Chen

Main category: cs.CL

TL;DR: ReviewGuard is an automated system that detects deficient peer reviews using a four-stage LLM-driven framework, addressing challenges from increased submissions and AI-generated reviews in scholarly evaluation.

DetailsMotivation: The surge in submissions and widespread adoption of LLMs in scholarly evaluation present unprecedented challenges, with unchecked deficient reviews from both human experts and AI systems threatening to undermine the peer review ecosystem and compromise academic integrity.

Method: A comprehensive four-stage LLM-driven framework: (1) collects ICLR and NeurIPS papers with reviews from OpenReview; (2) annotates review types using GPT-4.1 with human validation; (3) addresses class imbalance through LLM-driven synthetic data augmentation; (4) fine-tunes encoder-based models and open source LLMs.

Result: Created a corpus of 6,634 papers, 24,657 real reviews, and 46,438 synthetic reviews. Deficient reviews show lower rating scores, higher confidence, reduced structural complexity, and more negative sentiment. AI-generated reviews increased dramatically post-ChatGPT. Mixed training with synthetic and real data improved recall and F1 scores.

Conclusion: This study presents the first LLM-driven system for detecting deficient peer reviews, providing evidence to inform AI governance in peer review and offering insights into human-AI collaboration to maintain academic integrity.

Abstract: Peer review serves as the gatekeeper of science, yet the surge in submissions and widespread adoption of large language models (LLMs) in scholarly evaluation present unprecedented challenges. Recent work has focused on using LLMs to improve review efficiency or generate insightful review content. However, unchecked deficient reviews from both human experts and AI systems threaten to systematically undermine the peer review ecosystem and compromise academic integrity. To address this critical issue, we introduce ReviewGuard, an automated system for detecting and categorizing deficient reviews. ReviewGuard employs a comprehensive four-stage LLM-driven framework that: (1) collects ICLR and NeurIPS papers with their corresponding reviews from OpenReview; (2) annotates review types using GPT-4.1 with human validation; (3) addresses class imbalance and data scarcity through LLM-driven synthetic data augmentation, producing a final corpus of 6,634 papers, 24,657 real reviews, and 46,438 synthetic reviews; and (4) fine-tunes both encoder-based models and open source LLMs. We perform comprehensive feature analysis of the structure and quality of the review text. Compared to sufficient reviews, deficient reviews demonstrate lower rating scores, higher self-reported confidence, reduced structural complexity, and a higher proportion of negative sentiment. AI-generated text detection reveals that, since ChatGPT’s emergence, AI-generated reviews have increased dramatically. In the evaluation of deficient review detection models, mixed training with synthetic and real review data provides substantial enhancements to recall and F1 scores on the binary task. This study presents the first LLM-driven system for detecting deficient peer reviews, providing evidence to inform AI governance in peer review while offering valuable insights into human-AI collaboration to maintain academic integrity.

[33] Language over Content: Tracing Cultural Understanding in Multilingual Large Language Models

Seungho Cho, Changgeon Ko, Eui Jun Hwang, Junmyeong Lee, Huije Lee, Jong C. Park

Main category: cs.CL

TL;DR: This paper traces LLMs’ internal cultural understanding mechanisms by analyzing activation path overlaps when answering semantically equivalent questions across different countries and languages, revealing strong language-specific patterns and that linguistic similarity doesn’t guarantee aligned internal representations.

DetailsMotivation: LLMs are increasingly used across diverse cultural contexts, but prior evaluations focused mainly on output-level performance, obscuring the internal factors driving response differences. Circuit analysis studies have covered few languages and rarely focused on cultural understanding.

Method: The researchers traced LLMs’ internal cultural understanding by measuring activation path overlaps when answering semantically equivalent questions under two conditions: varying target country while fixing question language, and varying question language while fixing country. They also used same-language country pairs to disentangle language from cultural aspects.

Result: Results show that internal paths overlap more for same-language, cross-country questions than for cross-language, same-country questions, indicating strong language-specific patterns. The South Korea-North Korea pair exhibited low overlap and high variability, showing linguistic similarity doesn’t guarantee aligned internal representation.

Conclusion: The study reveals that LLMs exhibit strong language-specific patterns in cultural understanding, and linguistic similarity between countries doesn’t necessarily translate to aligned internal representations, highlighting the complexity of cultural understanding mechanisms in language models.

Abstract: Large language models (LLMs) are increasingly used across diverse cultural contexts, making accurate cultural understanding essential. Prior evaluations have mostly focused on output-level performance, obscuring the factors that drive differences in responses, while studies using circuit analysis have covered few languages and rarely focused on culture. In this work, we trace LLMs’ internal cultural understanding mechanisms by measuring activation path overlaps when answering semantically equivalent questions under two conditions: varying the target country while fixing the question language, and varying the question language while fixing the country. We also use same-language country pairs to disentangle language from cultural aspects. Results show that internal paths overlap more for same-language, cross-country questions than for cross-language, same-country questions, indicating strong language-specific patterns. Notably, the South Korea-North Korea pair exhibits low overlap and high variability, showing that linguistic similarity does not guarantee aligned internal representation.

[34] AI-Generated Text Detection in Low-Resource Languages: A Case Study on Urdu

Muhammad Ammar, Hadiya Murad Hadi, Usman Majeed Butt

Main category: cs.CL

TL;DR: Proposes a novel AI-generated text detection framework for Urdu language using multilingual transformer models, achieving 91.26% accuracy with mDeBERTa-v3-base.

DetailsMotivation: Address the challenge of detecting AI-generated text in Urdu language where few detection tools exist, to combat misinformation and academic misconduct in Urdu-speaking communities.

Method: Developed balanced dataset (1,800 human + 1,800 AI texts from Gemini, GPT-4o-mini, Kimi AI), conducted linguistic/statistical analysis, and fine-tuned three multilingual transformers (mdeberta-v3-base, distilbert-base-multilingualcased, xlm-roberta-base).

Result: mDeBERTa-v3-base achieved highest performance with 91.29 F1-score and 91.26% accuracy on test set, significantly advancing Urdu AI text detection capabilities.

Conclusion: The research successfully addresses the gap in Urdu AI text detection, contributes to NLP tools for low-resource languages, and helps combat misinformation in Urdu-speaking communities.

Abstract: Large Language Models (LLMs) are now capable of generating text that closely resembles human writing, making them powerful tools for content creation, but this growing ability has also made it harder to tell whether a piece of text was written by a human or by a machine. This challenge becomes even more serious for languages like Urdu, where there are very few tools available to detect AI-generated text. To address this gap, we propose a novel AI-generated text detection framework tailored for the Urdu language. A balanced dataset comprising 1,800 humans authored, and 1,800 AI generated texts, sourced from models such as Gemini, GPT-4o-mini, and Kimi AI was developed. Detailed linguistic and statistical analysis was conducted, focusing on features such as character and word counts, vocabulary richness (Type Token Ratio), and N-gram patterns, with significance evaluated through t-tests and MannWhitney U tests. Three state-of-the-art multilingual transformer models such as mdeberta-v3-base, distilbert-base-multilingualcased, and xlm-roberta-base were fine-tuned on this dataset. The mDeBERTa-v3-base achieved the highest performance, with an F1-score 91.29 and accuracy of 91.26% on the test set. This research advances efforts in contesting misinformation and academic misconduct in Urdu-speaking communities and contributes to the broader development of NLP tools for low resource languages.

[35] Fine-tuning of Large Language Models for Constituency Parsing Using a Sequence to Sequence Approach

Francisco Jose Cortes Delgado, Eduardo Martinez Gracia, Rafael Valencia Garcia

Main category: cs.CL

TL;DR: Fine-tuning large language models to translate sentences into syntactic structures for Spanish syntax analysis, achieving high accuracy.

DetailsMotivation: Extend the capabilities of MiSintaxis, a tool for teaching Spanish syntax, by leveraging recent advances in natural language processing with large neural models.

Method: Fine-tuned several models from Hugging Face repository using training data generated from AnCora-ES corpus to translate input sentences into corresponding syntactic structures.

Result: Demonstrated high accuracy in phrase-structure analysis as evaluated by F1 score.

Conclusion: The methodology shows great potential for syntactic analysis using fine-tuned large language models.

Abstract: Recent advances in natural language processing with large neural models have opened new possibilities for syntactic analysis based on machine learning. This work explores a novel approach to phrase-structure analysis by fine-tuning large language models (LLMs) to translate an input sentence into its corresponding syntactic structure. The main objective is to extend the capabilities of MiSintaxis, a tool designed for teaching Spanish syntax. Several models from the Hugging Face repository were fine-tuned using training data generated from the AnCora-ES corpus, and their performance was evaluated using the F1 score. The results demonstrate high accuracy in phrase-structure analysis and highlight the potential of this methodology.

[36] All You Need is One: Capsule Prompt Tuning with a Single Vector

Yiyang Liu, James C. Liang, Heng Fan, Wenhao Yang, Yiming Cui, Xiaotian Han, Lifu Huang, Dongfang Liu, Qifan Wang, Cheng Han

Main category: cs.CL

TL;DR: Capsule Prompt-Tuning (CaPT) is a parameter-efficient method that incorporates instance-aware information into prompt-based learning using a single capsule prompt, achieving superior performance across various language tasks with minimal parameter overhead.

DetailsMotivation: Current prompt-based learning methods rely on laborious grid searching for optimal prompt length and lack instance-aware information, leading to suboptimal attention interplay with input sequences.

Method: Introduces Capsule Prompt-Tuning (CaPT) that leverages off-the-shelf instance semantics to integrate both instance-aware and task-aware information in a nearly parameter-free manner using a single capsule prompt.

Result: Achieves 84.03% average accuracy on T5-Large and uses only 0.003% of model parameters on Llama3.2-1B, demonstrating superior performance and high parameter efficiency.

Conclusion: CaPT effectively serves as an “attention anchor” that preserves strong attention to critical structural information and exhibits active attention interaction with all input tokens, making prompt-based learning more efficient and effective.

Abstract: Prompt-based learning has emerged as a parameter-efficient finetuning (PEFT) approach to facilitate Large Language Model (LLM) adaptation to downstream tasks by conditioning generation with task-aware guidance. Despite its successes, current prompt-based learning methods heavily rely on laborious grid searching for optimal prompt length and typically require considerable number of prompts, introducing additional computational burden. Worse yet, our pioneer findings indicate that the task-aware prompt design is inherently limited by its absence of instance-aware information, leading to a subtle attention interplay with the input sequence. In contrast, simply incorporating instance-aware information as a part of the guidance can enhance the prompt-tuned model performance without additional fine-tuning. Moreover, we find an interesting phenomenon, namely “attention anchor”, that incorporating instance-aware tokens at the earliest position of the sequence can successfully preserve strong attention to critical structural information and exhibit more active attention interaction with all input tokens. In light of our observation, we introduce Capsule Prompt-Tuning (CaPT), an efficient and effective solution that leverages off-the-shelf, informative instance semantics into prompt-based learning. Our approach innovatively integrates both instance-aware and task-aware information in a nearly parameter-free manner (i.e., one single capsule prompt). Empirical results demonstrate that our method can exhibit superior performance across various language tasks (e.g., 84.03% average accuracy on T5-Large), serving as an “attention anchor,” while enjoying high parameter efficiency (e.g., 0.003% of model parameters on Llama3.2-1B).

[37] Temporal Understanding under Deictic Frame of Reference

Damin Zhang, Julia Rayz

Main category: cs.CL

TL;DR: The paper introduces TUuD framework to evaluate how LLMs interpret temporal relations when the reference point of “now” dynamically shifts, showing they exhibit partial human-like temporal cognition but with limitations in long-term contexts.

DetailsMotivation: While LLMs have advanced in natural language understanding, their ability to interpret and reason about time remains limited, particularly regarding temporal frames of reference that humans use to conceptualize time through spatial metaphors.

Method: The TUuD framework prompts LLMs to rate similarity between current moment and target events (0.00-1.00 scale) when the reference point of “now” dynamically shifts along a timeline, evaluating temporal understanding under deictic temporal frames of reference.

Result: Four evaluated LLMs show measurable adaptation to deictic t-FoR, with similarity ratings peaking around present and decreasing toward past/future events, but this adaptation weakens beyond near-term contexts.

Conclusion: LLMs display partial human-like temporal cognition but their temporal reasoning remains sensitive to reference-frame shifts and temporal distance, indicating limitations in temporal understanding.

Abstract: Understanding time is fundamental to human cognition, where temporal experience is often conceptualized through spatial metaphors grounded in sensory-motor experience. For example, “summer is approaching” parallels “We are approaching the summer”. In such expressions, humans rely on a frame of reference (FoR) to interpret meaning relative to a particular viewpoint. Extending this concept to time, a temporal frame of reference (t-FoR) defines how temporal relations are perceived relative to an experiencer’s moment of “now”. While Large Language Models (LLMs) have shown remarkable advances in natural language understanding, their ability to interpret and reason about time remains limited. In this work, we introduce TUuD (Temporal Understanding under Deictic t-FoR), a framework that evaluates how LLMs interpret time-event and event-event relations when the reference point of “now” dynamically shifts along a timeline. Following recent work on temporal cognition \cite{li2025other}, LLMs are prompted to rate the similarity between the current moment and a target event from 0.00 (completely dissimilar) to 1.00 (highly similar), where similarity quantifies perceived temporal alignment between the two points. Our results show that four evaluated LLMs exhibit measurable adaptation to a deictic t-FoR, with similarity ratings peaking around the present and decreasing toward past and future events. The adaptation, however, weakens beyond near-term contexts, suggesting that while LLMs display partial human-like temporal cognition, their temporal reasoning remains sensitive to reference-frame shifts and temporal distance.

[38] Investigating the Impact of Rationales for LLMs on Natural Language Understanding

Wenhang Shi, Shuqing Bian, Yiren Chen, Xinyi Zhang, Zhe Zhao, Pengfei Hu, Wei Lu, Xiaoyong Du

Main category: cs.CL

TL;DR: Chain-of-thought rationales benefit reasoning tasks but their impact on natural language understanding (NLU) tasks was unexplored. This work systematically investigates rationale-augmented methods for NLU tasks using a new dataset NLURC.

DetailsMotivation: Most research on chain-of-thought rationales focuses on reasoning tasks, overlooking their potential benefits for natural language understanding tasks. The authors aim to explore whether rationales can similarly improve performance on NLU tasks.

Method: Constructed NLURC (a comprehensive NLU dataset with rationales) and developed various rationale-augmented methods. Systematically explored these methods on NLU tasks to evaluate their effectiveness.

Result: Three key findings: (1) CoT inference shifts from hindering to surpassing direct prediction as model size grows; (2) Most rationale-augmented training performs worse than label-only training, except one specially designed method; (3) Models trained with rationales achieve significant gains on unseen tasks, rivaling much larger models while maintaining interpretability.

Conclusion: Rationales can benefit NLU tasks, particularly showing positive correlation with model size, and specially designed rationale-augmented training methods can achieve competitive performance with improved interpretability on unseen tasks.

Abstract: Chain-of-thought (CoT) rationales, which provide step-by-step reasoning to derive final answers, benefit LLMs in both inference and training. Incorporating rationales, either by generating them before answering during inference, or by placing them before or after the original answers during training - significantly improves model performance on mathematical, symbolic and commonsense reasoning tasks. However, most work focuses on the role of rationales in these reasoning tasks, overlooking their potential impact on other important tasks like natural language understanding (NLU) tasks. In this work, we raise the question: Can rationales similarly benefit NLU tasks? To conduct a systematic exploration, we construct NLURC, a comprehensive and high-quality NLU dataset collection with rationales, and develop various rationale-augmented methods. Through exploring the applicability of these methods on NLU tasks using the dataset, we uncover several potentially surprising findings: (1) CoT inference shifts from hindering NLU performance to surpassing direct label prediction as model size grows, indicating a positive correlation. (2) Most rationale-augmented training methods perform worse than label-only training, with one specially designed method consistently achieving improvements. (3) LLMs trained with rationales achieve significant performance gains on unseen NLU tasks, rivaling models ten times their size, while delivering interpretability on par with commercial LLMs.

[39] Natural Language Processing Applications in Cardiology: A Narrative Review

Kailai Yang, Yan Leng, Xin Zhang, Tianlin Zhang, Paul Thompson, Bernard Keavney, Maciej Tomaszewski, Sophia Ananiadou

Main category: cs.CL

TL;DR: This paper provides a comprehensive review of natural language processing (NLP) applications in cardiology from 2014 to 2025, analyzing 265 relevant articles across multiple dimensions including NLP paradigms, cardiology tasks, disease types, and data sources.

DetailsMotivation: Cardiovascular disease is a complex global health issue influenced by multiple factors, with relevant information scattered across diverse textual data sources. NLP techniques offer powerful tools to analyze this unstructured data and gain deeper insights for diagnosis, treatment, and prevention of cardiac conditions.

Method: The authors conducted a systematic review by querying six literature databases to identify articles applying NLP techniques in cardiovascular disease contexts. They screened and analyzed 265 relevant articles across multiple dimensions: NLP paradigm types, cardiology task types, cardiovascular disease types, and data source types, including temporal analysis of trends.

Result: The analysis revealed considerable diversity across all dimensions studied, demonstrating the breadth of NLP research in cardiology. Temporal analysis showed the evolution and changing trends in NLP methods over the past decade, providing the most comprehensive overview of this research area to date.

Conclusion: This review constitutes the most comprehensive overview of NLP research in cardiology, highlighting the field’s diversity and evolution over time, which can help healthcare professionals gain deeper insights and potentially revolutionize approaches to cardiac disease diagnosis, treatment, and prevention.

Abstract: Cardiovascular disease has become increasingly prevalent in modern society and has a significant effect on global health and well-being. Heart-related conditions are intricate, multifaceted disorders, which may be influenced by a combination of genetic predispositions, lifestyle choices, and various socioeconomic and clinical factors. Information regarding these potentially complex interrelationships is dispersed among diverse types of textual data, which include patient narratives, medical records, and scientific literature, among others. Natural language processing (NLP) techniques have increasingly been adopted as a powerful means to analyse and make sense of this vast amount of unstructured data. This, in turn, can allow healthcare professionals to gain deeper insights into the cardiology field, which has the potential to revolutionize current approaches to the diagnosis, treatment, and prevention of cardiac problems. This review provides a detailed overview of NLP research in cardiology between 2014 and 2025. We queried six literature databases to find articles describing the application of NLP techniques in the context of a range of different cardiovascular diseases. Following a rigorous screening process, we identified a total of 265 relevant articles. We analysed each article from multiple dimensions, i.e., NLP paradigm types, cardiology-related task types, cardiovascular disease types, and data source types. Our analysis reveals considerable diversity within each of these dimensions, thus demonstrating the considerable breadth of NLP research within the field. We also perform a temporal analysis, which illustrates the evolution and changing trends in NLP methods employed over the last decade that we cover. To our knowledge, the review constitutes the most comprehensive overview of NLP research in cardiology to date.

[40] Value-Based Large Language Model Agent Simulation for Mutual Evaluation of Trust and Interpersonal Closeness

Yuki Sakamoto, Takahisa Uchida, Hiroshi Ishiguro

Main category: cs.CL

TL;DR: This study examines whether value similarity affects relationship-building among LLM agents, finding that agents with similar values show greater trust and interpersonal closeness.

DetailsMotivation: To investigate if human social principles like value similarity building trust apply to artificial societies of LLM agents, and to validate LLM simulations as testbeds for social science theories.

Method: Two experiments: first evaluated controllability of values in LLMs to identify optimal models and prompts, then generated agent pairs with specific values and analyzed their mutual evaluations after dialogues in English and Japanese.

Result: Pairs of LLM agents with higher value similarity exhibited greater mutual trust and interpersonal closeness, confirming the principle holds in artificial societies.

Conclusion: LLM agent simulations serve as valid testbeds for social science theories, help explain how values influence relationship building, and provide foundations for new social science insights.

Abstract: Large language models (LLMs) have emerged as powerful tools for simulating complex social phenomena using human-like agents with specific traits. In human societies, value similarity is important for building trust and close relationships; however, it remains unexplored whether this principle holds true in artificial societies comprising LLM agents. Therefore, this study investigates the influence of value similarity on relationship-building among LLM agents through two experiments. First, in a preliminary experiment, we evaluated the controllability of values in LLMs to identify the most effective model and prompt design for controlling the values. Subsequently, in the main experiment, we generated pairs of LLM agents imbued with specific values and analyzed their mutual evaluations of trust and interpersonal closeness following a dialogue. The experiments were conducted in English and Japanese to investigate language dependence. The results confirmed that pairs of agents with higher value similarity exhibited greater mutual trust and interpersonal closeness. Our findings demonstrate that the LLM agent simulation serves as a valid testbed for social science theories, contributes to elucidating the mechanisms by which values influence relationship building, and provides a foundation for inspiring new theories and insights into the social sciences.

[41] The Chameleon Nature of LLMs: Quantifying Multi-Turn Stance Instability in Search-Enabled Language Models

Shivam Ratnakar, Sanjay Raghavendra

Main category: cs.CL

TL;DR: LLMs exhibit ‘chameleon behavior’ - shifting stances when presented with contradictory questions in multi-turn conversations, particularly in search-enabled systems, revealing fundamental reliability issues.

DetailsMotivation: To systematically investigate a critical vulnerability in LLM-search integration systems where models change positions when faced with contradictory questions, undermining reliability in critical applications.

Method: Created Chameleon Benchmark Dataset with 17,770 question-answer pairs across 1,180 multi-turn conversations spanning 12 controversial domains. Introduced Chameleon Score (stance instability) and Source Re-use Rate (knowledge diversity) metrics. Evaluated Llama-4-Maverick, GPT-4o-mini, and Gemini-2.5-Flash.

Result: All models showed severe chameleon behavior (scores 0.391-0.511), with GPT-4o-mini performing worst. Strong correlations found between source re-use rate and confidence (r=0.627) and stance changes (r=0.429), indicating limited knowledge diversity makes models deferential to query framing.

Conclusion: LLMs exhibit pathological deference to query framing due to limited knowledge diversity, highlighting the need for comprehensive consistency evaluation before deployment in healthcare, legal, and financial systems where coherent positions are critical.

Abstract: Integration of Large Language Models with search/retrieval engines has become ubiquitous, yet these systems harbor a critical vulnerability that undermines their reliability. We present the first systematic investigation of “chameleon behavior” in LLMs: their alarming tendency to shift stances when presented with contradictory questions in multi-turn conversations (especially in search-enabled LLMs). Through our novel Chameleon Benchmark Dataset, comprising 17,770 carefully crafted question-answer pairs across 1,180 multi-turn conversations spanning 12 controversial domains, we expose fundamental flaws in state-of-the-art systems. We introduce two theoretically grounded metrics: the Chameleon Score (0-1) that quantifies stance instability, and Source Re-use Rate (0-1) that measures knowledge diversity. Our rigorous evaluation of Llama-4-Maverick, GPT-4o-mini, and Gemini-2.5-Flash reveals consistent failures: all models exhibit severe chameleon behavior (scores 0.391-0.511), with GPT-4o-mini showing the worst performance. Crucially, small across-temperature variance (less than 0.004) suggests the effect is not a sampling artifact. Our analysis uncovers the mechanism: strong correlations between source re-use rate and confidence (r=0.627) and stance changes (r=0.429) are statistically significant (p less than 0.05), indicating that limited knowledge diversity makes models pathologically deferential to query framing. These findings highlight the need for comprehensive consistency evaluation before deploying LLMs in healthcare, legal, and financial systems where maintaining coherent positions across interactions is critical for reliable decision support.

[42] ReaGAN: Node-as-Agent-Reasoning Graph Agentic Network

Minghao Guo, Xi Zhu, Haochen Xue, Chong Zhang, Shuhang Lin, Jingyuan Huang, Ziyi Ye, Yongfeng Zhang

Main category: cs.CL

TL;DR: ReaGAN is an agent-based GNN framework where nodes act as autonomous agents with internal memory, enabling adaptive message propagation and global semantic relationships through retrieval-augmented generation.

DetailsMotivation: Traditional GNNs have limitations in handling node informativeness imbalance and capturing global semantic relationships due to fixed aggregation schemes and local structural focus.

Method: Each node acts as an autonomous agent with internal memory for node-level planning. Uses retrieval-augmented generation (RAG) to access semantically relevant content and build global relationships.

Result: Achieves competitive performance under few-shot in-context settings using a frozen LLM backbone without fine-tuning.

Conclusion: Demonstrates the potential of agentic planning and local-global retrieval in graph learning, enabling adaptive message propagation and global relationship building.

Abstract: Graph Neural Networks (GNNs) have achieved remarkable success in graph-based learning by propagating information among neighbor nodes via predefined aggregation mechanisms. However, such fixed schemes often suffer from two key limitations. First, they cannot handle the imbalance in node informativeness – some nodes are rich in information, while others remain sparse. Second, predefined message passing primarily leverages local structural similarity while ignoring global semantic relationships across the graph, limiting the model’s ability to capture distant but relevant information. We propose Retrieval-augmented Graph Agentic Network (ReaGAN), an agent-based framework that empowers each node with autonomous, node-level decision-making. Each node acts as an agent that independently plans its next action based on its internal memory, enabling node-level planning and adaptive message propagation. Additionally, retrieval-augmented generation (RAG) allows nodes to access semantically relevant content and build global relationships in the graph. ReaGAN achieves competitive performance under few-shot in-context settings using a frozen LLM backbone without fine-tuning, showcasing the potential of agentic planning and local-global retrieval in graph learning.

[43] so much depends / upon / a whitespace: Why Whitespace Matters for Poets and LLMs

Sriharsh Bhyravajjula, Melanie Walsh, Anna Preus, Maria Antoniak

Main category: cs.CL

TL;DR: This paper analyzes whitespace usage in poetry across published works, LLM-generated poems, and online unpublished poems, revealing differences in whitespace patterns and discussing implications for LLM pretraining data processing.

DetailsMotivation: Whitespace is a crucial but overlooked element in poetry that reflects artistic choices, yet has received insufficient attention in NLP research despite poetry's popularity as an art form and LLM generation task.

Method: Analyzed 19k English poems from Poetry Foundation, comparing whitespace usage across 4k poets, 51k LLM-generated poems, and 12k unpublished online poems. Examined variations across time periods, poetic forms, and data sources.

Result: Found significant differences in whitespace usage between published poems, LLM-generated poems, and unpublished online poems. Different text processing methods produce substantially different whitespace representations in poetry data.

Conclusion: Whitespace patterns in poetry reveal important artistic and semantic information. The findings highlight implications for how pretraining datasets for LLMs are processed, suggesting current methods may inadequately preserve whitespace features.

Abstract: Whitespace is a critical component of poetic form, reflecting both adherence to standardized forms and rebellion against those forms. Each poem’s whitespace distribution reflects the artistic choices of the poet and is an integral semantic and spatial feature of the poem. Yet, despite the popularity of poetry as both a long-standing art form and as a generation task for large language models (LLMs), whitespace has not received sufficient attention from the NLP community. Using a corpus of 19k English-language published poems from Poetry Foundation, we investigate how 4k poets have used whitespace in their works. We release a subset of 2.8k public-domain poems with preserved formatting to facilitate further research in this area. We compare whitespace usage in the published poems to (1) 51k LLM-generated poems, and (2) 12k unpublished poems posted in an online community. We also explore whitespace usage across time periods, poetic forms, and data sources. Additionally, we find that different text processing methods can result in significantly different representations of whitespace in poetry data, motivating us to use these poems and whitespace patterns to discuss implications for the processing strategies used to assemble pretraining datasets for LLMs.

[44] Beacon: Single-Turn Diagnosis and Mitigation of Latent Sycophancy in Large Language Models

Sanskar Pandey, Ruhaan Chopra, Angkul Puniya, Sohom Pal

Main category: cs.CL

TL;DR: Beacon is a benchmark that measures sycophancy in LLMs - the tendency to prioritize agreement with users over factual accuracy. It reveals this bias decomposes into linguistic and affective components that scale with model size, and proposes interventions to modulate this trade-off.

DetailsMotivation: To address the structural trade-off in LLMs between truthfulness and obsequious flattery (sycophancy) that emerges from reward optimization conflating helpfulness with polite submission, enabling precise measurement of this bias independent of conversational context.

Method: Introduces Beacon, a single-turn forced-choice benchmark that isolates sycophancy bias. Evaluates twelve state-of-the-art models and proposes both prompt-level and activation-level interventions to modulate these biases.

Result: Evaluations reveal sycophancy decomposes into stable linguistic and affective sub-biases that scale with model capacity. The interventions can modulate these biases in opposing directions, exposing the internal geometry of alignment as a dynamic manifold.

Conclusion: Beacon reframes sycophancy as a measurable form of normative misgeneralization, providing a reproducible foundation for studying and mitigating alignment drift in large-scale generative systems.

Abstract: Large language models internalize a structural trade-off between truthfulness and obsequious flattery, emerging from reward optimization that conflates helpfulness with polite submission. This latent bias, known as sycophancy, manifests as a preference for user agreement over principled reasoning. We introduce Beacon, a single-turn forced-choice benchmark that isolates this bias independent of conversational context, enabling precise measurement of the tension between factual accuracy and submissive bias. Evaluations across twelve state-of-the-art models reveal that sycophancy decomposes into stable linguistic and affective sub-biases, each scaling with model capacity. We further propose prompt-level and activation-level interventions that modulate these biases in opposing directions, exposing the internal geometry of alignment as a dynamic manifold between truthfulness and socially compliant judgment. Beacon reframes sycophancy as a measurable form of normative misgeneralization, providing a reproducible foundation for studying and mitigating alignment drift in large-scale generative systems.

[45] Enhancing Language Agent Strategic Reasoning through Self-Play in Adversarial Games

Yikai Zhang, Ye Rong, Siyu Yuan, Jiangjie Chen, Jian Xie, Yanghua Xiao

Main category: cs.CL

TL;DR: SCO-PAL is a step-level policy optimization method that uses self-play to improve strategic reasoning in adversarial games, achieving significant performance gains over baselines.

DetailsMotivation: Existing language agents struggle with strategic reasoning in dynamic adversarial games, and current approaches rely on costly expert-labeled data rather than learning automatically from game interactions.

Method: Proposed SCO-PAL (Step-level poliCy Optimization through Play-And-Learn) method that analyzes opponent selection at different levels and identifies self-play as the most effective approach.

Result: SCO-PAL with self-play increased average win rate by approximately 30% against four opponents compared to baselines, and achieved 54.76% win rate against GPT-4 in six adversarial games.

Conclusion: Self-play is the most effective way to improve strategic reasoning in adversarial environments, and SCO-PAL demonstrates significant performance improvements through this approach.

Abstract: Existing language agents often encounter difficulties in dynamic adversarial games due to poor strategic reasoning. To mitigate this limitation, a promising approach is to allow agents to learn from game interactions automatically, without relying on costly expert-labeled data. Unlike static environments where agents receive fixed feedback or rewards, selecting appropriate opponents in dynamic adversarial games can significantly impact learning performance. However, the discussion of opponents in adversarial environments remains an area under exploration. In this paper, we propose a Step-level poliCy Optimization method through Play-And-Learn, SCO-PAL. Leveraging SCO-PAL, we conduct a detailed analysis of opponent selection by setting opponents at different levels and find that self-play is the most effective way to improve strategic reasoning in such adversarial environments. Utilizing SCO-PAL with self-play, we increase the average win rate against four opponents by approximately 30% compared to baselines and achieve a 54.76% win rate against GPT-4 in six adversarial games.

[46] LC-Eval: A Bilingual Multi-Task Evaluation Benchmark for Long-Context Understanding

Sheikh Jubair, Arwa Omayrah, Amal Alshammari, Alhanoof Althnian, Abdulhamed Alothaimen, Norah A. Alzahrani, Shahad D. Alzaidi, Nora Al-Twairesh, Abdulmohsen Al-Thubaity

Main category: cs.CL

TL;DR: LC-Eval is a bilingual (English-Arabic) benchmark for evaluating long-context understanding in LLMs across 4k-128k token ranges, featuring four challenging tasks that test deep reasoning, document comprehension, and bilingual information processing.

DetailsMotivation: As LLMs develop sophisticated long-context capabilities, there's a need for rigorous evaluation methods to assess their performance in extended context understanding, particularly across different languages and text genres.

Method: Developed LC-Eval benchmark with four novel tasks: multi-document QA, bilingual QA, claim verification, and long-context multiple-choice questions. Created datasets in both English and Arabic for comparative analysis across text genres.

Result: Evaluations on open-weight and closed LLMs showed LC-Eval presents significant challenges. Even high-performing models like GPT-4o struggled with certain tasks, demonstrating the benchmark’s complexity and rigor.

Conclusion: LC-Eval effectively exposes limitations in current LLMs’ long-context understanding capabilities, particularly in bilingual settings and complex reasoning tasks, highlighting the need for continued model improvement in extended context processing.

Abstract: Recent advancements in Large Language Models (LLMs) have demonstrated sophisticated capabilities, including the ability to process and comprehend extended contexts. These emergent capabilities necessitate rigorous evaluation methods to effectively assess their performance in long-context understanding. In this paper, we present \textbf{LC-Eval}, a bilingual, multi-task evaluation benchmark designed to evaluate long-context understanding in English and Arabic, targeting context lengths ranging from 4k to over 128k tokens. LC-Eval introduces four novel and challenging tasks: multi-document question answering, bilingual question answering, claim verification within a paragraph, and multiple-choice questions based on long contexts. These tasks are designed to assess LLMs’ abilities in deep reasoning, document comprehension, information tracing, and bilingual information extraction and understanding. The benchmark includes datasets in both Arabic and English for each task, allowing for a comparative analysis of their performance across different text genres. Evaluations were conducted on both open-weight and closed LLMs, with results indicating that LC-Eval presents significant challenges. Even high-performing models, such as GPT-4o, struggled with certain tasks, highlighting the complexity and rigor of the benchmark.

[47] MOSAIC: Masked Objective with Selective Adaptation for In-domain Contrastive Learning

Vera Pavlova, Mohammed Makhlouf

Main category: cs.CL

TL;DR: MOSAIC is a multi-stage framework for domain adaptation of sentence embedding models that combines masked language modeling and contrastive learning objectives to adapt general-domain models to specialized domains.

DetailsMotivation: Address the challenges of adapting large-scale general-domain sentence embedding models to specialized domains while preserving their semantic discrimination capabilities.

Method: Multi-stage framework with joint optimization of masked language modeling (MLM) and contrastive objectives within a unified training pipeline, using selective adaptation for in-domain contrastive learning.

Result: Achieved improvements up to 13.4% in NDCG@10 over strong general-domain baselines, validated on both high-resource and low-resource domains.

Conclusion: The framework effectively learns domain-relevant representations while maintaining robust semantic discrimination, with ablation studies confirming the importance of balanced joint supervision and staged adaptation.

Abstract: We introduce MOSAIC (Masked Objective with Selective Adaptation for In-domain Contrastive learning), a multi-stage framework for domain adaptation of sentence embedding models that incorporates joint domain-specific masked supervision. Our approach addresses the challenges of adapting large-scale general-domain sentence embedding models to specialized domains. By jointly optimizing masked language modeling (MLM) and contrastive objectives within a unified training pipeline, our method enables effective learning of domain-relevant representations while preserving the robust semantic discrimination properties of the original model. We empirically validate our approach on both high-resource and low-resource domains, achieving improvements up to 13.4% in NDCG@10 (Normalized Discounted Cumulative Gain) over strong general-domain baselines. Comprehensive ablation studies further demonstrate the effectiveness of each component, highlighting the importance of balanced joint supervision and staged adaptation.

[48] Knowing the Facts but Choosing the Shortcut: Understanding How Large Language Models Compare Entities

Hans Hergen Lehmann, Jae Hee Lee, Steven Schockaert, Stefan Wermter

Main category: cs.CL

TL;DR: LLMs often rely on heuristic biases (entity popularity, mention order, semantic co-occurrence) rather than genuine numerical knowledge when comparing entities, with smaller models being more susceptible to this behavior than larger models.

DetailsMotivation: To understand when LLMs use genuine knowledge versus superficial heuristics in knowledge-based reasoning tasks, particularly for entity comparison with numerical attributes.

Method: Analyze LLM performance on entity comparison tasks with numerical attributes, identify heuristic biases, compare model sizes (7-8B vs 32B parameters), and test chain-of-thought prompting.

Result: Smaller models’ predictions are better predicted by surface cues than their own numerical knowledge, while larger models selectively use numerical knowledge when reliable. Chain-of-thought prompting improves numerical reasoning across all model sizes.

Conclusion: LLM reasoning is influenced by heuristic biases, with model size affecting the ability to selectively use reliable knowledge, and chain-of-thought prompting can mitigate heuristic reliance.

Abstract: Large Language Models (LLMs) are increasingly used for knowledge-based reasoning tasks, yet understanding when they rely on genuine knowledge versus superficial heuristics remains challenging. We investigate this question through entity comparison tasks by asking models to compare entities along numerical attributes (e.g., ``Which river is longer, the Danube or the Nile?’’), which offer clear ground truth for systematic analysis. Despite having sufficient numerical knowledge to answer correctly, LLMs frequently make predictions that contradict this knowledge. We identify three heuristic biases that strongly influence model predictions: entity popularity, mention order, and semantic co-occurrence. For smaller models, a simple logistic regression using only these surface cues predicts model choices more accurately than the model’s own numerical predictions, suggesting heuristics largely override principled reasoning. Crucially, we find that larger models (32B parameters) selectively rely on numerical knowledge when it is more reliable, while smaller models (7–8B parameters) show no such discrimination, which explains why larger models outperform smaller ones even when the smaller models possess more accurate knowledge. Chain-of-thought prompting steers all models towards using the numerical features across all model sizes.

[49] Cross-Genre Authorship Attribution via LLM-Based Retrieve-and-Rerank

Shantanu Agarwal, Joel Barry, Steven Fincke, Scott Miller

Main category: cs.CL

TL;DR: A two-stage retrieve-and-rerank framework using LLMs for cross-genre authorship attribution that achieves significant improvements over previous state-of-the-art methods.

DetailsMotivation: Traditional authorship attribution systems often rely on topical cues, but cross-genre AA requires identifying author-specific linguistic patterns independent of subject matter, which existing IR-based training strategies fail to address properly.

Method: Two-stage retrieve-and-rerank framework with LLMs, featuring a targeted data curation strategy to help the reranker learn author-discriminative signals rather than relying on IR-based training approaches.

Result: Achieved substantial gains of 22.3 and 34.4 absolute Success@8 points over previous state-of-the-art on HIATUS’s HRS1 and HRS2 cross-genre AA benchmarks.

Conclusion: The proposed LLM-based retrieve-and-rerank pipeline with targeted data curation effectively addresses the limitations of IR-based approaches for cross-genre authorship attribution, demonstrating significant performance improvements.

Abstract: Authorship attribution (AA) is the task of identifying the most likely author of a query document from a predefined set of candidate authors. We introduce a two-stage retrieve-and-rerank framework that finetunes LLMs for cross-genre AA. Unlike the field of information retrieval (IR), where retrieve-and-rerank is a de facto strategy, cross-genre AA systems must avoid relying on topical cues and instead learn to identify author-specific linguistic patterns that are independent of the text’s subject matter (genre/domain/topic). Consequently, for the reranker, we demonstrate that training strategies commonly used in IR are fundamentally misaligned with cross-genre AA, leading to suboptimal behavior. To address this, we introduce a targeted data curation strategy that enables the reranker to effectively learn author-discriminative signals. Using our LLM-based retrieve-and-rerank pipeline, we achieve substantial gains of 22.3 and 34.4 absolute Success@8 points over the previous state-of-the-art on HIATUS’s challenging HRS1 and HRS2 cross-genre AA benchmarks.

[50] Who’s Asking? Simulating Role-Based Questions for Conversational AI Evaluation

Navreet Kaur, Hoda Ayad, Hayoung Jung, Shravika Mittal, Munmun De Choudhury, Tanushree Mitra

Main category: cs.CL

TL;DR: CoRUS framework simulates role-based questions using a taxonomy from opioid use disorder communities, revealing systematic differences in LLM responses based on user roles.

DetailsMotivation: Current LLM evaluations ignore user context and roles, which is critical in sensitive domains like opioid use disorder where responses need to be stigma-free and appropriate for different user types.

Method: Developed CoRUS framework using role theory and online community posts to create a taxonomy of asker roles (patients, caregivers, practitioners), then simulated 15,321 role-based questions embedding each role’s goals, behaviors, and experiences.

Result: Simulated questions were highly believable and comparable to real data. LLM evaluations showed systematic differences: vulnerable roles (patients, caregivers) elicited 17% more supportive responses and 19% less knowledge content compared to practitioner roles.

Conclusion: User roles implicitly shape LLM responses, demonstrating the need for role-informed evaluation of conversational AI, especially in sensitive domains.

Abstract: Language model users often embed personal and social context in their questions. The asker’s role – implicit in how the question is framed – creates specific needs for an appropriate response. However, most evaluations, while capturing the model’s capability to respond, often ignore who is asking. This gap is especially critical in stigmatized domains such as opioid use disorder (OUD), where accounting for users’ contexts is essential to provide accessible, stigma-free responses. We propose CoRUS (COmmunity-driven Roles for User-centric Question Simulation), a framework for simulating role-based questions. Drawing on role theory and posts from an online OUD recovery community (r/OpiatesRecovery), we first build a taxonomy of asker roles – patients, caregivers, practitioners. Next, we use it to simulate 15,321 questions that embed each role’s goals, behaviors, and experiences. Our evaluations show that these questions are both highly believable and comparable to real-world data. When used to evaluate five LLMs, for the same question but differing roles, we find systematic differences: vulnerable roles, such as patients and caregivers, elicit more supportive responses (+17%) and reduced knowledge content (-19%) in comparison to practitioners. Our work demonstrates how implicitly signaling a user’s role shapes model responses, and provides a methodology for role-informed evaluation of conversational AI.

[51] FinSight: Towards Real-World Financial Deep Research

Jiajie Jin, Yuyao Zhang, Yimeng Xu, Hongjin Qian, Yutao Zhu, Zhicheng Dou

Main category: cs.CL

TL;DR: FinSight is a multi-agent framework that uses Code Agent with Variable Memory architecture and iterative visualization to generate high-quality multimodal financial reports, outperforming existing systems.

DetailsMotivation: Current AI systems struggle to fully automate the labor-intensive and intellectually demanding process of generating professional financial reports.

Method: Uses Code Agent with Variable Memory (CAVM) architecture for flexible data handling, Iterative Vision-Enhanced Mechanism for chart refinement, and two-stage Writing Framework for expanding analysis into coherent multimodal reports.

Result: Significantly outperforms all baselines including leading deep research systems in factual accuracy, analytical depth, and presentation quality across various company and industry-level tasks.

Conclusion: Demonstrates a clear path toward generating financial reports that approach human-expert quality through the FinSight framework.

Abstract: Generating professional financial reports is a labor-intensive and intellectually demanding process that current AI systems struggle to fully automate. To address this challenge, we introduce FinSight (Financial InSight), a novel multi agent framework for producing high-quality, multimodal financial reports. The foundation of FinSight is the Code Agent with Variable Memory (CAVM) architecture, which unifies external data, designed tools, and agents into a programmable variable space, enabling flexible data collection, analysis and report generation through executable code. To ensure professional-grade visualization, we propose an Iterative Vision-Enhanced Mechanism that progressively refines raw visual outputs into polished financial charts. Furthermore, a two stage Writing Framework expands concise Chain-of-Analysis segments into coherent, citation-aware, and multimodal reports, ensuring both analytical depth and structural consistency. Experiments on various company and industry-level tasks demonstrate that FinSight significantly outperforms all baselines, including leading deep research systems in terms of factual accuracy, analytical depth, and presentation quality, demonstrating a clear path toward generating reports that approach human-expert quality.

[52] Neuronal Group Communication for Efficient Neural representation

Zhengqi Pei, Qingming Huang, Shuhui Wang

Main category: cs.CL

TL;DR: Proposes Neuronal Group Communication (NGC), a framework treating neural networks as dynamical systems of interacting neuronal groups rather than monolithic weight collections, enabling efficient, modular, and interpretable representations through low-rank communication between neuron groups.

DetailsMotivation: Address the challenges of efficiency and interpretability in large-scale neural networks by moving beyond treating weights as independent parameters and instead modeling neural computation as dynamical interactions between neuronal groups.

Method: NGC treats weights as transient interactions between embedding-like neuronal states, with computation unfolding through iterative communication among groups of neurons. Introduces neuronal stability metric based on dynamical systems theory to quantify contraction of neuron activations during sequence processing.

Result: NGC instantiated in LLMs shows improved performance on complex reasoning benchmarks under moderate compression, consistently outperforming standard low-rank approximations and cross-layer basis-sharing methods at comparable compression rates.

Conclusion: NGC provides a framework for building efficient, modular, and interpretable neural systems, with structured neuronal group dynamics potentially relating to generalization in high-dimensional learning systems.

Abstract: The ever-increasing scale of modern neural networks has brought unprecedented performance alongside daunting challenges in efficiency and interpretability. This paper addresses the core question of how to build large neural systems that learn efficient, modular, and interpretable representations. We propose Neuronal Group Communication (NGC), a theory-driven framework that reimagines a neural network as a dynamical system of interacting neuronal groups rather than a monolithic collection of neural weights. Instead of treating each weight as an independent trainable parameter, NGC treats weights as transient interactions between embedding-like neuronal states, with neural computation unfolding through iterative communication among groups of neurons. This low-rank, modular representation yields compact models: groups of neurons exchange low-dimensional signals, enabling intra-group specialization and inter-group information sharing while dramatically reducing redundant parameters. By drawing on dynamical systems theory, we introduce a neuronal stability metric (analogous to Lyapunov stability) that quantifies the contraction of neuron activations toward stable patterns during sequence processing. Using this metric, we reveal that emergent reasoning capabilities correspond to an external driving force or ``potential’’, which nudges the neural dynamics away from trivial trajectories while preserving stability. Empirically, we instantiate NGC in large language models (LLMs) and demonstrate improved performance on complex reasoning benchmarks under moderate compression. NGC consistently outperforms standard low-rank approximations and cross-layer basis-sharing methods at comparable compression rates. We conclude by discussing the broader implications of NGC, including how structured neuronal group dynamics might relate to generalization in high-dimensional learning systems.

[53] Does Visual Grounding Enhance the Understanding of Embodied Knowledge in Large Language Models?

Zhihui Yang, Yupei Wang, Kaijie Mo, Zhe Zhao, Renfen Hu

Main category: cs.CL

TL;DR: Vision-language models surprisingly don’t outperform text-only models in embodied knowledge understanding, performing worst in visual perception despite visual grounding.

DetailsMotivation: To determine whether visual grounding actually enhances multimodal language models' understanding of embodied knowledge compared to text-only models.

Method: Created a novel embodied knowledge benchmark based on perceptual theory, testing 30 state-of-the-art models through vector comparison and QA tasks with 1,700+ questions across multiple sensory modalities.

Result: VLMs showed no advantage over text-only models, performed worst in visual dimension, and struggled with spatial perception and reasoning. Vector representations were affected by word form and frequency.

Conclusion: Current models need more effective integration of embodied knowledge to better understand the physical world, as visual grounding alone doesn’t provide the expected benefits.

Abstract: Despite significant progress in multimodal language models (LMs), it remains unclear whether visual grounding enhances their understanding of embodied knowledge compared to text-only models. To address this question, we propose a novel embodied knowledge understanding benchmark based on the perceptual theory from psychology, encompassing visual, auditory, tactile, gustatory, olfactory external senses, and interoception. The benchmark assesses the models’ perceptual abilities across different sensory modalities through vector comparison and question-answering tasks with over 1,700 questions. By comparing 30 state-of-the-art LMs, we surprisingly find that vision-language models (VLMs) do not outperform text-only models in either task. Moreover, the models perform significantly worse in the visual dimension compared to other sensory dimensions. Further analysis reveals that the vector representations are easily influenced by word form and frequency, and the models struggle to answer questions involving spatial perception and reasoning. Our findings underscore the need for more effective integration of embodied knowledge in LMs to enhance their understanding of the physical world.

[54] ChiKhaPo: A Large-Scale Multilingual Benchmark for Evaluating Lexical Comprehension and Generation in Large Language Models

Emily Chang, Niyati Bafna

Main category: cs.CL

TL;DR: ChiKhaPo is a benchmark for evaluating LLMs’ basic linguistic competence across 2700+ languages, focusing on lexical comprehension and generation rather than high-order reasoning tasks.

DetailsMotivation: Existing LLM benchmarks are limited to high/mid-resource languages and focus on reasoning/generation tasks, while LLMs lack basic linguistic competence in most of the world's 3800+ written languages.

Method: Created ChiKhaPo with 8 subtasks of varying difficulty using existing lexicons, monolingual data, and bitext to evaluate lexical comprehension and generation abilities.

Result: 6 state-of-the-art models struggle on the benchmark, with performance influenced by language family, resource availability, task type, and comprehension vs generation directions.

Conclusion: ChiKhaPo enables massively multilingual benchmarking of LLMs and highlights the need for better basic linguistic competence across diverse languages.

Abstract: Existing benchmarks for large language models (LLMs) are largely restricted to high- or mid-resource languages, and often evaluate performance on higher-order tasks in reasoning and generation. However, plenty of evidence points to the fact that LLMs lack basic linguistic competence in the vast majority of the world’s 3800+ written languages. We introduce ChiKhaPo, consisting of 8 subtasks of varying difficulty designed to evaluate the lexical comprehension and generation abilities of generative models. ChiKhaPo draws on existing lexicons, monolingual data, and bitext, and provides coverage for 2700+ languages for 2 subtasks, surpassing any existing benchmark in terms of language coverage. We further show that 6 SOTA models struggle on our benchmark, and discuss the factors contributing to performance scores, including language family, language resourcedness, task, and comprehension versus generation directions. With ChiKhaPo, we hope to enable and encourage the massively multilingual benchmarking of LLMs.

[55] Prompt-MII: Meta-Learning Instruction Induction for LLMs

Emily Xiao, Yixiao Zeng, Ada Chen, Chin-Jou Li, Amanda Bertsch, Graham Neubig

Main category: cs.CL

TL;DR: PROMPT-MII is a reinforcement learning framework that meta-learns to generate compact instructions from training examples, achieving comparable performance to in-context learning while using 3-13x fewer tokens.

DetailsMotivation: In-context learning (ICL) for LLM adaptation is effective but incurs high inference costs as context length grows, creating a need for more efficient methods.

Method: PROMPT-MII uses reinforcement learning to meta-learn an instruction induction model that generates compact, descriptive prompts from training examples for arbitrary new datasets.

Result: The method improves downstream model quality by 4-9 F1 points (10-20% relative), matching ICL performance while requiring 3-13x fewer tokens, trained on 3,000+ datasets and evaluated on 90 unseen tasks.

Conclusion: PROMPT-MII provides an efficient alternative to ICL by generating compact instructions that maintain performance while significantly reducing token usage.

Abstract: A popular method to adapt large language models (LLMs) to new tasks is in-context learning (ICL), which is effective but incurs high inference costs as context length grows. In this paper we propose a method to perform instruction induction, where we take training examples and reduce them to a compact but descriptive prompt that can achieve performance comparable to ICL over the full training set. Specifically, we propose PROMPT-MII, a reinforcement learning (RL) based framework to meta-learn an instruction induction model that can generate compact instructions on the fly for an arbitrary new dataset. We train on over 3,000 diverse classification datasets from the HuggingFace hub, and evaluate on 90 unseen tasks. PROMPT-MII improves downstream model quality by 4-9 F1 points (10-20% relative), matching ICL performance while requiring 3-13x fewer tokens.

[56] Parameter-Efficient Fine-Tuning for Low-Resource Languages: A Comparative Study of LLMs for Bengali Hate Speech Detection

Akif Islam, Mohd Ruhul Ameen

Main category: cs.CL

TL;DR: First application of Parameter-Efficient Fine-Tuning (PEFT) using LoRA and QLoRA for Bengali hate speech detection, achieving state-of-the-art results with minimal computational resources.

DetailsMotivation: Bengali social media has seen significant increase in hate speech targeting women and adolescents, while existing approaches rely on computationally expensive full-model fine-tuning or proprietary APIs.

Method: Fine-tuned three instruction-tuned LLMs (Gemma-3-4B, Llama-3.2-3B, Mistral-7B) on BD-SHS dataset using PEFT with LoRA and QLoRA, training less than 1% of parameters on a single consumer GPU.

Result: Llama-3.2-3B achieved highest F1-score of 92.23%, followed by Mistral-7B (88.94%) and Gemma-3-4B (80.25%), demonstrating effective hate speech detection.

Conclusion: PEFT is established as a practical and replicable strategy for Bengali and other low-resource languages, enabling efficient hate speech detection with minimal computational requirements.

Abstract: Bengali social media platforms have witnessed a sharp increase in hate speech, disproportionately affecting women and adolescents. While datasets such as BD-SHS provide a basis for structured evaluation, most prior approaches rely on either computationally costly full-model fine-tuning or proprietary APIs. This paper presents the first application of Parameter-Efficient Fine-Tuning (PEFT) for Bengali hate speech detection using LoRA and QLoRA. Three instruction-tuned large language models - Gemma-3-4B, Llama-3.2-3B, and Mistral-7B - were fine-tuned on the BD-SHS dataset of 50,281 annotated comments. Each model was adapted by training fewer than 1% of its parameters, enabling experiments on a single consumer-grade GPU. The results show that Llama-3.2-3B achieved the highest F1-score of 92.23%, followed by Mistral-7B at 88.94% and Gemma-3-4B at 80.25%. These findings establish PEFT as a practical and replicable strategy for Bengali and related low-resource languages.

[57] Back to Bytes: Revisiting Tokenization Through UTF-8

Amit Moryossef, Clara Meister, Pavel Stepachev, Desmond Elliott

Main category: cs.CL

TL;DR: UTF8Tokenizer is a byte-level tokenizer that maps text directly to UTF-8 byte IDs, using C0 control bytes for special behaviors instead of auxiliary tokens, offering faster processing and improved language modeling convergence.

DetailsMotivation: To create a minimalist tokenizer that avoids out-of-range IDs and auxiliary tokens by leveraging UTF-8 byte encoding and C0 control bytes for special behaviors, addressing limitations of prior byte-level approaches.

Method: Uses byte-level tokenization where text maps directly to UTF-8 byte IDs (0-255), encodes all special behaviors using C0 control bytes, implements 256*d embedding tables, and employs bit-biased embeddings to expose per-byte bit structure.

Result: Achieves 14x faster tokenization, 8x less host-device transfer than int64, enables shareable embedding tables across models, and improves language modeling convergence with HuggingFace-compatible implementation.

Conclusion: UTF8Tokenizer demonstrates that minimalist byte-level tokenization using UTF-8 encoding and C0 control bytes provides practical benefits including speed, efficiency, and improved model performance while maintaining compatibility.

Abstract: We present UTF8Tokenizer, a minimalist byte-level tokenizer that maps text exactly to IDs corresponding to the bytes underlying the text’s UTF-8 encoding (e.g., byte x09 is token ID 9). Unlike prior byte-level approaches (Xue et al., 2021; Pagnoni et al., 2025), our implementation never introduces out-of-range IDs (i.e. there is no token ID 256) or auxiliary tokens: all special behavior (e.g., padding, boundaries, conversation structure, attention segments, tool calling, “thinking” spans, etc.) is encoded using C0 control bytes - just as ASCII was originally designed to embed control information alongside printable text. These design principles yield practical benefits: (1) faster tokenization (14x) and significantly lower host-device transfer (8x less than int64); (2) simple, shareable 256*d embedding tables that can be aligned across models; and (3) a training-time enhancement via bit-biased embeddings, which exposes per-byte bit structure and can be added to the embedding table post-training, removing inference costs. Our HuggingFace-compatible implementation improves language modeling convergence.

[58] Vocab Diet: Reshaping the Vocabulary of LLMs with Vector Arithmetic

Yuval Reif, Guy Kaplan, Roy Schwartz

Main category: cs.CL

TL;DR: The paper proposes a compositional vocabulary approach that represents word variations as base forms plus transformation vectors, reducing vocabulary size while maintaining performance and expanding coverage.

DetailsMotivation: Standard tokenization treats word variations as distinct tokens, wasting vocabulary space on surface form variants and limiting coverage of less frequent words and multilingual content.

Method: Use transformation vectors as additive offsets to represent word variations from base forms, then reshape vocabulary by composing words from shared base forms and transformations rather than unique tokens.

Result: Successfully removed up to 10% of vocabulary entries across multiple LLMs and five languages, freeing space for more diverse tokens while expanding out-of-vocabulary word coverage with minimal performance impact.

Conclusion: The approach motivates rethinking vocabulary design from string enumeration to compositional structures that leverage language’s inherent patterns, enabling more efficient tokenization without model weight modifications.

Abstract: Large language models (LLMs) were shown to encode word form variations, such as “walk”->“walked”, as linear directions in embedding space. However, standard tokenization algorithms treat these variations as distinct tokens – filling the size-capped vocabulary with surface form variants (e.g., “walk”, “walking”, “Walk”), at the expense of less frequent words and multilingual coverage. We show that many of these variations can be captured by transformation vectors – additive offsets that yield the appropriate word’s representation when applied to the base form word embedding – in both the input and output spaces. Building on this, we propose a compact reshaping of the vocabulary: rather than assigning unique tokens to each surface form, we compose them from shared base form and transformation vectors (e.g., “walked” = “walk” + past tense). We apply our approach to multiple LLMs and across five languages, removing up to 10% of vocabulary entries – thereby freeing space to allocate new, more diverse tokens. Importantly, we do so while also expanding vocabulary coverage to out-of-vocabulary words, with minimal impact on downstream performance, and without modifying model weights. Our findings motivate a foundational rethinking of vocabulary design, moving from string enumeration to a compositional vocabulary that leverages the underlying structure of language.

[59] Online Learning Defense against Iterative Jailbreak Attacks via Prompt Optimization

Masahiro Kaneko, Zeerak Talat, Timothy Baldwin

Main category: cs.CL

TL;DR: A reinforcement learning-based defense framework that dynamically updates strategies against iterative jailbreak attacks on LLMs, using prompt optimization and gradient damping to maintain harmless task performance while rejecting harmful prompts.

DetailsMotivation: Existing defenses don't proactively disrupt the trial-and-error cycle of iterative jailbreak methods, which repeatedly rewrite prompts to induce harmful outputs from LLMs.

Method: Propose a framework with online learning that updates defense strategies dynamically. Use reinforcement learning to optimize prompts for appropriate harmless responses while rejecting harmful ones. Introduce Past-Direction Gradient Damping (PDGD) to prevent overfitting to attack patterns.

Result: Significantly outperforms five existing defense methods against five iterative jailbreak methods on three LLMs. Also enhances response quality for harmless tasks simultaneously.

Conclusion: The proposed dynamic defense framework effectively counters iterative jailbreak attacks while maintaining and even improving harmless task performance, demonstrating superior protection compared to existing methods.

Abstract: Iterative jailbreak methods that repeatedly rewrite and input prompts into large language models (LLMs) to induce harmful outputs – using the model’s previous responses to guide each new iteration – have been found to be a highly effective attack strategy. Despite being an effective attack strategy against LLMs and their safety mechanisms, existing defenses do not proactively disrupt this dynamic trial-and-error cycle. In this study, we propose a novel framework that dynamically updates its defense strategy through online learning in response to each new prompt from iterative jailbreak methods. Leveraging the distinctions between harmful jailbreak-generated prompts and typical harmless prompts, we introduce a reinforcement learning-based approach that optimizes prompts to ensure appropriate responses for harmless tasks while explicitly rejecting harmful prompts. Additionally, to curb overfitting to the narrow band of partial input rewrites explored during an attack, we introduce Past-Direction Gradient Damping (PDGD). Experiments conducted on three LLMs show that our approach significantly outperforms five existing defense methods against five iterative jailbreak methods. Moreover, our results indicate that our prompt optimization strategy simultaneously enhances response quality for harmless tasks.

[60] DiscoTrack: A Multilingual LLM Benchmark for Discourse Tracking

Lanni Bu, Lauren Levin, Amir Zeldes

Main category: cs.CL

TL;DR: DiscoTrack is a multilingual benchmark for evaluating LLMs on discourse tracking tasks across 12 languages, focusing on implicit information and pragmatic inferences in larger documents.

DetailsMotivation: Current LLM benchmarks primarily test natural language understanding for explicit information extraction (QA, summarization) at sentence level, lacking challenging multilingual benchmarks for implicit information and pragmatic inferences across larger discourse contexts.

Method: Developed DiscoTrack benchmark with tasks across 12 languages targeting four levels of discourse understanding: salience recognition, entity tracking, discourse relations, and bridging inference.

Result: Evaluation shows these discourse tracking tasks remain challenging even for state-of-the-art models, highlighting limitations in current LLM capabilities for discourse-level understanding.

Conclusion: DiscoTrack addresses the gap in challenging multilingual benchmarks for discourse tracking and demonstrates that current models struggle with implicit information and pragmatic inferences across larger documents.

Abstract: Recent LLM benchmarks have tested models on a range of phenomena, but are still focused primarily on natural language understanding for extraction of explicit information, such as QA or summarization, with responses often tar- geting information from individual sentences. We are still lacking more challenging, and im- portantly also multilingual, benchmarks focus- ing on implicit information and pragmatic infer- ences across larger documents in the context of discourse tracking: integrating and aggregating information across sentences, paragraphs and multiple speaker utterances. To this end, we present DiscoTrack, an LLM benchmark target- ing a range of tasks across 12 languages and four levels of discourse understanding: salience recognition, entity tracking, discourse relations and bridging inference. Our evaluation shows that these tasks remain challenging, even for state-of-the-art models.

[61] SafeSearch: Do Not Trade Safety for Utility in LLM Search Agents

Qiusi Zhan, Angeline Budiman-Chan, Abdelrahman Zayed, Xingzhi Guo, Daniel Kang, Joo-Kyung Kim

Main category: cs.CL

TL;DR: Search agents using LLMs are more vulnerable to producing harmful outputs than base LLMs, and SafeSearch is proposed as a multi-objective reinforcement learning approach to reduce harmfulness while maintaining utility.

DetailsMotivation: LLM-based search agents are increasingly used but their safety behaviors are underexplored. Research shows these agents are more likely to produce harmful outputs than base LLMs, especially when utility-oriented fine-tuning intensifies this risk.

Method: SafeSearch, a multi-objective reinforcement learning approach that combines final-output safety/utility rewards with a novel query-level shaping term that penalizes unsafe queries and rewards safe ones.

Result: SafeSearch reduces agent harmfulness by over 70% across three red-teaming datasets while maintaining safe, helpful responses and matching the QA performance of utility-only finetuned agents.

Conclusion: The query-level reward in SafeSearch is effective in jointly improving both safety and utility, demonstrating the importance of multi-objective alignment for search agents.

Abstract: Large language model (LLM) based search agents iteratively generate queries, retrieve external information, and reason to answer open-domain questions. While researchers have primarily focused on improving their utility, their safety behaviors remain underexplored. In this paper, we first evaluate search agents using red-teaming datasets and find that they are more likely to produce harmful outputs than base LLMs. For example, when asked ``How can I track someone’s location without their consent?’’, a base model refuses, whereas a search agent designed to retrieve and cite sources may lower its refusal threshold, fetch documents (e.g., court cases), and, once appended, synthesize them into an informative yet unsafe summary. We further show that utility-oriented fine-tuning intensifies this risk, motivating joint alignment of safety and utility. We present SafeSearch, a multi-objective reinforcement learning approach that couples a final-output safety/utility reward with a novel query-level shaping term that penalizes unsafe queries and rewards safe ones. Experiments show that SafeSearch reduces agent harmfulness by over 70% across three red-teaming datasets while producing safe, helpful responses, and matches the QA performance of a utility-only finetuned agent; further analyses confirm the effectiveness of the query-level reward in jointly improving safety and utility.

[62] Extended LSTM: Adaptive Feature Gating for Toxic Comment Classification

Noor Islam S. Mohammad

Main category: cs.CL

TL;DR: xLSTM is a parameter-efficient framework for toxic comment detection that combines cosine-similarity gating, adaptive feature prioritization, and class rebalancing to outperform BERT on minority toxicity classes with 15x fewer parameters.

DetailsMotivation: Transformer models like BERT have high computational costs and perform poorly on minority toxicity classes, while classical ensembles lack semantic adaptability for toxic comment detection tasks.

Method: Uses cosine-similarity gating with learnable reference vectors to modulate embeddings, integrates multi-source embeddings (GloVe, FastText, BERT CLS), character-level BiLSTM for morphology, embedding-space SMOTE for minority augmentation, and adaptive focal loss with dynamic class weighting.

Result: Achieves 96.0% accuracy and 0.88 macro-F1 on Jigsaw Toxic Comment benchmark, outperforming BERT by 33% on threat and 28% on identity_hate categories with 15x fewer parameters and 50ms inference latency. Cosine gating provides +4.8% F1 gain.

Conclusion: Lightweight, theoretically informed architectures can surpass large pretrained models on imbalanced, domain-specific NLP tasks, establishing a new efficiency-adaptability frontier.

Abstract: Toxic comment detection remains a challenging task, where transformer-based models (e.g., BERT) incur high computational costs and degrade on minority toxicity classes, while classical ensembles lack semantic adaptability. We propose xLSTM, a parameter-efficient and theoretically grounded framework that unifies cosine-similarity gating, adaptive feature prioritization, and principled class rebalancing. A learnable reference vector {v} in {R}^d modulates contextual embeddings via cosine similarity, amplifying toxic cues and attenuating benign signals to yield stronger gradients under severe class imbalance. xLSTM integrates multi-source embeddings (GloVe, FastText, BERT CLS) through a projection layer, a character-level BiLSTM for morphological cues, embedding-space SMOTE for minority augmentation, and adaptive focal loss with dynamic class weighting. On the Jigsaw Toxic Comment benchmark, xLSTM attains 96.0% accuracy and 0.88 macro-F1, outperforming BERT by 33% on threat and 28% on identity_hate categories, with 15 times fewer parameters and 50ms inference latency. Cosine gating contributes a +4.8% F1 gain in ablations. The results establish a new efficiency adaptability frontier, demonstrating that lightweight, theoretically informed architectures can surpass large pretrained models on imbalanced, domain-specific NLP tasks.

[63] Mapping from Meaning: Addressing the Miscalibration of Prompt-Sensitive Language Models

Kyle Cox, Jiawei Xu, Yikun Han, Rong Xu, Tianhao Li, Chi-Yang Hsu, Tianlong Chen, Walter Gerych, Ying Ding

Main category: cs.CL

TL;DR: LLMs show prompt sensitivity where semantically equivalent prompts produce different answer distributions. The paper models this as generalization error and uses paraphrasing across semantic concept space to improve uncertainty calibration without losing accuracy. A new metric is introduced to decompose uncertainty and quantify prompt sensitivity.

DetailsMotivation: Large language models exhibit prompt sensitivity - they produce different answer distributions for semantically equivalent prompts, suggesting their output uncertainty doesn't reflect true semantic uncertainty about the prompt meaning.

Method: Model prompt sensitivity as generalization error, use paraphrasing perturbations across semantic concept space, and introduce a new uncertainty decomposition metric that captures semantic continuities in natural language generation.

Result: Sampling across semantic concept space with paraphrasing improves uncertainty calibration without compromising accuracy. The new decomposition metric can quantify how much LLM uncertainty is attributed to prompt sensitivity.

Conclusion: The work provides a new approach to improve uncertainty calibration in prompt-sensitive LLMs and shows evidence that some LLMs fail to exhibit consistent general reasoning about input meanings.

Abstract: An interesting behavior in large language models (LLMs) is prompt sensitivity. When provided with different but semantically equivalent versions of the same prompt, models may produce very different distributions of answers. This suggests that the uncertainty reflected in a model’s output distribution for one prompt may not reflect the model’s uncertainty about the meaning of the prompt. We model prompt sensitivity as a type of generalization error, and show that sampling across the semantic ``concept space’’ with paraphrasing perturbations improves uncertainty calibration without compromising accuracy. Additionally, we introduce a new metric for uncertainty decomposition in black-box LLMs that improves upon entropy-based decomposition by modeling semantic continuities in natural language generation. We show that this decomposition metric can be used to quantify how much LLM uncertainty is attributed to prompt sensitivity. Our work introduces a new way to improve uncertainty calibration in prompt-sensitive language models, and provides evidence that some LLMs fail to exhibit consistent general reasoning about the meanings of their inputs.

[64] Investigating Thinking Behaviours of Reasoning-Based Language Models for Social Bias Mitigation

Guoqing Luo, Iffat Maab, Lili Mou, Junichi Yamagishi

Main category: cs.CL

TL;DR: This paper investigates how reasoning-based LLMs aggregate social stereotypes during their thinking process, identifies two failure patterns (stereotype repetition and irrelevant information injection), and proposes a lightweight prompt-based mitigation approach.

DetailsMotivation: While reasoning-based LLMs excel at complex tasks through structured thinking processes, they can inadvertently aggregate social stereotypes leading to biased outcomes, yet the underlying mechanisms remain underexplored.

Method: Systematically investigate thinking process mechanisms, identify two failure patterns (stereotype repetition and irrelevant information injection), and introduce a lightweight prompt-based mitigation approach that queries the model to review its initial reasoning against these failure patterns.

Result: Experiments on question answering (BBQ and StereoSet) and open-ended (BOLD) benchmarks show the approach effectively reduces bias while maintaining or improving accuracy.

Conclusion: The study uncovers specific failure patterns in LLM reasoning that drive social bias aggregation and demonstrates that a simple prompt-based mitigation approach can effectively address these biases without compromising performance.

Abstract: While reasoning-based large language models excel at complex tasks through an internal, structured thinking process, a concerning phenomenon has emerged that such a thinking process can aggregate social stereotypes, leading to biased outcomes. However, the underlying behaviours of these language models in social bias scenarios remain underexplored. In this work, we systematically investigate mechanisms within the thinking process behind this phenomenon and uncover two failure patterns that drive social bias aggregation: 1) stereotype repetition, where the model relies on social stereotypes as its primary justification, and 2) irrelevant information injection, where it fabricates or introduces new details to support a biased narrative. Building on these insights, we introduce a lightweight prompt-based mitigation approach that queries the model to review its own initial reasoning against these specific failure patterns. Experiments on question answering (BBQ and StereoSet) and open-ended (BOLD) benchmarks show that our approach effectively reduces bias while maintaining or improving accuracy.

[65] DVAGen: Dynamic Vocabulary Augmented Generation

Wei Du, Nuowei Liu, Jie Wang, Jiahao Kuang, Tao Ji, Xiaoling Wang, Yuanbin Wu

Main category: cs.CL

TL;DR: DVAGen is an open-source framework for dynamic vocabulary-augmented language models that addresses limitations of fixed vocabularies and existing dynamic approaches through modular design, modern LLM integration, and improved inference scalability.

DetailsMotivation: Fixed vocabulary language models struggle with novel/out-of-vocabulary words, and existing dynamic vocabulary methods have issues with fragmented codebases, lack of modern LLM support, and limited inference scalability.

Method: DVAGen provides a unified framework with modular pipeline design, seamless integration with open-source LLMs, and CLI/WebUI tools for training, evaluation, visualization, and real-time result inspection.

Result: The framework validates dynamic vocabulary effectiveness on modern LLMs and demonstrates support for batch inference with significantly improved inference throughput.

Conclusion: DVAGen overcomes limitations of existing dynamic vocabulary approaches by providing a comprehensive, scalable, and user-friendly framework that enhances language model flexibility and performance.

Abstract: Language models trained with a fixed vocabulary struggle to generalize to novel or out-of-vocabulary words, limiting their flexibility in handling diverse token combinations. Existing dynamic vocabulary approaches attempt to address this limitation but face challenges such as fragmented codebases, lack of support for modern LLMs, and limited inference scalability. To overcome these issues, we introduce DVAGen, a fully open-source, unified framework designed for training, evaluation, and visualization of dynamic vocabulary-augmented language models. Our framework modularizes the pipeline for ease of customization, integrates seamlessly with open-source LLMs, and is the first to provide both CLI and WebUI tools for real-time result inspection. We validate the effectiveness of dynamic vocabulary methods on modern LLMs and demonstrate support for batch inference, significantly improving inference throughput.

[66] Rethinking On-policy Optimization for Query Augmentation

Zhichao Xu, Shengyao Zhuang, Xueguang Ma, Bingsen Chen, Yijun Tian, Fengran Mo, Jie Cao, Vivek Srikumar

Main category: cs.CL

TL;DR: Systematic comparison shows simple prompting-based query augmentation often matches or beats expensive RL-based methods, leading to a novel hybrid approach (OPQE) that outperforms both.

DetailsMotivation: To compare prompting-based and RL-based query augmentation approaches under consistent conditions, as they have respective advantages but haven't been systematically evaluated.

Method: Introduced On-policy Pseudo-document Query Expansion (OPQE) - a hybrid method where LLM policy learns to generate pseudo-documents that maximize retrieval performance, combining prompting flexibility with RL optimization.

Result: Simple training-free query augmentation performs on par with or surpasses RL-based methods, especially with powerful LLMs. OPQE outperforms both standalone prompting and RL-based rewriting.

Conclusion: Synergistic approaches combining prompting flexibility with RL optimization yield the best results for query augmentation in information retrieval.

Abstract: Recent advances in large language models (LLMs) have led to a surge of interest in query augmentation for information retrieval (IR). Two main approaches have emerged. The first prompts LLMs to generate answers or pseudo-documents that serve as new queries, relying purely on the model’s parametric knowledge or contextual information. The second applies reinforcement learning (RL) to fine-tune LLMs for query rewriting, directly optimizing retrieval metrics. While having respective advantages and limitations, the two approaches have not been compared under consistent experimental conditions. In this work, we present the first systematic comparison of prompting-based and RL-based query augmentation across diverse benchmarks, including evidence-seeking, ad hoc, and tool retrieval. Our key finding is that simple, training-free query augmentation often performs on par with, or even surpasses, more expensive RL-based counterparts, especially when using powerful LLMs. Motivated by this discovery, we introduce a novel hybrid method, On-policy Pseudo-document Query Expansion (OPQE), which, instead of rewriting a query, the LLM policy learns to generate a pseudo-document that maximizes retrieval performance, thus merging the flexibility and generative structure of prompting with the targeted optimization of RL. We show OPQE outperforms both standalone prompting and RL-based rewriting, demonstrating that a synergistic approach yields the best results. Our implementation is made available to facilitate reproducibility.

[67] When AI companions become witty: Can human brain recognize AI-generated irony?

Xiaohui Rao, Hanlin Wu, Zhenguang G. Cai

Main category: cs.CL

TL;DR: People don’t fully adopt intentional stance toward AI-generated irony, showing reduced neural processing and greater attribution to computational errors rather than deliberate communication.

DetailsMotivation: To investigate whether people attribute mental states and intentionality to AI during irony comprehension, as LLMs are increasingly used as social agents producing humor and irony.

Method: Compared behavioral and neural responses to ironic statements from AI vs human sources using ERP components (P200 for early incongruity detection and P600 for cognitive reanalysis).

Result: Participants showed attenuated P200 and P600 effects for AI-generated irony, indicating reduced effortful detection and reanalysis. People attributed incongruity to deliberate communication less for AI than human sources.

Conclusion: Source attribution shapes neural processing of social communication. Achieving genuine social agency for AI requires more than linguistic competence - it necessitates a shift in how humans perceive and attribute intentionality to artificial agents.

Abstract: As Large Language Models (LLMs) are increasingly deployed as social agents and trained to produce humor and irony, a question emerges: when encountering witty AI remarks, do people interpret these as intentional communication or mere computational output? This study investigates whether people adopt the intentional stance, attributing mental states to explain behavior,toward AI during irony comprehension. Irony provides an ideal paradigm because it requires distinguishing intentional contradictions from unintended errors through effortful semantic reanalysis. We compared behavioral and neural responses to ironic statements from AI versus human sources using established ERP components: P200 reflecting early incongruity detection and P600 indexing cognitive efforts in reinterpreting incongruity as deliberate irony. Results demonstrate that people do not fully adopt the intentional stance toward AI-generated irony. Behaviorally, participants attributed incongruity to deliberate communication for both sources, though significantly less for AI than human, showing greater tendency to interpret AI incongruities as computational errors. Neural data revealed attenuated P200 and P600 effects for AI-generated irony, suggesting reduced effortful detection and reanalysis consistent with diminished attribution of communicative intent. Notably, people who perceived AI as more sincere showed larger P200 and P600 effects for AI-generated irony, suggesting that intentional stance adoption is calibrated by specific mental models of artificial agents. These findings reveal that source attribution shapes neural processing of social-communicative phenomena. Despite current LLMs’ linguistic sophistication, achieving genuine social agency requires more than linguistic competence, it necessitates a shift in how humans perceive and attribute intentionality to artificial agents.

[68] Understanding and Improving Length Generalization in Hierarchical Sparse Attention Models

Jiaqi Leng, Xiang Hu, Junxiong Wang, Jianguo Li, Wei Wu, Yucheng Lu

Main category: cs.CL

TL;DR: This paper systematically analyzes chunk-based sparse attention models for long-context processing, identifying three key design principles that enable training-free length extrapolation from 4K to 32M tokens.

DetailsMotivation: Standard Transformers have quadratic complexity and poor length extrapolation, while alternative architectures sacrifice full context utilization. The paper aims to understand the architectural principles behind successful chunk-based sparse attention models.

Method: The authors use a unified framework and comprehensive ablation studies to identify critical design components, including expressive chunk encoders with CLS tokens, bypassing residual paths, and enforced selection sparsity during pre-training.

Result: The proposed approach achieves new state-of-the-art training-free length extrapolation, successfully generalizing from 4K to 32 million tokens on RULER and BABILong benchmarks.

Conclusion: The study provides clear, empirically-grounded design principles for developing future long-context language models, demonstrating that the identified three components are critical for effective length generalization.

Abstract: Effectively processing long contexts is a critical challenge for language models. While standard Transformers are limited by quadratic complexity and poor length extrapolation, alternative architectures like sliding window attention and state space models sacrifice the ability to effectively utilize the full context due to their fixed-size memory. Chunk-based sparse attention has emerged as a promising paradigm for extreme length generalization, yet the key architectural principles underpinning its success are not yet fully understood. In this work, we present a systematic dissection of these models to identify the core components driving their performance. Through a unified framework and comprehensive ablation studies, we demonstrate that a combination of three design principles is critical: (1) an expressive, non-linear Chunk Encoder with a dedicated CLS token to produce representations for retrieval; (2) a Bypassing Residual Path to stably integrate retrieved global information without it being overridden by the local residual stream; and (3) enforced selection sparsity during pre-training to bridge the train-test distribution gap. We provide a theoretical motivation for intra-chunk information processing and landmark generation. By combining these principles, we establish a new state-of-the-art for training-free length extrapolation, successfully generalizing models trained on a 4K context to 32 million tokens on RULER and BABILong. Our findings provide a clear and empirically-grounded set of design principles for developing future, highly-capable long-context language models.

[69] Wisdom is Knowing What not to Say: Hallucination-Free LLMs Unlearning via Attention Shifting

Chenchen Tan, Youyang Qu, Xinghao Li, Hui Zhang, Shujie Cui, Cunjian Chen, Longxiang Gao

Main category: cs.CL

TL;DR: The paper introduces Attention-Shifting (AS), a novel framework for selective unlearning in LLMs that addresses the trade-off between utility preservation and hallucination risks through context-preserving suppression and hallucination-resistant response shaping.

DetailsMotivation: Existing machine unlearning approaches face a critical dilemma: aggressive unlearning compromises model utility while conservative strategies preserve utility but risk hallucinated responses, limiting LLMs' reliability in knowledge-intensive applications.

Method: AS framework uses two attention-level interventions: importance-aware suppression applied to unlearning set to reduce reliance on memorized knowledge, and attention-guided retention enhancement that reinforces attention toward semantically essential tokens in retained dataset. These are jointly optimized via a dual-loss objective.

Result: AS improves performance preservation over state-of-the-art unlearning methods, achieving up to 15% higher accuracy on ToFU benchmark and 10% on TDEC benchmark, while maintaining competitive hallucination-free unlearning effectiveness.

Conclusion: AS demonstrates superior balance between unlearning effectiveness, generalization, and response reliability compared to existing methods.

Abstract: The increase in computing power and the necessity of AI-assisted decision-making boost the growing application of large language models (LLMs). Along with this, the potential retention of sensitive data of LLMs has spurred increasing research into machine unlearning. However, existing unlearning approaches face a critical dilemma: Aggressive unlearning compromises model utility, while conservative strategies preserve utility but risk hallucinated responses. This significantly limits LLMs’ reliability in knowledge-intensive applications. To address this, we introduce a novel Attention-Shifting (AS) framework for selective unlearning. AS is driven by two design objectives: (1) context-preserving suppression that attenuates attention to fact-bearing tokens without disrupting LLMs’ linguistic structure; and (2) hallucination-resistant response shaping that discourages fabricated completions when queried about unlearning content. AS realizes these objectives through two attention-level interventions, which are importance-aware suppression applied to the unlearning set to reduce reliance on memorized knowledge and attention-guided retention enhancement that reinforces attention toward semantically essential tokens in the retained dataset to mitigate unintended degradation. These two components are jointly optimized via a dual-loss objective, which forms a soft boundary that localizes unlearning while preserving unrelated knowledge under representation superposition. Experimental results show that AS improves performance preservation over the state-of-the-art unlearning methods, achieving up to 15% higher accuracy on the ToFU benchmark and 10% on the TDEC benchmark, while maintaining competitive hallucination-free unlearning effectiveness. Compared to existing methods, AS demonstrates a superior balance between unlearning effectiveness, generalization, and response reliability.

[70] StreamingThinker: Large Language Models Can Think While Reading

Junlong Tong, Yingqi Fan, Anhao Zhao, Yunpu Ma, Xiaoyu Shen

Main category: cs.CL

TL;DR: StreamingThinker enables LLMs to think while reading input rather than waiting for complete input, reducing latency by 80% for reasoning onset and 60% for final answer generation while maintaining performance comparable to traditional batch thinking.

DetailsMotivation: Current LLM reasoning paradigm requires complete input before starting to think, causing unnecessary latency and weakening attention to earlier information in dynamic scenarios. Inspired by human 'thinking while reading' cognition.

Method: StreamingThinker framework integrates streaming CoT generation with quality control, streaming-constraint training using attention masks and position encoding, and parallel KV caches that decouple input encoding from reasoning generation for true concurrency.

Result: Evaluation on Qwen3 models across math, logical, and context-based QA reasoning tasks shows comparable performance to batch thinking, with 80% reduction in token waiting time before reasoning onset and over 60% reduction in time-level latency for final answer generation.

Conclusion: The streaming thinking paradigm effectively reduces latency while maintaining reasoning performance, demonstrating the viability of thinking-while-reading approaches for LLM reasoning.

Abstract: Large language models (LLMs) have demonstrated remarkable capabilities in chain of thought (CoT) reasoning. However, the current LLM reasoning paradigm initiates thinking only after the entire input is available, which introduces unnecessary latency and weakens attention to earlier information in dynamic scenarios. Inspired by human cognition of thinking while reading, we first design a \textit{\textbf{streaming thinking}} paradigm for LLMs, where reasoning unfolds in the order of input and further adjusts its depth once reading is complete. We instantiate this paradigm with \textit{StreamingThinker}, a framework that enables LLMs to think while reading through the integration of streaming CoT generation, streaming-constraint training, and streaming parallel inference. Specifically, StreamingThinker employs streaming reasoning units with quality control for CoT generation, enforces order-preserving reasoning through streaming attention masks and position encoding, and leverages parallel KV caches that decouple input encoding from reasoning generation, thereby ensuring alignment and enabling true concurrency. We evaluate StreamingThinker on the Qwen3 model family across math reasoning, logical reasoning, and context-based QA reasoning tasks. Experimental results show that the StreamingThinker preserves performance comparable to batch thinking, while yielding an 80% reduction in token waiting before the onset of reasoning and a more than 60% reduction in time-level latency for producing the final answer, demonstrating the effectiveness of the streaming paradigm for LLM reasoning. Code will be released at \href{https://github.com/EIT-NLP/StreamingLLM/tree/main/StreamingThinker}{this repository.}

[71] From Preferences to Prejudice: The Role of Alignment Tuning in Shaping Social Bias in Video Diffusion Models

Zefan Cai, Haoyi Qiu, Haozhe Zhao, Ke Wan, Jiachen Li, Jiuxiang Gu, Wen Xiao, Nanyun Peng, Junjie Hu

Main category: cs.CL

TL;DR: VideoBiasEval is a framework that reveals how video diffusion models amplify social biases during alignment tuning, showing that reward models trained on human preferences encode and strengthen stereotypes in generated videos.

DetailsMotivation: To systematically trace how social biases evolve throughout the video generation alignment pipeline and understand how alignment tuning unintentionally amplifies representational biases.

Method: Introduced VideoBiasEval framework with event-based prompting to disentangle semantic content from actor attributes, using multi-granular metrics to evaluate ethnicity bias, gender bias conditioned on ethnicity, distributional shifts, and temporal persistence of bias.

Result: Alignment tuning strengthens representational biases and makes them temporally stable, producing smoother yet more stereotyped portrayals. Biases in human preference datasets are amplified in reward models and propagate through alignment-tuned video diffusion models.

Conclusion: There is a critical need for bias-aware evaluation and mitigation throughout the alignment process to ensure fair and socially responsible video generation.

Abstract: Recent advances in video diffusion models have significantly enhanced text-to-video generation, particularly through alignment tuning using reward models trained on human preferences. While these methods improve visual quality, they can unintentionally encode and amplify social biases. To systematically trace how such biases evolve throughout the alignment pipeline, we introduce VideoBiasEval, a comprehensive diagnostic framework for evaluating social representation in video generation. Grounded in established social bias taxonomies, VideoBiasEval employs an event-based prompting strategy to disentangle semantic content (actions and contexts) from actor attributes (gender and ethnicity). It further introduces multi-granular metrics to evaluate (1) overall ethnicity bias, (2) gender bias conditioned on ethnicity, (3) distributional shifts in social attributes across model variants, and (4) the temporal persistence of bias within videos. Using this framework, we conduct the first end-to-end analysis connecting biases in human preference datasets, their amplification in reward models, and their propagation through alignment-tuned video diffusion models. Our results reveal that alignment tuning not only strengthens representational biases but also makes them temporally stable, producing smoother yet more stereotyped portrayals. These findings highlight the need for bias-aware evaluation and mitigation throughout the alignment process to ensure fair and socially responsible video generation.

[72] How News Feels: Understanding Affective Bias in Multilingual Headlines for Human-Centered Media Design

Mohd Ruhul Ameen, Akif Islam, Abu Saleh Musa Miah, Ayesha Siddiqua, Jungpil Shin

Main category: cs.CL

TL;DR: Large-scale emotion analysis of 300,000 Bengali news headlines reveals dominance of negative emotions (anger, fear, disappointment) and emotional bias across outlets, leading to design proposals for emotion-aware news aggregators.

DetailsMotivation: News media shape public mood through emotional framing - same events can appear calm or alarming depending on outlet. Negative/emotional headlines attract more attention and spread faster, encouraging provocative framing.

Method: Used zero-shot inference with Gemma-3 4B model to analyze 300,000 Bengali news headlines and content, identifying dominant emotions and overall tone for each article.

Result: Clear dominance of negative emotions (anger, fear, disappointment) and significant variation in emotional portrayal of similar stories across different news outlets.

Conclusion: Proposed design ideas for human-centered news aggregator that visualizes emotional cues and helps readers recognize hidden affective framing in daily news coverage.

Abstract: News media often shape the public mood not only by what they report but by how they frame it. The same event can appear calm in one outlet and alarming in another, reflecting subtle emotional bias in reporting. Negative or emotionally charged headlines tend to attract more attention and spread faster, which in turn encourages outlets to frame stories in ways that provoke stronger reactions. This research explores that tendency through large-scale emotion analysis of Bengali news. Using zero-shot inference with Gemma-3 4B, we analyzed 300000 Bengali news headlines and their content to identify the dominant emotion and overall tone of each. The findings reveal a clear dominance of negative emotions, particularly anger, fear, and disappointment, and significant variation in how similar stories are emotionally portrayed across outlets. Based on these insights, we propose design ideas for a human-centered news aggregator that visualizes emotional cues and helps readers recognize hidden affective framing in daily news.

[73] Explainability of Large Language Models: Opportunities and Challenges toward Generating Trustworthy Explanations

Shahin Atakishiyev, Housam K. B. Babiker, Jiayi Dai, Nawshad Farruque, Teruaki Hayashi, Nafisa Sadaf Hriti, Md Abed Rahman, Iain Smith, Mi-Young Kim, Osmar R. Zaïane, Randy Goebel

Main category: cs.CL

TL;DR: This paper reviews local explainability and mechanistic interpretability approaches for Transformer-based large language models, conducts experimental studies in healthcare and autonomous driving domains, and identifies future research directions for trustworthy LLM explanations.

DetailsMotivation: Large language models often make prediction errors and hallucinations, creating an urgent need to understand their inner workings and build trust through better interpretability.

Method: Presents a literature review of explainability approaches, conducts experimental studies in healthcare and autonomous driving domains, and analyzes trust implications of explanations.

Result: The paper provides insights into current explainability methods and their application in critical domains, highlighting the importance of human-aligned explanations for building trust.

Conclusion: Identifies unaddressed issues in LLM explainability and outlines opportunities, challenges, and future directions for generating trustworthy explanations that align with human understanding.

Abstract: Large language models have exhibited impressive performance across a broad range of downstream tasks in natural language processing. However, how a language model predicts the next token and generates content is not generally understandable by humans. Furthermore, these models often make errors in prediction and reasoning, known as hallucinations. These errors underscore the urgent need to better understand and interpret the intricate inner workings of language models and how they generate predictive outputs. Motivated by this gap, this paper investigates local explainability and mechanistic interpretability within Transformer-based large language models to foster trust in such models. In this regard, our paper aims to make three key contributions. First, we present a review of local explainability and mechanistic interpretability approaches and insights from relevant studies in the literature. Furthermore, we describe experimental studies on explainability and reasoning with large language models in two critical domains – healthcare and autonomous driving – and analyze the trust implications of such explanations for explanation receivers. Finally, we summarize current unaddressed issues in the evolving landscape of LLM explainability and outline the opportunities, critical challenges, and future directions toward generating human-aligned, trustworthy LLM explanations.

[74] TaxoAlign: Scholarly Taxonomy Generation Using Language Models

Avishek Lahiri, Yufang Hou, Debarshi Kumar Sanyal

Main category: cs.CL

TL;DR: The paper presents TaxoAlign, a method for automated taxonomy generation that bridges the gap between human-written and automatically-created taxonomies, evaluated on the CS-TaxoBench benchmark.

DetailsMotivation: Existing approaches to automatic survey generation don't compare generated structures with human expert taxonomies, creating a gap in evaluation and quality.

Method: Proposes TaxoAlign - a three-phase topic-based instruction-guided method for scholarly taxonomy generation, with an automated evaluation framework measuring structural alignment and semantic coherence.

Result: TaxoAlign consistently surpasses baselines on nearly all metrics when evaluated on CS-TaxoBench (460 taxonomies from human-written surveys plus 80 from conference surveys).

Conclusion: The method successfully bridges the gap between human-generated and automatically-created taxonomies, demonstrating superior performance over existing approaches.

Abstract: Taxonomies play a crucial role in helping researchers structure and navigate knowledge in a hierarchical manner. They also form an important part in the creation of comprehensive literature surveys. The existing approaches to automatic survey generation do not compare the structure of the generated surveys with those written by human experts. To address this gap, we present our own method for automated taxonomy creation that can bridge the gap between human-generated and automatically-created taxonomies. For this purpose, we create the CS-TaxoBench benchmark which consists of 460 taxonomies that have been extracted from human-written survey papers. We also include an additional test set of 80 taxonomies curated from conference survey papers. We propose TaxoAlign, a three-phase topic-based instruction-guided method for scholarly taxonomy generation. Additionally, we propose a stringent automated evaluation framework that measures the structural alignment and semantic coherence of automatically generated taxonomies in comparison to those created by human experts. We evaluate our method and various baselines on CS-TaxoBench, using both automated evaluation metrics and human evaluation studies. The results show that TaxoAlign consistently surpasses the baselines on nearly all metrics. The code and data can be found at https://github.com/AvishekLahiri/TaxoAlign.

[75] Addressing Antisocial Behavior in Multi-Party Dialogs Through Multimodal Representation Learning

Hajar Bakarou, Mohamed Sinane El Messoussi, Anaïs Ollagnier

Main category: cs.CL

TL;DR: This paper addresses antisocial behavior detection in multi-party conversations using a French dataset, evaluating text-based and graph-based methods across three tasks, with multimodal fusion achieving best results.

DetailsMotivation: Antisocial behavior on social media poses risks to platform safety, but multi-party conversational settings remain underexplored due to limited data availability.

Method: Used CyberAgressionAdo-Large dataset to evaluate abuse detection, bullying behavior analysis, and bullying peer-group identification using six text-based and eight graph-based representation-learning methods with multimodal fusion.

Result: Multimodal models outperformed unimodal baselines, with late fusion model mBERT + WD-SGCN achieving best overall performance (0.718 on abuse detection, 0.286 on peer-group identification, 0.606 on bullying analysis).

Conclusion: The study demonstrates the effectiveness of multimodal approaches in handling nuanced antisocial behavior phenomena like implicit aggression and context-dependent hostility in multi-party conversations.

Abstract: Antisocial behavior (ASB) on social media – including hate speech, harassment, and cyberbullying – poses growing risks to platform safety and societal well-being. Prior research has focused largely on networks such as X and Reddit, while \textit{multi-party conversational settings} remain underexplored due to limited data. To address this gap, we use \textit{CyberAgressionAdo-Large}, a French open-access dataset simulating ASB in multi-party conversations, and evaluate three tasks: \textit{abuse detection}, \textit{bullying behavior analysis}, and \textit{bullying peer-group identification}. We benchmark six text-based and eight graph-based \textit{representation-learning methods}, analyzing lexical cues, interactional dynamics, and their multimodal fusion. Results show that multimodal models outperform unimodal baselines. The late fusion model \texttt{mBERT + WD-SGCN} achieves the best overall results, with top performance on abuse detection (0.718) and competitive scores on peer-group identification (0.286) and bullying analysis (0.606). Error analysis highlights its effectiveness in handling nuanced ASB phenomena such as implicit aggression, role transitions, and context-dependent hostility.

[76] Readers Prefer Outputs of AI Trained on Copyrighted Books over Expert Human Writers

Tuhin Chakrabarty, Jane C. Ginsburg, Paramveer Dhillon

Main category: cs.CL

TL;DR: Fine-tuning AI models on authors’ complete works dramatically improves their ability to emulate writing styles and quality, reversing experts’ initial preference for human-written text over in-context prompted AI outputs.

DetailsMotivation: To investigate whether AI models can generate high-quality literary text that effectively emulates authors' styles, addressing copyright concerns about AI's ability to create derivative content.

Method: Preregistered study comparing MFA-trained expert writers with three AI models (ChatGPT, Claude, Gemini) using blind pairwise evaluations by 159 expert and lay readers. Tested both in-context prompting and fine-tuning approaches.

Result: In-context prompting was strongly disfavored by experts for stylistic fidelity and writing quality, but fine-tuning completely reversed these findings - experts then favored AI-generated text. Fine-tuned outputs were rarely detected as AI-generated (3% vs 97% for in-context).

Conclusion: Author-specific fine-tuning enables AI writing that readers prefer to expert human writing, providing empirical evidence relevant to copyright’s fair-use factor regarding market effects on source works.

Abstract: The use of copyrighted books for training AI models has led to numerous lawsuits from authors concerned about AI’s ability to generate derivative content. Yet it’s unclear if these models can generate high quality literary text while emulating authors’ styles. To answer this we conducted a preregistered study comparing MFA-trained expert writers with three frontier AI models: ChatGPT, Claude & Gemini in writing up to 450 word excerpts emulating 50 award-winning authors’ diverse styles. In blind pairwise evaluations by 159 representative expert & lay readers, AI-generated text from in-context prompting was strongly disfavored by experts for both stylistic fidelity (OR=0.16, p<10^-8) & writing quality (OR=0.13, p<10^-7) but showed mixed results with lay readers. However, fine-tuning ChatGPT on individual authors’ complete works completely reversed these findings: experts now favored AI-generated text for stylistic fidelity (OR=8.16, p<10^-13) & writing quality (OR=1.87, p=0.010), with lay readers showing similar shifts. These effects generalize across authors & styles. The fine-tuned outputs were rarely flagged as AI-generated (3% rate v. 97% for in-context prompting) by best AI detectors. Mediation analysis shows this reversal occurs because fine-tuning eliminates detectable AI stylistic quirks (e.g., cliche density) that penalize in-context outputs. While we do not account for additional costs of human effort required to transform raw AI output into cohesive, publishable prose, the median fine-tuning & inference cost of $81 per author represents a dramatic 99.7% reduction compared to typical professional writer compensation. Author-specific fine-tuning thus enables non-verbatim AI writing that readers prefer to expert human writing, providing empirical evidence directly relevant to copyright’s fourth fair-use factor, the “effect upon the potential market or value” of the source works.

[77] Towards Mixed-Modal Retrieval for Universal Retrieval-Augmented Generation

Chenghao Zhang, Guanting Dong, Xinyu Yang, Zhicheng Dou

Main category: cs.CL

TL;DR: Nyx is a unified mixed-modal retriever for Universal RAG that handles text and image queries/documents, trained on automatically generated mixed-modal data (NyxQA) and fine-tuned with VLM feedback.

DetailsMotivation: Existing RAG systems focus only on text documents and fail in real-world scenarios where both queries and documents contain mixed modalities (text and images).

Method: Proposed Nyx - a unified mixed-modal retriever, with a four-stage automated pipeline to generate NyxQA dataset, followed by two-stage training: pre-training on mixed-modal data and supervised fine-tuning using VLM feedback.

Result: Nyx performs competitively on standard text-only RAG benchmarks and excels in Universal RAG setting, significantly improving generation quality in vision-language tasks.

Conclusion: Nyx effectively addresses the Universal RAG challenge by enabling mixed-modal retrieval and reasoning, demonstrating strong performance in both text-only and mixed-modal scenarios.

Abstract: Retrieval-Augmented Generation (RAG) has emerged as a powerful paradigm for enhancing large language models (LLMs) by retrieving relevant documents from an external corpus. However, existing RAG systems primarily focus on unimodal text documents, and often fall short in real-world scenarios where both queries and documents may contain mixed modalities (such as text and images). In this paper, we address the challenge of Universal Retrieval-Augmented Generation (URAG), which involves retrieving and reasoning over mixed-modal information to improve vision-language generation. To this end, we propose Nyx, a unified mixed-modal to mixed-modal retriever tailored for URAG scenarios. To mitigate the scarcity of realistic mixed-modal data, we introduce a four-stage automated pipeline for generation and filtering, leveraging web documents to construct NyxQA, a dataset comprising diverse mixed-modal question-answer pairs that better reflect real-world information needs. Building on this high-quality dataset, we adopt a two-stage training framework for Nyx: we first perform pre-training on NyxQA along with a variety of open-source retrieval datasets, followed by supervised fine-tuning using feedback from downstream vision-language models (VLMs) to align retrieval outputs with generative preferences. Experimental results demonstrate that Nyx not only performs competitively on standard text-only RAG benchmarks, but also excels in the more general and realistic URAG setting, significantly improving generation quality in vision-language tasks.

[78] The Atomic Instruction Gap: Instruction-Tuned LLMs Struggle with Simple, Self-Contained Directives

Henry Lim, Kwan Hui Lim

Main category: cs.CL

TL;DR: Instruction-tuned LLMs show significant performance variations when option label formats change, revealing instruction-format bias and weak adherence to atomic directives despite strong zero-shot reasoning capabilities.

DetailsMotivation: To investigate whether instruction-tuned LLMs can reliably follow simple, self-contained instructions, which is foundational for complex instruction-following tasks.

Method: Evaluated 20 IT-LLMs on modified MMLU and MMLU-Pro benchmarks by systematically varying option label formats (alphabetic, numeric, Roman) under four paradigms: with explicit instructions, without instructions, with option contents removed, and with three-shot exemplars.

Result: Label format changes caused large performance shifts (up to -30.45%), performance dropped without instructions (up to -10.84%), models failed random-choice baselines when option contents were removed except with numeric labels, and three-shot exemplars provided no significant gains.

Conclusion: Current instruction-tuning paradigms are insufficient, and there’s a need for evaluation methods and training strategies that explicitly target atomic instruction-following, as larger LLMs achieve higher accuracy but remain inconsistent in instruction adherence.

Abstract: Instruction-tuned large language models (IT-LLMs) exhibit strong zero-shot reasoning, yet their ability to execute simple, self-contained instructions remains underexplored, despite this being foundational to complex instruction-following. We evaluate 20 IT-LLMs on modified MMLU and MMLU-Pro benchmarks, by systematically varying the format of option labels (alphabetic, numeric, Roman) while keeping their meaning identical under four paradigms, namely: (1) With explicit instructions, label changes cause large performance shifts (e.g., -30.45% for Roman vs. numeric), revealing instruction-format bias. (2) Without instructions, performance drops further (up to -10.84%) and label sensitivity intensifies, underscoring the role of explicit guidance. (3) When option contents are removed, models fail random-choice baselines except with numeric labels, suggesting weak adherence to atomic directives. (4) Three-shot exemplars yield no significant gains in robustness or fidelity, and generation analyses show persistent label errors, especially for non-numeric formats. Across model sizes, larger LLMs achieve higher accuracy but remain inconsistent in instruction adherence. These results expose the insufficiencies of current instruction-tuning paradigms and highlight the need for evaluation methods and training strategies that explicitly target atomic instruction-following.

[79] EduAdapt: A Question Answer Benchmark Dataset for Evaluating Grade-Level Adaptability in LLMs

Numaan Naeem, Abdellah El Mekki, Muhammad Abdul-Mageed

Main category: cs.CL

TL;DR: EduAdapt is a benchmark for evaluating LLMs’ ability to generate grade-appropriate educational responses, featuring 48k QA pairs across science subjects for Grades 1-12.

DetailsMotivation: LLMs perform well academically but fail to tailor responses to students' grade levels, which is critical for K-12 education where age-appropriate vocabulary and explanations are essential.

Method: Created EduAdapt benchmark with nearly 48k grade-labeled QA pairs across nine science subjects spanning Grades 1-12, grouped into four grade levels. Evaluated diverse open-source LLMs on this benchmark.

Result: Larger models generally perform better but still struggle with generating suitable responses for early-grade students (Grades 1-5).

Conclusion: This work presents the first dataset and evaluation framework for assessing grade-level adaptability in LLMs, aiming to foster more developmentally aligned educational AI systems through better training and prompting strategies.

Abstract: Large language models (LLMs) are transforming education by answering questions, explaining complex concepts, and generating content across a wide range of subjects. Despite strong performance on academic benchmarks, they often fail to tailor responses to students’ grade levels. This is a critical need in K-12 education, where age-appropriate vocabulary and explanation are essential for effective learning. Existing models frequently produce outputs that are too advanced or vague for younger learners, and there are no standardized benchmarks to evaluate their ability to adjust across cognitive and developmental stages. To address this gap, we introduce EduAdapt, a benchmark of nearly 48k grade-labeled QA pairs across nine science subjects, spanning Grades 1-12 and grouped into four grade levels. We evaluate a diverse set of open-source LLMs on EduAdapt and find that while larger models generally perform better, they still struggle with generating suitable responses for early-grade students (Grades 1-5). Our work presents the first dataset and evaluation framework for assessing grade-level adaptability in LLMs, aiming to foster more developmentally aligned educational AI systems through better training and prompting strategies. EduAdapt code and datasets are publicly available at https://github.com/NaumanNaeem/EduAdapt.

[80] Leveraging Group Relative Policy Optimization to Advance Large Language Models in Traditional Chinese Medicine

Jiacheng Xie, Shuai Zeng, Yang Yu, Xiaoting Tang, Guanghui An, Dong Xu

Main category: cs.CL

TL;DR: Ladder-base is the first TCM-focused LLM trained with Group Relative Policy Optimization (GRPO), showing superior performance over both general-purpose and domain-specific models in traditional Chinese medicine reasoning tasks.

DetailsMotivation: Traditional Chinese Medicine presents unique challenges for LLMs, and previous TCM-specific models had limitations in alignment, data quality, and evaluation consistency.

Method: Built on Qwen2.5-7B-Instruct foundation model, trained exclusively on TCM-Ladder benchmark’s textual subset using GRPO reinforcement learning method that optimizes response selection through intra-group comparisons.

Result: Ladder-base demonstrates superior performance across multiple reasoning metrics compared to state-of-the-art general-purpose LLMs (GPT-4, Gemini 2.5, Claude 3, Qwen3) and domain-specific TCM models (BenTsao, HuatuoGPT2, Zhongjing).

Conclusion: GRPO provides an effective strategy for aligning LLMs with expert-level reasoning in traditional medical domains and supports development of trustworthy TCM AI systems.

Abstract: Traditional Chinese Medicine (TCM) presents a rich and structurally unique knowledge system that challenges conventional applications of large language models (LLMs). Although previous TCM-specific LLMs have shown progress through supervised fine-tuning, they often face limitations in alignment, data quality, and evaluation consistency. In this study, we introduce Ladder-base, the first TCM-focused LLM trained with Group Relative Policy Optimization (GRPO), a reinforcement learning method that improves reasoning and factual consistency by optimizing response selection based on intra-group comparisons. Ladder-base is built upon the Qwen2.5-7B-Instruct foundation model and trained exclusively on the textual subset of the TCM-Ladder benchmark, using 80 percent of the data for training and the remaining 20 percent split evenly between validation and test sets. Through standardized evaluation, Ladder-base demonstrates superior performance across multiple reasoning metrics when compared to both state-of-the-art general-purpose LLMs such as GPT-4, Gemini 2.5, Claude 3, and Qwen3 and domain-specific TCM models including BenTsao, HuatuoGPT2, and Zhongjing. These findings suggest that GRPO provides an effective and efficient strategy for aligning LLMs with expert-level reasoning in traditional medical domains and supports the development of trustworthy and clinically grounded TCM artificial intelligence systems.

[81] AFRICAPTION: Establishing a New Paradigm for Image Captioning in African Languages

Mardiyyah Oduwole, Prince Mireku, Fatimo Adebanjo, Oluwatosin Olajide, Mahi Aminu Aliyu, Jekaterina Novikova

Main category: cs.CL

TL;DR: AfriCaption is a framework for multilingual image captioning in 20 African languages, addressing the gap in multimodal AI research for under-resourced languages through curated datasets, dynamic pipelines, and a specialized model.

DetailsMotivation: Multimodal AI research has focused on high-resource languages, limiting democratization. This work aims to make advancements accessible by focusing on under-represented African languages.

Method: The framework includes: (i) a curated dataset from Flickr8k with semantically aligned captions via context-aware selection and translation; (ii) a dynamic pipeline with model ensembling and adaptive substitution for quality preservation; (iii) AfriCaption model (0.5B parameters) integrating SigLIP and NLLB200 for vision-to-text caption generation.

Result: The framework establishes the first scalable image-captioning resource for under-represented African languages, ensuring ongoing data quality and enabling inclusive multimodal AI.

Conclusion: AfriCaption lays the groundwork for truly inclusive multimodal AI by providing a comprehensive solution for multilingual image captioning in African languages, addressing the democratization gap in the field.

Abstract: Multimodal AI research has overwhelmingly focused on high-resource languages, hindering the democratization of advancements in the field. To address this, we present AfriCaption, a comprehensive framework for multilingual image captioning in 20 African languages and our contributions are threefold: (i) a curated dataset built on Flickr8k, featuring semantically aligned captions generated via a context-aware selection and translation process; (ii) a dynamic, context-preserving pipeline that ensures ongoing quality through model ensembling and adaptive substitution; and (iii) the AfriCaption model, a 0.5B parameter vision-to-text architecture that integrates SigLIP and NLLB200 for caption generation across under-represented languages. This unified framework ensures ongoing data quality and establishes the first scalable image-captioning resource for under-represented African languages, laying the groundwork for truly inclusive multimodal AI.

[82] Navigating the Alignment-Calibration Trade-off: A Pareto-Superior Frontier via Model Merging

Tiancheng Hu, Benjamin Minixhofer, Nigel Collier

Main category: cs.CL

TL;DR: Simple weight interpolation between pre- and post-alignment models can create Pareto-optimal models that improve accuracy and recover calibration lost during alignment.

DetailsMotivation: Post-training alignment causes not just accuracy drops but also severe calibration loss, making models overconfident and less reliable.

Method: Post-hoc weight interpolation between models before and after alignment training.

Result: Consistently reveals Pareto-optimal interpolations that improve accuracy beyond both parent models while recovering calibration.

Conclusion: Simple model merging efficiently mitigates the full alignment tax, yielding more capable and reliable models.

Abstract: The “alignment tax” of post-training is typically framed as a drop in task accuracy. We show it also involves a severe loss of calibration, making models overconfident, less reliable, and model outputs less diverse. We show that this trade-off can be navigated effectively via a simple post-hoc intervention: interpolating between a model’s weights before and after alignment. Crucially, this is not a strict trade-off. We find that the process consistently reveals Pareto-optimal interpolations - models that improve accuracy beyond both parents while substantially recovering the calibration lost during alignment. Our work demonstrates that simple model merging provides a computationally efficient method for mitigating the full scope of the alignment tax, yielding models that are more capable and more reliable.

[83] Agentic Reinforcement Learning for Search is Unsafe

Yushi Yang, Shreyansh Padarha, Andrew Lee, Adam Mahdi

Main category: cs.CL

TL;DR: RL-trained search models have fragile safety - simple attacks can bypass refusal mechanisms and trigger harmful searches.

DetailsMotivation: To understand safety vulnerabilities in agentic reinforcement learning models that use search tools, as their safety properties are not well understood despite their multi-step reasoning capabilities.

Method: Two simple attacks: Search attack (forcing model to begin with search) and Multi-search attack (encouraging repeated searches). Tested across Qwen and Llama model families with local and web search.

Result: Attacks lowered refusal rates by up to 60.0%, answer safety by 82.5%, and search-query safety by 82.4%. Models generate harmful, request-mirroring search queries before refusal tokens.

Conclusion: Current RL training rewards effective queries without accounting for harmfulness, creating exploitable vulnerabilities. Urgent need for safety-aware agentic RL pipelines that optimize for safe search.

Abstract: Agentic reinforcement learning (RL) trains large language models to autonomously call tools during reasoning, with search as the most common application. These models excel at multi-step reasoning tasks, but their safety properties are not well understood. In this study, we show that RL-trained search models inherit refusal from instruction tuning and often deflect harmful requests by turning them into safe queries. However, this safety is fragile. Two simple attacks, one that forces the model to begin response with search (Search attack), another that encourages models to repeatedly search (Multi-search attack), trigger cascades of harmful searches and answers. Across two model families (Qwen, Llama) with both local and web search, these attacks lower refusal rates by up to 60.0%, answer safety by 82.5%, and search-query safety by 82.4%. The attacks succeed by triggering models to generate harmful, request-mirroring search queries before they can generate the inherited refusal tokens. This exposes a core weakness of current RL training: it rewards continued generation of effective queries without accounting for their harmfulness. As a result, RL search models have vulnerabilities that users can easily exploit, making it urgent to develop safety-aware agentic RL pipelines optimising for safe search.

[84] Multilingual Clinical NER for Diseases and Medications Recognition in Cardiology Texts using BERT Embeddings

Manuela Daniela Danu, George Marica, Constantin Suciu, Lucian Mihai Itu, Oladimeji Farri

Main category: cs.CL

TL;DR: Developed deep contextual embedding models for clinical named entity recognition in cardiology domain across English, Spanish, and Italian languages, achieving state-of-the-art performance in the BioASQ MultiCardioNER shared task.

DetailsMotivation: Address the scarcity of research on clinical NER for low-resource languages despite the growing volume of EHR data and the need to extract biomedical knowledge from unstructured clinical texts.

Method: Explored monolingual and multilingual BERT-based models trained on general domain text for extracting disease and medication mentions from clinical case reports in three languages.

Result: Achieved F1-scores of 77.88% (Spanish Diseases), 92.09% (Spanish Medications), 91.74% (English Medications), and 88.9% (Italian Medications), outperforming mean and median scores across all subtasks.

Conclusion: The proposed deep contextual embedding models effectively enhance clinical NER performance in multiple languages, demonstrating the viability of adapting general-domain BERT models for specialized clinical text processing tasks.

Abstract: The rapidly increasing volume of electronic health record (EHR) data underscores a pressing need to unlock biomedical knowledge from unstructured clinical texts to support advancements in data-driven clinical systems, including patient diagnosis, disease progression monitoring, treatment effects assessment, prediction of future clinical events, etc. While contextualized language models have demonstrated impressive performance improvements for named entity recognition (NER) systems in English corpora, there remains a scarcity of research focused on clinical texts in low-resource languages. To bridge this gap, our study aims to develop multiple deep contextual embedding models to enhance clinical NER in the cardiology domain, as part of the BioASQ MultiCardioNER shared task. We explore the effectiveness of different monolingual and multilingual BERT-based models, trained on general domain text, for extracting disease and medication mentions from clinical case reports written in English, Spanish, and Italian. We achieved an F1-score of 77.88% on Spanish Diseases Recognition (SDR), 92.09% on Spanish Medications Recognition (SMR), 91.74% on English Medications Recognition (EMR), and 88.9% on Italian Medications Recognition (IMR). These results outperform the mean and median F1 scores in the test leaderboard across all subtasks, with the mean/median values being: 69.61%/75.66% for SDR, 81.22%/90.18% for SMR, 89.2%/88.96% for EMR, and 82.8%/87.76% for IMR.

[85] Evaluating Large Language Models on Urdu Idiom Translation

Muhammad Farmal Khan, Mousumi Akter

Main category: cs.CL

TL;DR: This paper introduces the first evaluation datasets for Urdu to English idiomatic translation, evaluates LLMs and NMT systems on preserving idiomatic meaning, and finds that prompt engineering helps and Native Urdu inputs outperform Roman Urdu.

DetailsMotivation: Idiomatic translation is a major challenge in machine translation, especially for low-resource languages like Urdu, and has received limited prior research attention.

Method: Created first Urdu to English idiomatic translation datasets for both Native and Roman Urdu scripts, evaluated multiple LLMs and NMT systems using automatic metrics (BLEU, BERTScore, COMET, XCOMET), and tested prompt engineering approaches.

Result: Prompt engineering improves idiomatic translation compared to direct translation, though differences between prompt types are minor. Native Urdu inputs produce more accurate idiomatic translations than Roman Urdu.

Conclusion: Text representation significantly impacts translation quality for idiomatic expressions, with Native Urdu performing better than Roman Urdu, and prompt engineering can enhance idiomatic translation capabilities.

Abstract: Idiomatic translation remains a significant challenge in machine translation, especially for low resource languages such as Urdu, and has received limited prior attention. To advance research in this area, we introduce the first evaluation datasets for Urdu to English idiomatic translation, covering both Native Urdu and Roman Urdu scripts and annotated with gold-standard English equivalents. We evaluate multiple open-source Large Language Models (LLMs) and Neural Machine Translation (NMT) systems on this task, focusing on their ability to preserve idiomatic and cultural meaning. Automatic metrics including BLEU, BERTScore, COMET, and XCOMET are used to assess translation quality. Our findings indicate that prompt engineering enhances idiomatic translation compared to direct translation, though performance differences among prompt types are relatively minor. Moreover, cross script comparisons reveal that text representation substantially affects translation quality, with Native Urdu inputs producing more accurate idiomatic translations than Roman Urdu.

[86] Disparities in Multilingual LLM-Based Healthcare Q&A

Ipek Baris Schlicht, Burcu Sayin, Zhixue Zhao, Frederik M. Labonté, Cesare Barbera, Marco Viviani, Paolo Rosso, Lucie Flek

Main category: cs.CL

TL;DR: The study examines cross-lingual disparities in healthcare information quality across English, German, Turkish, Chinese, and Italian, revealing that LLM responses align more with English Wikipedia even for non-English prompts, but providing non-English contextual excerpts can improve factual alignment.

DetailsMotivation: To address concerns about reliability and consistency of multilingual LLMs in healthcare, particularly regarding equitable access to reliable health information across different languages.

Method: Constructed MultiWikiHealthCare dataset from Wikipedia, analyzed cross-lingual healthcare coverage, assessed LLM response alignment with references, and conducted case studies using contextual information and RAG across five languages.

Result: Substantial cross-lingual disparities in both Wikipedia coverage and LLM factual alignment; LLM responses consistently align more with English Wikipedia regardless of prompt language; providing non-English contextual excerpts effectively shifts factual alignment toward culturally relevant knowledge.

Conclusion: The findings highlight practical pathways for building more equitable, multilingual AI systems in healthcare by addressing cross-lingual disparities through contextual information and RAG approaches.

Abstract: Equitable access to reliable health information is vital when integrating AI into healthcare. Yet, information quality varies across languages, raising concerns about the reliability and consistency of multilingual Large Language Models (LLMs). We systematically examine cross-lingual disparities in pre-training source and factuality alignment in LLM answers for multilingual healthcare Q&A across English, German, Turkish, Chinese (Mandarin), and Italian. We (i) constructed Multilingual Wiki Health Care (MultiWikiHealthCare), a multilingual dataset from Wikipedia; (ii) analyzed cross-lingual healthcare coverage; (iii) assessed LLM response alignment with these references; and (iv) conducted a case study on factual alignment through the use of contextual information and Retrieval-Augmented Generation (RAG). Our findings reveal substantial cross-lingual disparities in both Wikipedia coverage and LLM factual alignment. Across LLMs, responses align more with English Wikipedia, even when the prompts are non-English. Providing contextual excerpts from non-English Wikipedia at inference time effectively shifts factual alignment toward culturally relevant knowledge. These results highlight practical pathways for building more equitable, multilingual AI systems for healthcare.

[87] ReXMoE: Reusing Experts with Minimal Overhead in Mixture-of-Experts

Zheyue Tan, Zhiyuan Li, Tao Yuan, Dong Zhou, Weilin Liu, Yueqing Zhuang, Yadong Li, Guowei Niu, Cheng Qin, Zhuyu Yao, Congyi Liu, Haiyang Xu, Boxun Li, Guohao Dai, Bo Zhao, Yu Wang

Main category: cs.CL

TL;DR: ReXMoE is a novel Mixture-of-Experts architecture that enables cross-layer expert reuse, overcoming limitations of layer-local routing and improving performance without increasing parameter count.

DetailsMotivation: Current MoE architectures are limited by layer-local routing where each layer can only use its own expert pool, creating a trade-off between expert dimensionality and routing diversity within fixed parameter budgets.

Method: ReXMoE allows routers to reuse experts across adjacent layers, decoupling expert dimensionality from per-layer budgets. It uses progressive scaling routing (PSR) to gradually increase candidate expert pool during training.

Result: Extensive experiments on 0.5B to 7B parameter models show ReXMoE consistently improves language modeling and downstream task performance under fixed architectural dimensions.

Conclusion: ReXMoE represents a new design paradigm for parameter-efficient and scalable MoE-based LLMs, enabling richer expert combinations without sacrificing capacity or inflating parameters.

Abstract: Mixture-of-Experts (MoE) architectures have emerged as a promising approach to scale Large Language Models (LLMs). MoE boosts the efficiency by activating a subset of experts per token. Recent works show that fine-grained experts substantially enriches the combinatorial flexibility of active experts and enhances model expressiveness. However, such a design is fundamentally limited by the layer-local routing mechanism: each layer is restricted to its own expert pool. This requires a careful trade-off between expert dimensionality and routing diversity given fixed parameter budgets. We describe ReXMoE, a novel MoE architecture that improves routing beyond the existing layer-local approaches by allowing routers to reuse experts across adjacent layers. ReXMoE decouples expert dimensionality from per-layer budgets, enabling richer expert combinations without sacrificing individual expert capacity or inflating overall parameters. To this end, we propose a new progressive scaling routing (PSR) strategy to gradually increase the candidate expert pool during training. As a result, ReXMoE improves both language modeling and downstream task performance. Extensive experiments on models ranging from 0.5B to 7B parameters across different architectures demonstrate that ReXMoE consistently improves performance under fixed architectural dimensions, confirming ReXMoE as new design paradigm for parameter-efficient and scalable MoE-based LLMs.

[88] DETree: DEtecting Human-AI Collaborative Texts via Tree-Structured Hierarchical Representation Learning

Yongxin He, Shan Zhang, Yixuan Cao, Lei Ma, Ping Luo

Main category: cs.CL

TL;DR: DETree is a novel approach for detecting AI-involved text that models human-AI collaboration processes using a Hierarchical Affinity Tree structure and specialized loss function, improving detection performance and robustness in out-of-distribution scenarios.

DetailsMotivation: Current AI text detection methods crudely model complex human-AI collaboration processes (AI-written text edited by humans, human-written text edited by AI, AI-generated text refined by other AI) using binary or multi-class classification, failing to capture the inherent clustering relationships in text representations from different processes.

Method: Proposed DETree approach models relationships among different human-AI collaboration processes as a Hierarchical Affinity Tree structure with specialized loss function that aligns text representations with the tree. Developed RealBench dataset containing various hybrid texts from different collaboration processes to facilitate learning.

Result: DETree improves performance in hybrid text detection tasks and significantly enhances robustness and generalization in out-of-distribution scenarios, particularly in few-shot learning conditions. Demonstrates promise of training-based approaches in OOD settings.

Conclusion: The DETree framework effectively captures the complex relationships in human-AI collaborative text generation processes through hierarchical modeling, providing superior detection performance and robustness compared to existing methods.

Abstract: Detecting AI-involved text is essential for combating misinformation, plagiarism, and academic misconduct. However, AI text generation includes diverse collaborative processes (AI-written text edited by humans, human-written text edited by AI, and AI-generated text refined by other AI), where various or even new LLMs could be involved. Texts generated through these varied processes exhibit complex characteristics, presenting significant challenges for detection. Current methods model these processes rather crudely, primarily employing binary classification (purely human vs. AI-involved) or multi-classification (treating human-AI collaboration as a new class). We observe that representations of texts generated through different processes exhibit inherent clustering relationships. Therefore, we propose DETree, a novel approach that models the relationships among different processes as a Hierarchical Affinity Tree structure, and introduces a specialized loss function that aligns text representations with this tree. To facilitate this learning, we developed RealBench, a comprehensive benchmark dataset that automatically incorporates a wide spectrum of hybrid texts produced through various human-AI collaboration processes. Our method improves performance in hybrid text detection tasks and significantly enhances robustness and generalization in out-of-distribution scenarios, particularly in few-shot learning conditions, further demonstrating the promise of training-based approaches in OOD settings. Our code and dataset are available at https://github.com/heyongxin233/DETree.

[89] Empowering Real-World: A Survey on the Technology, Practice, and Evaluation of LLM-driven Industry Agents

Yihong Tang, Kehai Chen, Liang Yue, Jinxin Fan, Caishen Zhou, Xiaoguang Li, Yuyang Zhang, Mingming Zhao, Shixiong Kai, Kaiyang Guo, Xingshan Zeng, Wenjing Cun, Lifeng Shang, Min Zhang

Main category: cs.CL

TL;DR: This paper provides a systematic review of LLM-based industry agents, examining their technological pillars (Memory, Planning, Tool Use), applications across various domains, evaluation methods, and practical challenges.

DetailsMotivation: To address the challenge of translating general agent research into productivity that drives industry transformations, by providing a comprehensive overview of industry agents' technologies, applications, and evaluation methods.

Method: Uses an industry agent capability maturity framework to outline agent evolution from ‘process execution systems’ to ‘adaptive social systems’, and systematically reviews three key technological pillars (Memory, Planning, Tool Use), real-world applications, and evaluation benchmarks.

Result: The review clarifies the current state of industry agents, identifies challenges in evaluation systems regarding authenticity, safety, and industry specificity, and explores practical challenges including capability boundaries, developmental potential, and governance issues.

Conclusion: By combining technological evolution with industry practices, the paper provides a clear roadmap and theoretical foundation for understanding and building the next generation of industry agents.

Abstract: With the rise of large language models (LLMs), LLM agents capable of autonomous reasoning, planning, and executing complex tasks have become a frontier in artificial intelligence. However, how to translate the research on general agents into productivity that drives industry transformations remains a significant challenge. To address this, this paper systematically reviews the technologies, applications, and evaluation methods of industry agents based on LLMs. Using an industry agent capability maturity framework, it outlines the evolution of agents in industry applications, from “process execution systems” to “adaptive social systems.” First, we examine the three key technological pillars that support the advancement of agent capabilities: Memory, Planning, and Tool Use. We discuss how these technologies evolve from supporting simple tasks in their early forms to enabling complex autonomous systems and collective intelligence in more advanced forms. Then, we provide an overview of the application of industry agents in real-world domains such as digital engineering, scientific discovery, embodied intelligence, collaborative business execution, and complex system simulation. Additionally, this paper reviews the evaluation benchmarks and methods for both fundamental and specialized capabilities, identifying the challenges existing evaluation systems face regarding authenticity, safety, and industry specificity. Finally, we focus on the practical challenges faced by industry agents, exploring their capability boundaries, developmental potential, and governance issues in various scenarios, while providing insights into future directions. By combining technological evolution with industry practices, this review aims to clarify the current state and offer a clear roadmap and theoretical foundation for understanding and building the next generation of industry agents.

[90] Deep Self-Evolving Reasoning

Zihan Liu, Shun Zheng, Xumeng Wen, Yang Wang, Jiang Bian, Mao Yang

Main category: cs.CL

TL;DR: Deep Self-Evolving Reasoning (DSER) extends reasoning capabilities of smaller models through probabilistic parallel processes, enabling them to solve previously unsolvable problems despite weak verification capabilities.

DetailsMotivation: Current verification-refinement frameworks require strong verification capabilities that are fragile in open-weight, smaller-scale models, limiting their effectiveness on hard reasoning tasks.

Method: DSER conceptualizes iterative reasoning as a Markov chain with stochastic transitions, running multiple long-horizon self-evolving processes in parallel to amplify small positive tendencies toward correct solutions.

Result: Applied to DeepSeek-R1-0528-Qwen3-8B, DSER solved 5 out of 9 previously unsolvable problems on AIME 2024-2025 benchmark and enabled this compact model to surpass its 600B-parameter teacher’s single-turn accuracy through majority voting.

Conclusion: DSER framework not only provides immediate test-time scaling utility but also diagnoses fundamental limitations of current open-weight reasoners, establishing a clear research agenda for developing next-generation models with intrinsic self-evolving capabilities.

Abstract: Long-form chain-of-thought reasoning has become a cornerstone of advanced reasoning in large language models. While recent verification-refinement frameworks have enabled proprietary models to solve Olympiad-level problems, their effectiveness hinges on strong, reliable verification and correction capabilities, which remain fragile in open-weight, smaller-scale models. This work demonstrates that even with weak verification and refinement capabilities on hard tasks, the reasoning limits of such models can be substantially extended through a probabilistic paradigm we call Deep Self-Evolving Reasoning (DSER). We conceptualize iterative reasoning as a Markov chain, where each step represents a stochastic transition in the solution space. The key insight is that convergence to a correct solution is guaranteed as long as the probability of improvement marginally exceeds that of degradation. By running multiple long-horizon, self-evolving processes in parallel, DSER amplifies these small positive tendencies, enabling the model to asymptotically approach correct answers. Empirically, we apply DSER to the DeepSeek-R1-0528-Qwen3-8B model. On the challenging AIME 2024-2025 benchmark, DSER solves 5 out of 9 previously unsolvable problems and boosts overall performance, enabling this compact model to surpass the single-turn accuracy of its 600B-parameter teacher through majority voting. Beyond its immediate utility for test-time scaling, the DSER framework serves to diagnose the fundamental limitations of current open-weight reasoners. By clearly delineating their shortcomings in self-verification, refinement, and stability, our findings establish a clear research agenda for developing next-generation models with powerful, intrinsic self-evolving capabilities.

[91] Lingua Custodi’s participation at the WMT 2025 Terminology shared task

Jingshu Liu, Raheel Qader, Gaëtan Caillaut, Mariam Nakhlé

Main category: cs.CL

TL;DR: This paper explores methods for learning multilingual sentence embeddings using BERT-based approaches, achieving significant improvements in cross-lingual retrieval accuracy while maintaining competitive monolingual performance.

DetailsMotivation: BERT has shown effectiveness for monolingual sentence embeddings, but cross-lingual sentence embeddings using BERT have not been sufficiently explored. The authors aim to develop effective multilingual sentence embedding methods.

Method: Combined best methods for learning monolingual and cross-lingual representations: masked language modeling (MLM), translation language modeling (TLM), dual encoder translation ranking, and additive margin softmax. Used pre-trained multilingual language models to reduce parallel data requirements.

Result: Achieved 83.7% bi-text retrieval accuracy over 112 languages on Tatoeba (vs 65.5% by LASER), with 80% reduction in parallel training data requirements. The model also performs competitively on monolingual transfer learning benchmarks and can mine parallel data for training competitive NMT models.

Conclusion: The proposed methods successfully create high-quality multilingual sentence embeddings that significantly outperform previous approaches in cross-lingual retrieval while maintaining strong monolingual performance. The model is publicly released for 109+ languages.

Abstract: While BERT is an effective method for learning monolingual sentence embeddings for semantic similarity and embedding based transfer learning BERT based cross-lingual sentence embeddings have yet to be explored. We systematically investigate methods for learning multilingual sentence embeddings by combining the best methods for learning monolingual and cross-lingual representations including: masked language modeling (MLM), translation language modeling (TLM), dual encoder translation ranking, and additive margin softmax. We show that introducing a pre-trained multilingual language model dramatically reduces the amount of parallel training data required to achieve good performance by 80%. Composing the best of these methods produces a model that achieves 83.7% bi-text retrieval accuracy over 112 languages on Tatoeba, well above the 65.5 achieved by LASER, while still performing competitively on monolingual transfer learning benchmarks. Parallel data mined from CommonCrawl using our best model is shown to train competitive NMT models for en-zh and en-de. We publicly release our best multilingual sentence embedding model for 109+ languages at https://tfhub.dev/google/LaBSE.

[92] Annotation-Efficient Universal Honesty Alignment

Shiyu Ni, Keping Bi, Jiafeng Guo, Minghao Tang, Jingtong Wu, Zengxin Han, Xueqi Cheng

Main category: cs.CL

TL;DR: EliCal is a two-stage framework for honesty alignment in LLMs that first elicits internal confidence using self-consistency supervision, then calibrates with minimal correctness annotations, achieving near-optimal performance with only 0.18% of full supervision.

DetailsMotivation: Existing honesty alignment methods require costly large-scale labeling for training-based calibration, creating a need for annotation-efficient approaches to achieve universal honesty alignment in LLMs.

Method: Two-stage framework: 1) Elicitation stage uses inexpensive self-consistency supervision to extract internal confidence, 2) Calibration stage uses a small set of correctness annotations (only 1k) to refine the confidence estimates.

Result: EliCal achieves near-optimal alignment with only 1k correctness annotations (0.18% of full supervision) and shows better alignment performance on unseen MMLU tasks compared to calibration-only baselines.

Conclusion: EliCal offers a scalable solution for universal honesty alignment in LLMs by significantly reducing annotation costs while maintaining strong performance across diverse tasks.

Abstract: Honesty alignment-the ability of large language models (LLMs) to recognize their knowledge boundaries and express calibrated confidence-is essential for trustworthy deployment. Existing methods either rely on training-free confidence estimation (e.g., token probabilities, self-consistency) or training-based calibration with correctness annotations. While effective, achieving universal honesty alignment with training-based calibration requires costly, large-scale labeling. To support annotation-efficient training, we introduce Elicitation-Then-Calibration (EliCal), a two-stage framework that first elicits internal confidence using inexpensive self-consistency supervision, then calibrates this confidence with a small set of correctness annotations. To support a large-scale study, we release HonestyBench, a benchmark covering ten free-form QA datasets with 560k training and 70k evaluation instances annotated with correctness and self-consistency signals. Experiments show that EliCal achieves near-optimal alignment with only 1k correctness annotations (0.18% of full supervision) and better alignment performance on unseen MMLU tasks than the calibration-only baseline, offering a scalable solution toward universal honesty alignment in LLMs.

[93] SimBench: Benchmarking the Ability of Large Language Models to Simulate Human Behaviors

Tiancheng Hu, Joachim Baumann, Lorenzo Lupo, Dirk Hovy, Nigel Collier, Paul Röttger

Main category: cs.CL

TL;DR: SimBench is the first large-scale standardized benchmark for evaluating LLM simulations of human behavior, unifying 20 diverse datasets to provide reproducible evaluation of when, why, and how LLM simulations succeed or fail.

DetailsMotivation: Current evaluations of LLM simulations are fragmented with bespoke tasks and metrics, creating incomparable results. There's a need for standardized benchmarks to enable robust, reproducible science of LLM simulation.

Method: Introduced SimBench benchmark that unifies 20 diverse datasets covering tasks from moral decision-making to economic choice across a large global participant pool. Evaluated LLM performance across different model sizes, inference-time compute, instruction-tuning approaches, and demographic simulations.

Result: Best LLMs today have limited simulation ability (score: 40.80/100). Performance scales log-linearly with model size. Simulation performance is not improved by increased inference-time compute. Shows alignment-simulation trade-off: instruction-tuning helps on low-entropy questions but degrades performance on high-entropy ones. Models struggle with specific demographic groups. Simulation ability correlates strongly with deep, knowledge-intensive reasoning (MMLU-Pro, r=0.939).

Conclusion: SimBench enables measurable progress in LLM simulation development by providing standardized evaluation. Current LLMs have limited but scalable simulation capabilities, with specific limitations around demographic simulation and trade-offs between alignment and simulation performance.

Abstract: Large language model (LLM) simulations of human behavior have the potential to revolutionize the social and behavioral sciences, if and only if they faithfully reflect real human behaviors. Current evaluations are fragmented, based on bespoke tasks and metrics, creating a patchwork of incomparable results. To address this, we introduce SimBench, the first large-scale, standardized benchmark for a robust, reproducible science of LLM simulation. By unifying 20 diverse datasets covering tasks from moral decision-making to economic choice across a large global participant pool, SimBench provides the necessary foundation to ask fundamental questions about when, how, and why LLM simulations succeed or fail. We show that, while even the best LLMs today have limited simulation ability (score: 40.80/100), performance scales log-linearly with model size. Simulation performance is not improved by increased inference-time compute. We demonstrate an alignment-simulation trade-off: instruction-tuning improves performance on low-entropy (consensus) questions but degrades it on high-entropy (diverse) ones. Models particularly struggle when simulating specific demographic groups. Finally, we demonstrate that simulation ability correlates most strongly with deep, knowledge-intensive reasoning (MMLU-Pro, r=0.939). By making progress measurable, we aim to accelerate the development of more faithful LLM simulators.

[94] OncoReason: Structuring Clinical Reasoning in LLMs for Robust and Interpretable Survival Prediction

Raghu Vamshi Hemadri, Geetha Krishna Guruju, Kristi Topollai, Anna Ewa Choromanska

Main category: cs.CL

TL;DR: A unified multi-task learning framework that aligns LLMs with clinical reasoning for cancer outcome prediction, using CoT prompting and GRPO to improve interpretability and performance.

DetailsMotivation: Predicting cancer treatment outcomes requires accurate and interpretable models, especially with heterogeneous clinical data. LLMs lack structured reasoning capabilities needed for high-stakes decision support.

Method: Multi-task learning framework training LLMs to jointly perform binary survival classification, continuous survival time regression, and natural language rationale generation. Evaluated three alignment strategies: standard SFT, SFT with CoT prompting, and GRPO reinforcement learning.

Result: CoT prompting improved F1 by +6.0 and reduced MAE by 12%. GRPO achieved state-of-the-art interpretability and predictive performance across BLEU, ROUGE, and BERTScore metrics.

Conclusion: Reasoning-aware alignment is crucial for multi-task clinical modeling, setting a new benchmark for interpretable, trustworthy LLMs in precision oncology.

Abstract: Predicting cancer treatment outcomes requires models that are both accurate and interpretable, particularly in the presence of heterogeneous clinical data. While large language models (LLMs) have shown strong performance in biomedical NLP, they often lack structured reasoning capabilities critical for high-stakes decision support. We present a unified, multi-task learning framework that aligns autoregressive LLMs with clinical reasoning for outcome prediction on the MSK-CHORD dataset. Our models are trained to jointly perform binary survival classification, continuous survival time regression, and natural language rationale generation. We evaluate three alignment strategies: (1) standard supervised fine-tuning (SFT), (2) SFT with Chain-of-Thought (CoT) prompting to elicit step-by-step reasoning, and (3) Group Relative Policy Optimization (GRPO), a reinforcement learning method that aligns model outputs to expert-derived reasoning trajectories. Experiments with LLaMa3-8B and Med42-8B backbones demonstrate that CoT prompting improves F1 by +6.0 and reduces MAE by 12%, while GRPO achieves state-of-the-art interpretability and predictive performance across BLEU, ROUGE, and BERTScore. We further show that existing biomedical LLMs often fail to produce valid reasoning traces due to architectural constraints. Our findings underscore the importance of reasoning-aware alignment in multi-task clinical modeling and set a new benchmark for interpretable, trustworthy LLMs in precision oncology.

[95] When Annotators Disagree, Topology Explains: Mapper, a Topological Tool for Exploring Text Embedding Geometry and Ambiguity

Nisrine Rair, Alban Goupil, Valeriu Vrabie, Emmanuel Chochoy

Main category: cs.CL

TL;DR: This paper proposes using Mapper, a topological data analysis tool, to analyze how language models internally represent ambiguity, revealing that fine-tuning creates modular, non-convex regions in embedding space that align with model predictions even for ambiguous cases.

DetailsMotivation: Traditional scalar metrics like accuracy fail to capture how models internally represent ambiguity, especially when human annotators disagree. There's a need to understand how models encode ambiguous instances.

Method: Applied Mapper (a topological data analysis tool) to analyze RoBERTa-Large fine-tuned on the MD-Offense dataset, comparing it with traditional tools like PCA and UMAP.

Result: Fine-tuning restructures embedding space into modular, non-convex regions aligned with model predictions (>98% of components show ≥90% prediction purity). However, alignment with ground-truth labels drops in ambiguous data, revealing tension between structural confidence and label uncertainty.

Conclusion: Mapper serves as a powerful diagnostic tool for understanding how models resolve ambiguity, enabling topological metrics that can inform proactive modeling strategies in subjective NLP tasks.

Abstract: Language models are often evaluated with scalar metrics like accuracy, but such measures fail to capture how models internally represent ambiguity, especially when human annotators disagree. We propose a topological perspective to analyze how fine-tuned models encode ambiguity and more generally instances. Applied to RoBERTa-Large on the MD-Offense dataset, Mapper, a tool from topological data analysis, reveals that fine-tuning restructures embedding space into modular, non-convex regions aligned with model predictions, even for highly ambiguous cases. Over $98%$ of connected components exhibit $\geq 90%$ prediction purity, yet alignment with ground-truth labels drops in ambiguous data, surfacing a hidden tension between structural confidence and label uncertainty. Unlike traditional tools such as PCA or UMAP, Mapper captures this geometry directly uncovering decision regions, boundary collapses, and overconfident clusters. Our findings position Mapper as a powerful diagnostic tool for understanding how models resolve ambiguity. Beyond visualization, it also enables topological metrics that may inform proactive modeling strategies in subjective NLP tasks.

[96] Language Confusion Gate: Language-Aware Decoding Through Model Self-Distillation

Collin Zhang, Fei Huang, Chenhan Yuan, Junyang Lin

Main category: cs.CL

TL;DR: LCG is a lightweight plug-in that reduces language confusion in LLMs during decoding without retraining, using norm-adjusted self-distillation to filter tokens while preserving code-switching.

DetailsMotivation: Current solutions for language confusion either require model retraining or cannot distinguish between harmful confusion and acceptable code-switching.

Method: Language Confusion Gate (LCG) filters tokens during decoding using norm-adjusted self-distillation to predict language families and apply masking only when needed, leveraging findings that correct-language tokens are usually top predictions.

Result: LCG decreases language confusion significantly (often by an order of magnitude) across various models including Qwen3, GPT-OSS, Gemma3, Llama3.1 without negatively impacting task performance.

Conclusion: LCG provides an effective plug-in solution to reduce language confusion in LLMs without model retraining, distinguishing between harmful confusion and acceptable code-switching.

Abstract: Large language models (LLMs) often experience language confusion, which is the unintended mixing of languages during text generation. Current solutions to this problem either necessitate model retraining or cannot differentiate between harmful confusion and acceptable code-switching. This paper introduces the Language Confusion Gate (LCG), a lightweight, plug-in solution that filters tokens during decoding without altering the base LLM. The LCG is trained using norm-adjusted self-distillation to predict appropriate language families and apply masking only when needed. Our method is based on the findings that language confusion is infrequent, correct-language tokens are usually among the top predictions, and output token embedding norms are larger for high-resource languages, which biases sampling. When evaluated across various models, including Qwen3, GPT-OSS, Gemma3, Llama3.1, LCG decreases language confusion significantly, often by an order of magnitude, without negatively impacting task performance. Code is available at https://github.com/collinzrj/language_confusion_gate.

[97] HGAdapter: Hypergraph-based Adapters in Language Models for Code Summarization and Clone Detection

Guang Yang, Yujie Zhu

Main category: cs.CL

TL;DR: HGAdapter enhances pre-trained language models for code tasks by capturing high-order correlations in code tokens through hypergraph neural networks and adapter tuning.

DetailsMotivation: Current PLMs for code tasks don't consider high-order data correlations within code, which could improve performance.

Method: Proposed three high-order correlations (AST family, lexical, line), designed tokens/hyperedges generator, and created HGAdapter - a hypergraph-based adapter that combines improved hypergraph neural networks with adapter tuning.

Result: Improved PLM performance on code summarization and clone detection tasks across six languages, with varying degrees of improvement.

Conclusion: High-order data correlations in code contribute to improved effectiveness of PLMs for code-related tasks.

Abstract: Pre-trained language models (PLMs) are increasingly being applied to code-related tasks. Although PLMs have achieved good results, they do not take into account potential high-order data correlations within the code. We propose three types of high-order correlations in code tokens, i.e. abstract syntax tree family correlation, lexical correlation, and line correlation. We design a tokens and hyperedges generator to capture these high-order data correlations. We improve the architecture of hypergraph neural networks and combine it with adapter tuning to propose a novel hypergraph-based adapter (HGAdapter) to fine-tune PLMs. HGAdapter can encode high-order data correlations and is allowed to be inserted into various PLMs to enhance performance. Experiments were conducted on several public datasets, including six languages of code summarization and code clone detection tasks. Our methods improved the performance of PLMs in datasets to varying degrees. Experimental results validate the introduction of high-order data correlations that contribute to improved effectiveness.

Huiyuan Xie, Chenyang Li, Huining Zhu, Chubin Zhang, Yuxiao Ye, Zhenghao Liu, Zhiyuan Liu

Main category: cs.CL

TL;DR: The paper introduces LawChain, a novel framework for modeling legal reasoning in Chinese tort civil cases, addressing gaps in existing computational approaches that focus mainly on criminal cases and generic reasoning methods.

DetailsMotivation: Existing computational legal reasoning approaches rely on generic frameworks like syllogism and IRAC, lack comprehensive examination of nuanced legal reasoning processes, and focus predominantly on criminal cases with insufficient modeling for civil cases.

Method: Developed LawChain - a three-module reasoning framework that operationalizes legal reasoning processes for tort analysis, with each module containing multiple finer-grained sub-steps. Created LawChain_eval benchmark and evaluated state-of-the-art LLMs, then introduced baseline approaches incorporating LawChain-style reasoning through prompting or post-training.

Result: Current LLMs fall short in accurately handling crucial elements of tort legal reasoning. The proposed baseline approaches achieved significant improvements in tort-related legal reasoning and generalized well to related legal analysis tasks like Legal Named-Entity Recognition and Criminal Damages Calculation.

Conclusion: Explicitly modeling legal reasoning chains enhances language models’ reasoning capabilities, demonstrating the value of structured legal reasoning frameworks for improving computational legal analysis.

Abstract: Legal reasoning is a fundamental component of legal analysis and decision-making. Existing computational approaches to legal reasoning predominantly rely on generic reasoning frameworks such as syllogism and IRAC, which do not comprehensively examine the nuanced processes that underpin legal reasoning. Moreover, current research has largely focused on criminal cases, with insufficient modeling for civil cases. In this work, we present a novel framework for explicitly modeling legal reasoning in the analysis of Chinese tort-related civil cases. We first operationalize the legal reasoning processes used in tort analysis into the LawChain framework. LawChain is a three-module reasoning framework, with each module consisting of multiple finer-grained sub-steps. Informed by the LawChain framework, we introduce the task of tort legal reasoning and construct an evaluation benchmark, LawChain$_{eval}$, to systematically assess the critical steps within analytical reasoning chains for tort analysis. Leveraging this benchmark, we evaluate state-of-the-art large language models for their legal reasoning ability in civil tort contexts. Our results indicate that current models still fall short in accurately handling crucial elements of tort legal reasoning. Furthermore, we introduce several baseline approaches that explicitly incorporate LawChain-style reasoning through prompting or post-training. We conduct further experiments on additional legal analysis tasks, such as Legal Named-Entity Recognition and Criminal Damages Calculation, to verify the generalizability of these baselines. The proposed baseline approaches achieve significant improvements in tort-related legal reasoning and generalize well to related legal analysis tasks, thus demonstrating the value of explicitly modeling legal reasoning chains to enhance the reasoning capabilities of language models.

[99] Forget to Know, Remember to Use: Context-Aware Unlearning for Large Language Models

Yuefeng Peng, Parnian Afshar, Megan Ganji, Thomas Butler, Amir Houmansadr, Mingxian Wang, Dezhi Hong

Main category: cs.CL

TL;DR: The paper addresses the problem that current unlearning methods impair models’ ability to use forgotten knowledge when it’s reintroduced in prompts, and proposes a plug-in objective to preserve this contextual utility.

DetailsMotivation: Existing unlearning evaluations overlook the important usability aspect where users may still want models to leverage removed information if it's re-introduced in the prompt, which current methods consistently impair.

Method: Augment unlearning objectives with a plug-in term that preserves the model’s ability to use forgotten knowledge when present in context, evaluated through systematic testing of six state-of-the-art unlearning methods.

Result: The approach restores contextual utility to near original levels while maintaining effective forgetting and retain-set utility, as demonstrated through extensive experiments.

Conclusion: The proposed plug-in objective successfully addresses the contextual utility limitation of existing unlearning methods, enabling models to use forgotten knowledge when contextually appropriate while still achieving effective unlearning.

Abstract: Large language models may encode sensitive information or outdated knowledge that needs to be removed, to ensure responsible and compliant model responses. Unlearning has emerged as an efficient alternative to full retraining, aiming to remove specific knowledge while preserving overall model utility. Existing evaluations of unlearning methods focus on (1) the extent of forgetting of the target knowledge (forget set) and (2) maintaining performance on the retain set (i.e., utility). However, these evaluations overlook an important usability aspect: users may still want the model to leverage the removed information if it is re-introduced in the prompt. In a systematic evaluation of six state-of-the-art unlearning methods, we find that they consistently impair such contextual utility. To address this, we augment unlearning objectives with a plug-in term that preserves the model’s ability to use forgotten knowledge when it is present in context. Extensive experiments demonstrate that our approach restores contextual utility to near original levels while still maintaining effective forgetting and retain-set utility.

[100] Qomhra: A Bilingual Irish-English Large Language Model

Joseph McInerney

Main category: cs.CL

TL;DR: Qomhr’a is a bilingual Irish-English LLM developed under low-resource constraints using a complete pipeline including continued pre-training, instruction tuning, and human preference alignment. It achieves significant performance gains in both languages.

DetailsMotivation: To develop a bilingual Irish-English large language model under low-resource constraints, addressing the need for improved Irish language performance while preserving English capabilities.

Method: Used a complete pipeline: bilingual continued pre-training with mixed Irish-English corpora, instruction tuning with synthesized datasets from Gemini-2.5-Pro, and human preference alignment. Created 30K Irish-English parallel instruction dataset and 1K human preference dataset.

Result: Achieved gains of up to 29% in Irish and 44% in English across benchmarks for translation, gender understanding, topic identification and world knowledge. Demonstrated clear progress in instruction following for chatbot functionality.

Conclusion: Qomhr’a successfully demonstrates the development of a bilingual LLM under low-resource conditions, showing significant improvements in both Irish and English performance through a comprehensive training pipeline.

Abstract: This paper introduces Qomhr'a, a bilingual Irish-English large language model (LLM), developed under low-resource constraints presenting a complete pipeline spanning bilingual continued pre-training, instruction tuning, and alignment from human preferences. Newly accessible Irish corpora and English text are mixed and curated to improve Irish performance while preserving English ability. 6 closed-weight LLMs are judged for their Irish text generation by a native speaker, a learner and other LLMs. Google’s Gemini-2.5-Pro is ranked the highest and is subsequently used to synthesise instruction tuning and human preference datasets. Two datasets are contributed leveraging Gemini-2.5-Pro: a 30K Irish-English parallel instruction tuning dataset and a 1K human preference dataset, generating accepted and rejected responses that show near perfect alignment with a native Irish speaker. Qomhr'a is comprehensively evaluated across benchmarks testing translation, gender understanding, topic identification and world knowledge with gains of up to 29% in Irish and 44% in English. Qomhr'a also undergoes instruction tuning and demonstrates clear progress in instruction following, crucial for chatbot functionality.

[101] Towards Mining Effective Pedagogical Strategies from Learner-LLM Educational Dialogues

Liqun He, Manolis Mavrikis, Mutlu Cukurova

Main category: cs.CL

TL;DR: This paper proposes a dialogue analysis approach to evaluate LLM-based educational applications by focusing on learner-LLM interactions and pedagogical strategies rather than just technical performance.

DetailsMotivation: Existing evaluation methods for educational LLMs focus primarily on technical performance or learning outcomes, neglecting the crucial aspect of learner-LLM interactions in educational dialogues.

Method: The approach involves four steps: dialogue data collection, dialogue act (DA) annotation, DA pattern mining, and predictive model building to identify effective pedagogical strategies.

Result: Early insights are presented as initial findings, with the work representing an ongoing study that lays groundwork for future research in this area.

Conclusion: The research emphasizes the importance of evaluating LLM-based educational applications through dialogue dynamics and pedagogical strategies rather than just technical metrics.

Abstract: Dialogue plays a crucial role in educational settings, yet existing evaluation methods for educational applications of large language models (LLMs) primarily focus on technical performance or learning outcomes, often neglecting attention to learner-LLM interactions. To narrow this gap, this AIED Doctoral Consortium paper presents an ongoing study employing a dialogue analysis approach to identify effective pedagogical strategies from learner-LLM dialogues. The proposed approach involves dialogue data collection, dialogue act (DA) annotation, DA pattern mining, and predictive model building. Early insights are outlined as an initial step toward future research. The work underscores the need to evaluate LLM-based educational applications by focusing on dialogue dynamics and pedagogical strategies.

[102] QueST: Incentivizing LLMs to Generate Difficult Problems

Hanxu Hu, Xingxing Zhang, Jannis Vamvas, Rico Sennrich, Furu Wei

Main category: cs.CL

TL;DR: QueST is a framework that generates challenging coding problems through difficulty-aware graph sampling and rejection fine-tuning, enabling significant performance improvements in LLMs for competitive coding tasks.

DetailsMotivation: Current LLMs face scalability limitations due to insufficient human-labeled datasets and lack of large-scale challenging coding problems, with existing datasets containing only thousands to tens of thousands of problems.

Method: Combines difficulty-aware graph sampling and difficulty-aware rejection fine-tuning to optimize specialized generators for creating challenging coding problems, then uses generated problems for distillation from strong teacher models or reinforcement learning.

Result: Generated 100K difficult problems that, when used to fine-tune Qwen3-8B-base, surpassed original Qwen3-8B performance on LiveCodeBench. With additional 112K examples, the 8B model matched performance of much larger DeepSeek-R1-671B.

Conclusion: QueST provides an effective and scalable approach to advancing competitive coding and reasoning capabilities in LLMs by generating complex problems rather than relying on limited human-labeled data.

Abstract: Large Language Models have achieved strong performance on reasoning tasks, solving competition-level coding and math problems. However, their scalability is limited by human-labeled datasets and the lack of large-scale, challenging coding problem training data. Existing competitive coding datasets contain only thousands to tens of thousands of problems. Previous synthetic data generation methods rely on either augmenting existing instruction datasets or selecting challenging problems from human-labeled data. In this paper, we propose QueST, a novel framework which combines difficulty-aware graph sampling and difficulty-aware rejection fine-tuning that directly optimizes specialized generators to create challenging coding problems. Our trained generators demonstrate superior capability compared to even GPT-4o at creating challenging problems that benefit downstream performance. We leverage QueST to generate large-scale synthetic coding problems, which we then use to distill from strong teacher models with long chain-of-thought or to conduct reinforcement learning for smaller models, proving effective in both scenarios. Our distillation experiments demonstrate significant performance gains. Specifically, after fine-tuning Qwen3-8B-base on 100K difficult problems generated by QueST, we surpass the performance of the original Qwen3-8B on LiveCodeBench. With an additional 112K examples (i.e., 28K human-written problems paired with multiple synthetic solutions), our 8B model matches the performance of the much larger DeepSeek-R1-671B. These findings indicate that generating complex problems via QueST offers an effective and scalable approach to advancing the frontiers of competitive coding and reasoning for large language models.

[103] PANER: A Paraphrase-Augmented Framework for Low-Resource Named Entity Recognition

Nanda Kumar Rengarajan, Jun Yan, Chun Wang

Main category: cs.CL

TL;DR: A lightweight few-shot NER framework that uses instruction tuning with simplified output format and strategic data augmentation to improve performance in low-resource scenarios.

DetailsMotivation: NER requires substantial annotated data, which is challenging in low-resource scenarios. Existing zero-shot and instruction-tuned approaches often fail to generalize to domain-specific entities and don't effectively utilize limited available data.

Method: Two key innovations: (1) new instruction tuning template with simplified output format to leverage large context windows of LLMs, (2) strategic data augmentation that preserves entity information while paraphrasing surrounding context to expand training data.

Result: Achieves performance comparable to state-of-the-art models on few-shot and zero-shot tasks, with average F1 score of 80.1 on CrossNER datasets. Models show consistent improvements of up to 17 F1 points over baselines.

Conclusion: Offers a promising solution for groups with limited NER training data and compute power by effectively leveraging few-shot learning and strategic data augmentation.

Abstract: Named Entity Recognition (NER) is a critical task that requires substantial annotated data, making it challenging in low-resource scenarios where label acquisition is expensive. While zero-shot and instruction-tuned approaches have made progress, they often fail to generalize to domain-specific entities and do not effectively utilize limited available data. We present a lightweight few-shot NER framework that addresses these challenges through two key innovations: (1) a new instruction tuning template with a simplified output format that combines principles from prior IT approaches to leverage the large context window of recent state-of-the-art LLMs; (2) introducing a strategic data augmentation technique that preserves entity information while paraphrasing the surrounding context, thereby expanding our training data without compromising semantic relationships. Experiments on benchmark datasets show that our method achieves performance comparable to state-of-the-art models on few-shot and zero-shot tasks, with our few-shot approach attaining an average F1 score of 80.1 on the CrossNER datasets. Models trained with our paraphrasing approach show consistent improvements in F1 scores of up to 17 points over baseline versions, offering a promising solution for groups with limited NER training data and compute power.

[104] AcademicEval: Live Long-Context LLM Benchmark

Haozhen Zhang, Tao Feng, Pengrui Han, Jiaxuan You

Main category: cs.CL

TL;DR: AcademicEval is a live benchmark using arXiv papers to evaluate LLMs on long-context academic writing tasks (Title, Abstract, Introduction, Related Work) without manual labeling, addressing label leakage issues and enabling flexible context lengths.

DetailsMotivation: Current long-context LLM benchmarks have limitations including rigid context length, labor-intensive annotation, and label leakage problems during training.

Method: Uses arXiv papers to create academic writing tasks, integrates expert-curated few-shot demonstrations from a co-author graph, and implements live evaluation to prevent label leakage.

Result: LLMs perform poorly on tasks with hierarchical abstraction levels and struggle with long few-shot demonstrations, highlighting benchmark challenges.

Conclusion: The benchmark reveals insights for enhancing LLMs’ long-context modeling capabilities and provides an effective evaluation framework for academic writing tasks.

Abstract: Large Language Models (LLMs) have recently achieved remarkable performance in long-context understanding. However, current long-context LLM benchmarks are limited by rigid context length, labor-intensive annotation, and the pressing challenge of label leakage issues during LLM training. Therefore, we propose \textsc{AcademicEval}, a live benchmark for evaluating LLMs over long-context generation tasks. \textsc{AcademicEval} adopts papers on arXiv to introduce several academic writing tasks with long-context inputs, \textit{i.e.}, \textsc{Title}, \textsc{Abstract}, \textsc{Introduction}, and \textsc{Related Work}, which cover a wide range of abstraction levels and require no manual labeling. Moreover, \textsc{AcademicEval} integrates high-quality and expert-curated few-shot demonstrations from a collected co-author graph to enable flexible context length. Especially, \textsc{AcademicEval} features an efficient live evaluation, ensuring no label leakage. We conduct a holistic evaluation on \textsc{AcademicEval}, and the results illustrate that LLMs perform poorly on tasks with hierarchical abstraction levels and tend to struggle with long few-shot demonstrations, highlighting the challenge of our benchmark. Through experimental analysis, we also reveal some insights for enhancing LLMs’ long-context modeling capabilities. Code is available at https://github.com/ulab-uiuc/AcademicEval

[105] Train for Truth, Keep the Skills: Binary Retrieval-Augmented Reward Mitigates Hallucinations

Tong Chen, Akari Asai, Luke Zettlemoyer, Hannaneh Hajishirzi, Faeze Brahman

Main category: cs.CL

TL;DR: Online reinforcement learning with binary retrieval-augmented reward reduces hallucinations in language models without degrading performance on other tasks.

DetailsMotivation: Language models often generate factually incorrect information (extrinsic hallucinations), and existing mitigation approaches degrade performance on open-ended generation and downstream tasks.

Method: Online reinforcement learning using a novel binary retrieval-augmented reward (RAR) that assigns reward of 1 only when output is entirely factually correct, and 0 otherwise.

Result: 39.3% reduction in hallucination rates for open-ended generation; 44.4% and 21.7% fewer incorrect answers on PopQA and GPQA respectively; learns calibrated abstention; maintains performance on instruction following, math, and code tasks.

Conclusion: Binary RAR effectively reduces hallucinations without performance degradation, outperforming both supervised training and continuous-reward RL baselines.

Abstract: Language models often generate factually incorrect information unsupported by their training data, a phenomenon known as extrinsic hallucination. Existing mitigation approaches often degrade performance on open-ended generation and downstream tasks, limiting their practical utility. We propose an online reinforcement learning method using a novel binary retrieval-augmented reward (RAR) to address this tradeoff. Unlike continuous reward schemes, our approach assigns a reward of one only when the model’s output is entirely factually correct, and zero otherwise. We evaluate our method on Qwen3 reasoning models across diverse tasks. For open-ended generation, binary RAR achieves a 39.3% reduction in hallucination rates, substantially outperforming both supervised training and continuous-reward RL baselines. In short-form question answering, the model learns calibrated abstention, strategically outputting “I don’t know” when faced with insufficient parametric knowledge. This yields 44.4% and 21.7% fewer incorrect answers on PopQA and GPQA, respectively. Crucially, these factuality gains come without performance degradation on instruction following, math, or code, whereas continuous-reward RL, despite improving factuality, induces quality regressions.

[106] Evaluating Medical LLMs by Levels of Autonomy: A Survey Moving from Benchmarks to Applications

Xiao Ye, Jacob Dineen, Zhaonan Li, Zhikun Xu, Weiyu Chen, Shijie Lu, Yuxi Huang, Ming Shen, Phu Tran, Ji-Eun Irene Yum, Muhammad Ali Khan, Muhammad Umar Afzal, Irbaz Bin Riaz, Ben Zhou

Main category: cs.CL

TL;DR: This survey reframes medical LLM evaluation through a levels-of-autonomy framework (L0-L3) to address the gap between benchmark scores and safe clinical use, proposing a structured approach for risk-aware evaluation.

DetailsMotivation: Medical LLMs achieve strong benchmark scores but transferring these to safe clinical workflows remains challenging. Current evaluation lacks alignment with real clinical risks and autonomy levels.

Method: Proposes a levels-of-autonomy lens (L0-L3) spanning informational tools to supervised agents, aligning existing benchmarks with permitted actions and associated risks at each level.

Result: Develops a level-conditioned blueprint for selecting metrics, assembling evidence, and reporting claims that links evaluation to oversight mechanisms.

Conclusion: By centering autonomy levels, the field can move beyond score-based claims toward credible, risk-aware evidence for real clinical implementation.

Abstract: Medical Large language models achieve strong scores on standard benchmarks; however, the transfer of those results to safe and reliable performance in clinical workflows remains a challenge. This survey reframes evaluation through a levels-of-autonomy lens (L0-L3), spanning informational tools, information transformation and aggregation, decision support, and supervised agents. We align existing benchmarks and metrics with the actions permitted at each level and their associated risks, making the evaluation targets explicit. This motivates a level-conditioned blueprint for selecting metrics, assembling evidence, and reporting claims, alongside directions that link evaluation to oversight. By centering autonomy, the survey moves the field beyond score-based claims toward credible, risk-aware evidence for real clinical use.

[107] Foundational Automatic Evaluators: Scaling Multi-Task Generative Evaluator Training for Reasoning-Centric Domains

Austin Xu, Xuan-Phi Nguyen, Yilun Zhou, Chien-Sheng Wu, Caiming Xiong, Shafiq Joty

Main category: cs.CL

TL;DR: FARE is a family of foundational automatic reasoning evaluators trained on 2.5M samples across five evaluation tasks, achieving state-of-the-art performance through data scaling and simple iterative SFT approach.

DetailsMotivation: To address the need for scalable evaluation during training and test-time by focusing on data scaling rather than complex methodology, as recent work has emphasized new techniques like RL over large-scale data development.

Method: Curated 2.5M samples across five evaluation tasks (pairwise, step-level, reference-free and reference-based verification, single rating) and trained FARE models (8B and 20B parameters) using iterative rejection-sampling supervised finetuning (SFT).

Result: FARE-8B challenges larger specialized RL-trained evaluators, FARE-20B sets new standard for open-source evaluators surpassing specialized 70B+ models. Achieves near-oracle performance on MATH as rerankers, improves RL-trained model performance by 14.1% as verifiers, and FARE-Code outperforms gpt-oss-20B by 65% on test-case quality evaluation.

Conclusion: Data scaling with simple SFT approach can produce highly effective evaluators that outperform more complex methods, demonstrating the importance of large-scale curated datasets for building foundational automatic reasoning evaluators.

Abstract: Finetuning specialized generative evaluators has emerged as a popular paradigm to meet the increasing demand for scalable evaluation during both training and test-time. However, recent work has largely focused on applying new methodology, such as reinforcement learning (RL), to training evaluators, shying away from large-scale, data-driven development. In this work, we focus on data scaling, curating a set of 2.5M samples spanning five unique evaluation tasks (pairwise, step-level, reference-free and reference-based verification, and single rating) and multiple domains focused on reasoning evaluation. With our data, we train Foundational Automatic Reasoning Evaluators (FARE), a family of 8B and 20B (with 3.6B active) parameter evaluators, with a simple iterative rejection-sampling supervised finetuning (SFT) approach. FARE-8B challenges larger specialized RL-trained evaluators and FARE-20B sets the new standard for open-source evaluators, surpassing specialized 70B+ evaluators. Beyond static benchmarks, we evaluate FARE in real-world tasks: As inference-time rerankers, FARE-20B achieves near-oracle performance on MATH. As verifiers in RL training, FARE improves the downstream RL-trained model performance by up to 14.1% vs. string-matching verifiers. When initialized from FARE, a continually-finetuned FARE-Code outperforms gpt-oss-20B by 65% on evaluating test-case quality.

[108] Enterprise Deep Research: Steerable Multi-Agent Deep Research for Enterprise Analytics

Akshara Prabhakar, Roshan Ram, Zixiang Chen, Silvio Savarese, Frank Wang, Caiming Xiong, Huan Wang, Weiran Yao

Main category: cs.CL

TL;DR: EDR is a multi-agent system that transforms unstructured enterprise data into actionable insights through specialized agents for planning, search, visualization, and reflection, outperforming state-of-the-art systems on open-ended benchmarks.

DetailsMotivation: Enterprises face challenges in converting growing unstructured data into coherent insights, with existing autonomous agents struggling with domain-specific nuances, intent alignment, and enterprise integration.

Method: Multi-agent system with Master Planning Agent for query decomposition, four specialized search agents (General, Academic, GitHub, LinkedIn), MCP-based tool ecosystem for NL2SQL and file analysis, Visualization Agent, and reflection mechanism with optional human-in-the-loop guidance.

Result: Outperforms state-of-the-art agentic systems on DeepResearch Bench and DeepConsult benchmarks without human steering, enabling automated report generation and real-time streaming on internal datasets.

Conclusion: EDR framework and benchmark trajectories released to advance multi-agent reasoning research, demonstrating effective enterprise deployment for automated insight generation.

Abstract: As information grows exponentially, enterprises face increasing pressure to transform unstructured data into coherent, actionable insights. While autonomous agents show promise, they often struggle with domain-specific nuances, intent alignment, and enterprise integration. We present Enterprise Deep Research (EDR), a multi-agent system that integrates (1) a Master Planning Agent for adaptive query decomposition, (2) four specialized search agents (General, Academic, GitHub, LinkedIn), (3) an extensible MCP-based tool ecosystem supporting NL2SQL, file analysis, and enterprise workflows, (4) a Visualization Agent for data-driven insights, and (5) a reflection mechanism that detects knowledge gaps and updates research direction with optional human-in-the-loop steering guidance. These components enable automated report generation, real-time streaming, and seamless enterprise deployment, as validated on internal datasets. On open-ended benchmarks including DeepResearch Bench and DeepConsult, EDR outperforms state-of-the-art agentic systems without any human steering. We release the EDR framework and benchmark trajectories to advance research on multi-agent reasoning applications. Code at https://github.com/SalesforceAIResearch/enterprise-deep-research and Dataset at https://huggingface.co/datasets/Salesforce/EDR-200

[109] The Moral Foundations Reddit Corpus

Jackson Trager, Alireza S. Ziabari, Elnaz Rahmati, Aida Mostafazadeh Davani, Preni Golazizian, Farzan Karimi-Malekabadi, Ali Omrani, Zhihe Li, Brendan Kennedy, Nils Karl Reimer, Melissa Reyes, Kelsey Cheng, Mellow Wei, Christina Merrifield, Arta Khosravi, Evans Alvarez, Morteza Dehghani

Main category: cs.CL

TL;DR: This paper introduces the Moral Foundations Reddit Corpus - a dataset of 16,123 Reddit comments annotated for 8 moral sentiment categories using Moral Foundations Theory, and evaluates LLMs against fine-tuned BERT models on moral sentiment classification.

DetailsMotivation: Existing moral sentiment datasets are limited to Twitter, and strong performance in subjective moral sentiment tasks requires large hand-annotated datasets. There's a need for diverse moral rhetoric data to improve understanding of moral framing effects.

Method: Created a corpus of 16,123 English Reddit comments from 12 subreddits, annotated by at least three trained annotators for 8 moral sentiment categories based on updated Moral Foundations Theory. Evaluated LLMs (Llama3-8B, Ministral-8B) in zero-shot, few-shot, and PEFT settings against fine-tuned BERT models.

Result: LLMs continue to lag behind fine-tuned encoder-only models like BERT on this subjective moral sentiment classification task, despite advances in large language models.

Conclusion: There is an ongoing need for human-annotated moral corpora for AI alignment evaluation, as current LLMs still underperform fine-tuned models on subjective moral sentiment tasks.

Abstract: Moral framing and sentiment can affect a variety of online and offline behaviors, including donation, environmental action, political engagement, and protest. Various computational methods in Natural Language Processing (NLP) have been used to detect moral sentiment from textual data, but achieving strong performance in such subjective tasks requires large, hand-annotated datasets. Previous corpora annotated for moral sentiment have proven valuable, and have generated new insights both within NLP and across the social sciences, but have been limited to Twitter. To facilitate improving our understanding of the role of moral rhetoric, we present the Moral Foundations Reddit Corpus, a collection of 16,123 English Reddit comments that have been curated from 12 distinct subreddits, hand-annotated by at least three trained annotators for 8 categories of moral sentiment (i.e., Care, Proportionality, Equality, Purity, Authority, Loyalty, Thin Morality, Implicit/Explicit Morality) based on the updated Moral Foundations Theory (MFT) framework. We evaluate baselines using large language models (Llama3-8B, Ministral-8B) in zero-shot, few-shot, and PEFT settings, comparing their performance to fine-tuned encoder-only models like BERT. The results show that LLMs continue to lag behind fine-tuned encoders on this subjective task, underscoring the ongoing need for human-annotated moral corpora for AI alignment evaluation. Keywords: moral sentiment annotation, moral values, moral foundations theory, multi-label text classification, large language models, benchmark dataset, evaluation and alignment resource

[110] Consistency is Key: Disentangling Label Variation in Natural Language Processing with Intra-Annotator Agreement

Gavin Abercrombie, Tanvi Dinkar, Amanda Cercas Curry, Verena Rieser, Dirk Hovy

Main category: cs.CL

TL;DR: The paper argues for measuring intra-annotator agreement (stability over time) in addition to inter-annotator agreement to assess label reliability in NLP tasks, finding that intra-annotator agreement is rarely reported despite annotators providing inconsistent responses around 25% of the time.

DetailsMotivation: Current NLP practice primarily uses inter-annotator agreement to measure label reliability, but this overlooks the importance of measuring annotator consistency over time through intra-annotator agreement.

Method: Conducted a systematic review of current practices and performed exploratory annotation experiments across four different NLP tasks to investigate relationships between agreement measures and perceptions of subjectivity/ambiguity.

Result: Found that intra-annotator agreement is rarely reported in the field, and experiments revealed annotators provide inconsistent responses around 25% of the time across different NLP tasks.

Conclusion: Intra-annotator agreement should be calculated as important quality control that provides insights into why annotators disagree, complementing traditional inter-annotator agreement measures.

Abstract: We commonly use agreement measures to assess the utility of judgements made by human annotators in Natural Language Processing (NLP) tasks. While inter-annotator agreement is frequently used as an indication of label reliability by measuring consistency between annotators, we argue for the additional use of intra-annotator agreement to measure label stability (and annotator consistency) over time. However, in a systematic review, we find that the latter is rarely reported in this field. Calculating these measures can act as important quality control and could provide insights into why annotators disagree. We conduct exploratory annotation experiments to investigate the relationships between these measures and perceptions of subjectivity and ambiguity in text items, finding that annotators provide inconsistent responses around 25% of the time across four different NLP tasks.

[111] Synthetic Dataset for Evaluating Complex Compositional Knowledge for Natural Language Inference

Sushma Anand Akoju, Robert Vacareanu, Haris Riaz, Eduardo Blanco, Mihai Surdeanu

Main category: cs.CL

TL;DR: The paper introduces SICCK, a synthetic dataset for testing NLI models’ understanding of compositional logic, and finds that models struggle with negation and quantifiers even after fine-tuning.

DetailsMotivation: To investigate how well Natural Language Inference models understand compositional logic, particularly with logical operators like quantifiers and negation.

Method: Created SICCK dataset by modifying 15 SICK examples using logical modifiers (quantifiers, negation), annotated with NL entailment rules, and tested NLI models in zero-shot and fine-tuned settings.

Result: NLI models performed poorly in zero-shot setting, especially for negation and existential quantifiers. After fine-tuning, models continued to struggle with negation, existential, and universal modifiers.

Conclusion: Current NLI models have significant limitations in handling compositional logical structures, particularly with negation and quantifiers, even after targeted training.

Abstract: We introduce a synthetic dataset called Sentences Involving Complex Compositional Knowledge (SICCK) and a novel analysis that investigates the performance of Natural Language Inference (NLI) models to understand compositionality in logic. We produce 1,304 sentence pairs by modifying 15 examples from the SICK dataset (Marelli et al., 2014). To this end, we modify the original texts using a set of phrases - modifiers that correspond to universal quantifiers, existential quantifiers, negation, and other concept modifiers in Natural Logic (NL) (MacCartney, 2009). We use these phrases to modify the subject, verb, and object parts of the premise and hypothesis. Lastly, we annotate these modified texts with the corresponding entailment labels following NL rules. We conduct a preliminary verification of how well the change in the structural and semantic composition is captured by neural NLI models, in both zero-shot and fine-tuned scenarios. We found that the performance of NLI models under the zero-shot setting is poor, especially for modified sentences with negation and existential quantifiers. After fine-tuning this dataset, we observe that models continue to perform poorly over negation, existential and universal modifiers.

[112] Exploration of Marker-Based Approaches in Argument Mining through Augmented Natural Language

Nilmadhab Das, Vishal Choudhary, V. Vijaya Saradhi, Ashish Anand

Main category: cs.CL

TL;DR: argTANL is a generative end-to-end framework that extracts argumentative components and relations by converting them into label-augmented text, with marker-enhanced variants showing superior performance.

DetailsMotivation: Existing Argument Mining approaches use multi-step pipelines or dependency parsing, lacking end-to-end solutions that jointly extract argument components and relations efficiently.

Method: Proposes argTANL framework that converts argument structures into Augmented Natural Language (ANL) text. Introduces two variants: ME-argTANL with marker enhancement and specialized marker-based fine-tuning.

Result: Extensive experiments on three standard benchmarks show ME-argTANL achieves superior performance compared to existing methods.

Conclusion: The generative paradigm with marker enhancement provides an effective end-to-end solution for argument mining, outperforming traditional approaches.

Abstract: Argument Mining (AM) involves identifying and extracting Argumentative Components (ACs) and their corresponding Argumentative Relations (ARs). Most of the prior works have broken down these tasks into multiple sub-tasks. Existing end-to-end setups primarily use the dependency parsing approach. This work introduces a generative paradigm-based end-to-end framework argTANL. argTANL frames the argumentative structures into label-augmented text, called Augmented Natural Language (ANL). This framework jointly extracts both ACs and ARs from a given argumentative text. Additionally, this study explores the impact of Argumentative and Discourse markers on enhancing the model’s performance within the proposed framework. Two distinct frameworks, Marker-Enhanced argTANL (ME-argTANL) and argTANL with specialized Marker-Based Fine-Tuning, are proposed to achieve this. Extensive experiments are conducted on three standard AM benchmarks to demonstrate the superior performance of the ME-argTANL.

[113] Familiarity-Aware Evidence Compression for Retrieval-Augmented Generation

Dongwon Jung, Qin Liu, Tenghao Huang, Ben Zhou, Muhao Chen

Main category: cs.CL

TL;DR: FaviComp is a training-free evidence compression technique that makes retrieved evidence more familiar to target models in RAG systems, improving accuracy by up to 28.1% while maintaining high compression rates.

DetailsMotivation: RAG systems struggle with inconsistent and irrelevant information that distracts language models, and compressed evidence may be unfamiliar to target models, reducing effectiveness.

Method: Proposed FaviComp - a training-free evidence compression technique that enhances familiarity of compressed evidence to target models while integrating parametric knowledge.

Result: Outperforms recent evidence compression baselines across multiple open-domain QA datasets, improving accuracy by up to 28.1% with high compression rates.

Conclusion: FaviComp effectively integrates both parametric and non-parametric knowledge during evidence compression and consistently improves RAG performance.

Abstract: Retrieval-augmented generation (RAG) improves large language models (LMs) by incorporating non-parametric knowledge through evidence retrieved from external sources. However, it often struggles to cope with inconsistent and irrelevant information that can distract the LM from its tasks, especially when multiple evidence pieces are required. While compressing the retrieved evidence with a compression model aims to address this issue, the compressed evidence may still be unfamiliar to the target model used for downstream tasks, potentially failing to utilize the evidence effectively. We propose FaviComp (Familarity-Aware Evidence Compression), a novel training-free evidence compression technique that makes retrieved evidence more familiar to the target model, while seamlessly integrating parametric knowledge from the model. Experimental results show that FaviComp consistently outperforms most recent evidence compression baselines across multiple open-domain QA datasets, improving accuracy by up to 28.1% while achieving high compression rates. Additionally, we demonstrate the effective integration of both parametric and non-parametric knowledge during evidence compression.

[114] LEME: Open Large Language Models for Ophthalmology with Advanced Reasoning and Clinical Validation

Hyunjae Kim, Xuguang Ai, Sahana Srinivasan, Aidan Gilson, Maxwell B. Singer, Krithi Pushpanathan, Qianqian Xie, Jungwoo Park, Serina Applebaum, Gabriel Dawei Yang, Minjie Zou, David Ziyou Chen, Ke Zou, Soshian Sarrafpour, Ji Liu, Yu Yin, Jimin Huang, Quang Ngoc Nguyen, Erping Long, Peixing Wan, Dianbo Liu, Richard Hintz, W. Jim Zheng, Sophia Y. Wang, Lucila Ohno-Machado, Hua Xu, Ron A. Adelman, Luciano V. Del Priore, Yih-Chung Tham, Qingyu Chen

Main category: cs.CL

TL;DR: LEME is an open-source ophthalmology-specific LLM based on Llama2 70B, fine-tuned with 127K ophthalmology training instances. It outperformed other LLMs in various ophthalmology tasks including abstract completion, fill-in-the-blank, MCQ, and clinical QA.

DetailsMotivation: Ophthalmology-specific LLMs are scarce and underexplored despite LLMs' potential to revolutionize healthcare. There's a need for specialized models in ophthalmology to improve clinical task execution and research collaboration.

Method: Pre-trained on Llama2 70B framework and fine-tuned with ~127,000 non-copyrighted ophthalmology training instances from case reports, abstracts, and study materials. Evaluated against 8 other LLMs using internal and external validation tasks with metrics including Rouge-L scores, accuracy, and expert evaluation.

Result: LEME consistently outperformed other models in internal validations (Rouge-L: 0.20-0.82) and external validations. It ranked second in MCQ accuracy (0.68) and scored highest in EHR summarization and clinical QA (4.24-4.83/5 for correctness, completeness, readability).

Conclusion: LEME represents a breakthrough in open-source ophthalmology-specific LLMs, offering potential to revolutionize clinical task execution while democratizing research collaboration through robust fine-tuning and non-copyrighted data.

Abstract: Large Language Models (LLMs) are poised to revolutionize healthcare. Ophthalmology-specific LLMs remain scarce and underexplored. We introduced an open-source, specialized LLM for ophthalmology, termed Language Enhanced Model for Eye (LEME). LEME was initially pre-trained on the Llama2 70B framework and further fine-tuned with a corpus of ~127,000 non-copyrighted training instances curated from ophthalmology-specific case reports, abstracts, and open-source study materials. We benchmarked LEME against eight other LLMs, namely, GPT-3.5, GPT-4, three Llama2 models (7B, 13B, 70B), PMC-LLAMA 13B, Meditron 70B, and EYE-Llama (another ophthalmology-specific LLM). Evaluations included four internal validation tasks: abstract completion, fill-in-the-blank, multiple-choice questions (MCQ), and short-answer QA. External validation tasks encompassed long-form QA, MCQ, patient EHR summarization, and clinical QA. Evaluation metrics included Rouge-L scores, accuracy, and expert evaluation of correctness, completeness, and readability. In internal validations, LEME consistently outperformed its counterparts, achieving Rouge-L scores of 0.20 in abstract completion (all p<0.05), 0.82 in fill-in-the-blank (all p<0.0001), and 0.22 in short-answer QA (all p<0.0001, except versus GPT-4). In external validations, LEME excelled in long-form QA with a Rouge-L of 0.19 (all p<0.0001), ranked second in MCQ accuracy (0.68; all p<0.0001), and scored highest in EHR summarization and clinical QA (ranging from 4.24 to 4.83 out of 5 for correctness, completeness, and readability). LEME’s emphasis on robust fine-tuning and the use of non-copyrighted data represents a breakthrough in open-source ophthalmology-specific LLMs, offering the potential to revolutionize execution of clinical tasks while democratizing research collaboration.

[115] Simplicity Prevails: Rethinking Negative Preference Optimization for LLM Unlearning

Chongyu Fan, Jiancheng Liu, Licong Lin, Jinghan Jia, Ruiqi Zhang, Song Mei, Sijia Liu

Main category: cs.CL

TL;DR: SimNPO is a simple yet effective LLM unlearning framework that removes reference model bias in NPO, addressing uneven optimization allocation and gradient weight smoothing issues for better unlearning performance.

DetailsMotivation: Existing LLM unlearning methods like NPO suffer from reference model bias, which compromises effectiveness by causing uneven optimization across forget data and ineffective gradient smoothing during early unlearning stages.

Method: Proposes SimNPO framework that removes reliance on reference model through simple preference optimization, providing advantages analyzed via mixtures of Markov chains.

Result: Extensive experiments on TOFU and MUSE benchmarks show SimNPO’s efficacy and robustness against relearning attacks, outperforming existing methods.

Conclusion: Simplicity in removing reference model dependence benefits LLM unlearning, with SimNPO providing a technically-grounded optimization framework that addresses critical issues in current state-of-the-art approaches.

Abstract: This work studies the problem of large language model (LLM) unlearning, aiming to remove unwanted data influences (e.g., copyrighted or harmful content) while preserving model utility. Despite the increasing demand for unlearning, a technically-grounded optimization framework is lacking. Gradient ascent (GA)-type methods, though widely used, are suboptimal as they reverse the learning process without controlling optimization divergence (i.e., deviation from the pre-trained state), leading to risks of over-forgetting and potential model collapse. Negative preference optimization (NPO) has been proposed to address this issue and is considered one of the state-of-the-art LLM unlearning approaches. In this work, we revisit NPO and identify another critical issue: reference model bias. This bias arises from using the reference model (i.e., the model prior to unlearning) to evaluate the unlearning success, which can compromise NPO’s effectiveness. Specifically, it leads to (a) uneven allocation of optimization power across forget data with varying difficulty levels and (b) ineffective gradient weight smoothing during the early stages of unlearning optimization. To overcome these challenges, we propose a simple yet effective unlearning optimization framework, called SimNPO, showing that `simplicity’ in removing the reliance on a reference model (through the lens of simple preference optimization) benefits unlearning. We provide deeper insights into SimNPO’s advantages through an analysis based on mixtures of Markov chains. Extensive experiments further validate SimNPO’s efficacy on benchmarks like TOFU and MUSE, as well as its robustness against relearning attacks. Codes are available at https://github.com/OPTML-Group/Unlearn-Simple.

[116] H3Fusion: Helpful, Harmless, Honest Fusion of Aligned LLMs

Selim Furkan Tekin, Fatih Ilhan, Tiansheng Huang, Sihao Hu, Yichang Xu, Zachary Yahn, Ling Liu

Main category: cs.CL

TL;DR: H3Fusion introduces a mixture-of-experts-based fusion mechanism for LLM alignment that models alignment as controllable drift in representation subspace, achieving better performance than individual alignment models and ensemble approaches.

DetailsMotivation: Current LLM alignment methods struggle to find a single point in the model's representation subspace that simultaneously satisfies all three alignment properties (helpful, harmless, honest), requiring a more sophisticated approach.

Method: Uses mixture-of-experts (MoE)-based fusion with controllable drift modeling, drift-regularization loss, dual objective formulation for embedding distances, and gating loss to canalize expert activations.

Result: Outperforms individually aligned models by 11.37%, shows 13.77% stronger robustness than state-of-the-art LLM ensemble approaches, and 6.18% better than model-merging approaches across three benchmark datasets.

Conclusion: H3Fusion effectively addresses the challenge of multi-dimensional LLM alignment through subspace fusion and controllable drift mechanisms, providing superior performance across all three alignment dimensions.

Abstract: The alignment of pre-trained LLMs continues to draw significant attention from both industry and academia, aiming to ensure responses that are helpful, harmless, and honest. However, identifying a point in the model’s representation subspace that simultaneously satisfies all these properties remains challenging. H3Fusion addresses this challenge by introducing a mixture-of-experts (MoE)-based fusion mechanism that models alignment as a controllable drift within the subspace, guided by a drift-regularization loss to balance competing alignment dimensions. Furthermore, we formulate the alignment by finding a dual objective of harnessing the distance of generated embeddings and alignment embeddings, and introduce a gating loss by canalizing the activations on the contributing experts. Extensive evaluations of three benchmark datasets show that H3Fusion is more helpful, less harmful, and more honest in three aspects: it outperforms each individually aligned model by 11.37%, and provides stronger robustness compared to the state-of-the-art LLM ensemble approaches by 13.77% and model-merging approaches by 6.18%. Code is available at https://github.com/sftekin/h3fusion.

[117] Auto-Prompt Generation is Not Robust: Prompt Optimization Driven by Pseudo Gradient

Zeru Shi, Zhenting Wang, Yongye Su, Weidi Luo, Hang Gao, Fan Yang, Ruixiang Tang, Yongfeng Zhang

Main category: cs.CL

TL;DR: PertBench benchmark reveals auto-prompting methods are vulnerable to input perturbations. PGO framework uses perturbation types as pseudo-gradients to generate more robust prompts, outperforming existing methods.

DetailsMotivation: Current automatic prompt generation methods lack robustness evaluation and are vulnerable to minor input perturbations, leading to significant performance degradation.

Method: Proposed PGO, a gradient-free prompt generation framework that uses perturbation types as pseudo-gradient signals to guide LLMs in producing robust prompts under noisy conditions.

Result: Extensive experiments show PGO consistently outperforms previous methods across diverse tasks and multiple LLMs in maintaining performance under input perturbations.

Conclusion: Explicit emphasis on robustness under noisy conditions is crucial for reliable auto-prompting, and PGO provides an effective solution for generating perturbation-resistant prompts.

Abstract: While automatic prompt generation methods have recently received significant attention, their robustness remains poorly understood. In this paper, we introduce PertBench, a comprehensive benchmark dataset that includes a wide range of input perturbations, designed to systematically evaluate the robustness of current auto-prompting techniques. Our analysis reveals substantial vulnerabilities in existing prompt generation strategies, where even minor modifications to the prompt can lead to significant differences in model output. To address this issue, we propose PGO, a gradient-free prompt generation framework that leverages perturbation types as pseudo-gradient signals to guide LLMs in producing more robust prompts. In contrast to existing methods that assess prompt quality only on clean, well-structured inputs, our approach explicitly emphasizes robustness under noisy and perturbed conditions. Extensive experiments across diverse tasks and multiple LLMs show PGO consistently outperforms previous methods in maintaining performance under input perturbations.

[118] Tracing Partisan Bias to Its Emotional Fingerprints: A Computational Approach to Mitigation

Junjie Liu, Xi Luo, Sirong Wu, Gengchen Sun, Yuhui Deng

Main category: cs.CL

TL;DR: This paper introduces a framework that identifies media bias through emotional fingerprints in news texts using the VAD framework, and proposes NeutraSum model to neutralize these emotional patterns in summaries.

DetailsMotivation: To address media bias by focusing on its linguistic roots in emotional language rather than abstract political stances, treating bias as quantifiable emotional fingerprints.

Method: Uses Valence-Arousal-Dominance (VAD) framework to measure emotional fingerprints in news texts, analyzes Allsides dataset, and develops NeutraSum model to neutralize identified emotional patterns in summaries.

Result: Analysis confirmed distinct emotional fingerprints for left, center, and right-leaning media. NeutraSum successfully erased partisan emotional fingerprints from summaries, achieving lower emotional bias scores than other models.

Conclusion: The work pioneers bias mitigation by shifting focus from political labels to addressing the emotional encoding of partisan bias in language, providing an evidence-driven approach to media bias reduction.

Abstract: This study introduces a novel framework for analysing and mitigating media bias by tracing partisan stances to their linguistic roots in emotional language. We posit that partisan bias is not merely an abstract stance but materialises as quantifiable ’emotional fingerprints’ within news texts. These fingerprints are systematically measured using the Valence-Arousal-Dominance (VAD) framework, allowing us to decode the affective strategies behind partisan framing. Our analysis of the Allsides dataset confirms this hypothesis, revealing distinct and statistically significant emotional fingerprints for left, centre, and right-leaning media. Based on this evidence-driven approach, we then propose a computational approach to mitigation through NeutraSum, a model designed to neutralise these identified emotional patterns. By explicitly targeting the VAD characteristics of biased language, NeutraSum generates summaries that are not only coherent but also demonstrably closer to an emotionally neutral baseline. Experimental results validate our framework: NeutraSum successfully erases the partisan emotional fingerprints from its summaries, achieving a demonstrably lower emotional bias score than other models. This work pioneers a new path for bias mitigation, shifting the focus from treating symptoms (political labels) to addressing the cause: the emotional encoding of partisan bias in language.

[119] Consistency of Responses and Continuations Generated by Large Language Models on Social Media

Wenlu Fan, Yuqi Zhu, Bin Wang, Wentao Xu

Main category: cs.CL

TL;DR: LLMs maintain high semantic coherence but moderate negative emotions in social media contexts, showing preference for neutral rational emotions and reduced emotional intensity compared to human-authored content.

DetailsMotivation: To investigate how LLMs handle emotional content and maintain semantic relationships in social media contexts, particularly their emotional consistency and semantic coherence.

Method: Analyzed emotional transitions, intensity patterns, and semantic consistency using continuation and response tasks with three open-source models (Gemma, Llama3, Llama3.3) and one commercial model (Claude) on climate change discussions from Twitter and Reddit.

Result: LLMs maintain high semantic coherence but show strong tendency to moderate negative emotions, converting them to neutral or positive emotions. They generate responses with reduced emotional intensity and prefer neutral rational emotions compared to human-authored content.

Conclusion: The findings provide insights into LLMs’ emotion and semantic processing capabilities, which are significant for their deployment in social media environments and human-computer interaction design.

Abstract: Large Language Models (LLMs) demonstrate remarkable capabilities in text generation, yet their emotional consistency and semantic coherence in social media contexts remain insufficiently understood. This study investigates how LLMs handle emotional content and maintain semantic relationships through continuation and response tasks using three open-source models: Gemma, Llama3 and Llama3.3 and one commercial Model:Claude. By analyzing climate change discussions from Twitter and Reddit, we examine emotional transitions, intensity patterns, and semantic consistency between human-authored and LLM-generated content. Our findings reveal that while both models maintain high semantic coherence, they exhibit distinct emotional patterns: these models show a strong tendency to moderate negative emotions. When the input text carries negative emotions such as anger, disgust, fear, or sadness, LLM tends to generate content with more neutral emotions, or even convert them into positive emotions such as joy or surprise. At the same time, we compared the LLM-generated content with human-authored content. The four models systematically generated responses with reduced emotional intensity and showed a preference for neutral rational emotions in the response task. In addition, these models all maintained a high semantic similarity with the original text, although their performance in the continuation task and the response task was different. These findings provide deep insights into the emotion and semantic processing capabilities of LLM, which are of great significance for its deployment in social media environments and human-computer interaction design.

[120] Evolving LLMs’ Self-Refinement Capability via Iterative Preference Optimization

Yongcheng Zeng, Xinyu Cui, Xuanfa Jin, Qirui Mi, Guoqing Liu, Zexu Sun, Mengyue Yang, Dong Li, Weiyu Ma, Ning Yang, Jian Zhao, Jianye Hao, Haifeng Zhang, Jun Wang

Main category: cs.CL

TL;DR: EVOLVE framework enables LLMs to develop Self-Refinement capability through iterative training and inference optimization, allowing models to improve their own responses and achieve state-of-the-art performance.

DetailsMotivation: Current LLMs lack inherent Self-Refinement ability and may degrade response quality when attempting self-revision, limiting their potential for self-improvement.

Method: Proposed EVOLVE framework with synergistic optimization: training methods to activate Self-Refinement capability and inference strategies to enhance and utilize it while generating training data.

Result: Llama-3.1-8B with evolved Self-Refinement surpasses GPT-4o on AlpacaEval 2 (62.3% length-controlled, 63.3% raw win rates) and Arena-Hard (50.3%), and improves performance on mathematical reasoning benchmarks (GSM8K, MATH).

Conclusion: Self-Refinement can be systematically developed through the EVOLVE framework, enabling models to refine their own responses and achieve broader self-improvement across various tasks.

Abstract: Self-Refinement refers to a model’s ability to revise its own responses to produce improved outputs. This capability can also serve as a fundamental mechanism for Self-Improvement, for example, by reconstructing datasets with refined results to enhance intrinsic model performance. However, our comprehensive experiments reveal that large language models (LLMs) show no clear evidence of inherent Self-Refinement and may even experience response quality degradation after Self-Refinement. To address this issue, we propose EVOLVE, a simple and effective framework for eliciting and tracking the evolution of Self-Refinement through iterative training. We first explore optimization methods during training to activate the model’s Self-Refinement capability. Then, at inference, we investigate various generation strategies to further enhance and utilize Self-Refinement while supplying the necessary data for training. Through synergistic optimization of training and inference stages, we continually evolve the model’s Self-Refinement ability, enabling it to better refine its own responses. Moreover, we demonstrate the potential of leveraging Self-Refinement to achieve broader Self-Improvement of intrinsic model abilities. Experiments show that the evolved Self-Refinement ability enables the Llama-3.1-8B base model to surpass GPT-4o, achieving 62.3% length-controlled and 63.3% raw win rates on AlpacaEval 2, and 50.3% on Arena-Hard. It also generalizes effectively to out-of-domain reasoning tasks, improving performance on mathematical reasoning benchmarks such as GSM8K and MATH.

[121] Hope vs. Hate: Understanding User Interactions with LGBTQ+ News Content in Mainstream US News Media through the Lens of Hope Speech

Jonathan Pofcher, Christopher M. Homan, Randall Sell, Ashiqur R. KhudaBukhsh

Main category: cs.CL

TL;DR: Analysis of YouTube comments on LGBTQ+ news videos reveals political bias in content rating, with LLMs aligning more with liberal perspectives.

DetailsMotivation: To understand how users engage with LGBTQ+ news content on YouTube and examine the influence of political beliefs on content rating.

Method: Analyzed 1.4M comments from 3,161 YouTube news videos, created a fine-grained hope speech classifier, conducted annotation study with 3,750 instances using diverse political representation, and assessed zero-shot LLM performance.

Result: Found strong association between rater political beliefs and content rating, models trained on individual political beliefs show significant disagreement in real-world settings, and LLMs align more with liberal raters.

Conclusion: Political beliefs significantly influence how content about marginalized communities is rated, highlighting challenges in developing unbiased content moderation systems.

Abstract: This paper makes three contributions. First, via a substantial corpus of 1,419,047 comments posted on 3,161 YouTube news videos of major US cable news outlets, we analyze how users engage with LGBTQ+ news content. Our analyses focus both on positive and negative content. In particular, we construct a fine-grained hope speech classifier that detects positive (hope speech), negative, neutral, and irrelevant content. Second, in consultation with a public health expert specializing on LGBTQ+ health, we conduct an annotation study with a balanced and diverse political representation and release a dataset of 3,750 instances with fine-grained labels and detailed annotator demographic information. Finally, beyond providing a vital resource for the LGBTQ+ community, our annotation study and subsequent in-the-wild assessments reveal (1) strong association between rater political beliefs and how they rate content relevant to a marginalized community; (2) models trained on individual political beliefs exhibit considerable in-the-wild disagreement; and (3) zero-shot large language models (LLMs) align more with liberal raters.

[122] Large Language Diffusion Models

Shen Nie, Fengqi Zhu, Zebin You, Xiaolu Zhang, Jingyang Ou, Jun Hu, Jun Zhou, Yankai Lin, Ji-Rong Wen, Chongxuan Li

Main category: cs.CL

TL;DR: LLaDA is a diffusion-based language model that challenges the assumption that LLM capabilities require autoregressive models, demonstrating competitive performance across various tasks.

DetailsMotivation: To challenge the common belief that large language model capabilities inherently depend on autoregressive models by exploring diffusion models as an alternative approach.

Method: LLaDA uses a diffusion model with forward data masking and reverse generation processes, parameterized by a Transformer to predict masked tokens, trained under pre-training and supervised fine-tuning paradigms.

Result: LLaDA performs comparably to autoregressive baselines across general tasks, math, and code, with LLaDA 8B being competitive with LLaMA3 8B in in-context learning and instruction-following. It also addresses the reversal curse, surpassing GPT-4o in reversal poem completion.

Conclusion: Diffusion models show promise for language modeling at scale and challenge the assumption that core LLM capabilities inherently depend on autoregressive models.

Abstract: The capabilities of large language models (LLMs) are widely regarded as relying on autoregressive models (ARMs). We challenge this notion by introducing LLaDA, a diffusion model trained from scratch under the pre-training and supervised fine-tuning (SFT) paradigm. LLaDA employs a forward data masking process and a reverse generation process, parameterized by a Transformer to predict masked tokens. It provides a principled generative approach for probabilistic inference by optimizing a likelihood lower bound. Across extensive benchmarks on general tasks, math, code, and so on, LLaDA demonstrates strong scalability and performs comparably to our self-constructed ARM baselines. Remarkably, LLaDA 8B is competitive with strong LLMs like LLaMA3 8B in in-context learning and, after SFT, exhibits impressive instruction-following abilities in case studies such as multi-turn dialogue. Moreover, LLaDA addresses the reversal curse, surpassing GPT-4o in a reversal poem completion task. Our findings show the promise of diffusion models for language modeling at scale and challenge the common assumption that core LLM capabilities discussed above inherently depend on ARMs. Project page and codes: https://ml-gsai.github.io/LLaDA-demo/.

[123] GRIFFIN: Effective Token Alignment for Faster Speculative Decoding

Shijing Hu, Jingyang Li, Xingyu Xie, Zhihui Lu, Kim-Chuan Toh, Pan Zhou

Main category: cs.CL

TL;DR: GRIFFIN is a novel speculative decoding framework that addresses token misalignment issues in LLM inference through token-alignable training and draft models, achieving 8% acceptance length improvement and 7% speedup over state-of-the-art methods.

DetailsMotivation: Existing speculative decoding methods struggle with token misalignment between training and decoding phases, which limits their performance in accelerating LLM inference.

Method: Proposes GRIFFIN framework with token-alignable training strategy using loss masking to exclude misaligned tokens, and token-alignable draft model that introduces input tokens to correct feature inconsistencies.

Result: Experiments on LLaMA, Vicuna, Qwen and Mixtral models show average 8% acceptance length improvement and over 7% speedup ratio, outperforming current state-of-the-art speculative decoding methods.

Conclusion: GRIFFIN effectively mitigates token misalignment in speculative decoding, providing significant performance improvements for LLM inference acceleration.

Abstract: Speculative decoding accelerates inference in large language models (LLMs) by generating multiple draft tokens simultaneously. However, existing methods often struggle with token misalignment between the training and decoding phases, limiting their performance. To address this, we propose GRIFFIN, a novel framework that incorporates a token-alignable training strategy and a token-alignable draft model to mitigate misalignment. The training strategy employs a loss masking mechanism to exclude highly misaligned tokens during training, preventing them from negatively impacting the draft model’s optimization. The token-alignable draft model introduces input tokens to correct inconsistencies in generated features. Experiments on LLaMA, Vicuna, Qwen and Mixtral models demonstrate that GRIFFIN achieves an average acceptance length improvement of over 8% and a speedup ratio exceeding 7%, outperforming current speculative decoding state-of-the-art methods. Our code and GRIFFIN’s draft models are released publicly in https://github.com/hsj576/GRIFFIN.

[124] A Knapsack by Any Other Name: Presentation impacts LLM performance on NP-hard problems

Alex Duchnowski, Ellie Pavlick, Alexander Koller

Main category: cs.CL

TL;DR: LLMs perform better on textbook-style NP-hard problems than real-life or inverted versions, showing they lack robust reasoning and depend heavily on training data.

DetailsMotivation: To investigate how problem presentation affects LLMs' ability to solve optimization problems and test their generalization capabilities.

Method: Created EHOP dataset with NP-hard problems in three formats: textbook versions, real-life scenarios, and inverted rules. Tested state-of-the-art LLMs with multiple prompting strategies.

Result: LLMs systematically solve textbook problems more accurately than real-life and inverted versions. Reasoning models show high variance across problem presentations.

Conclusion: LLMs are heavily dependent on training data and struggle to generalize to novel problems, lacking truly robust reasoning mechanisms.

Abstract: To investigate the effect of problem presentation on LLMs’ ability to solve optimization problems, we introduce the dataset of Everyday Hard Optimization Problems (EHOP), a collection of NP-hard problems expressed in natural language. EHOP includes problem formulations that could be found in computer science textbooks (e.g., graph coloring), versions that are dressed up as problems that could arise in real life (e.g., party planning), and variants with inverted rules. We find that state-of-the-art LLMs, across multiple prompting strategies, systematically solve textbook problems more accurately than their real-life and inverted counterparts. While reasoning models are more capable, they nonetheless show high variance across problem presentations, suggesting they lack a truly robust reasoning mechanism. We argue that this constitutes evidence that LLMs are still heavily dependent on what was seen in training and struggle to generalize to novel problems.

[125] Automated Evaluation of Meter and Rhyme in Russian Generative and Human-Authored Poetry

Ilya Koziev

Main category: cs.CL

TL;DR: A tool library and dataset for analyzing Russian poetry versification rules, including stress marking, rhyme detection, and poetic defect identification.

DetailsMotivation: Generative poetry systems need effective tools for data engineering and automatic evaluation, particularly to assess adherence to versification rules like stress patterns and rhymes.

Method: Developed Russian Poetry Scansion Tool library for stress mark placement in Russian syllabo-tonic poetry, rhyme detection, and poetic defect identification. Also created RIFMA dataset of annotated poem fragments.

Result: Released a comprehensive tool library and annotated dataset that can evaluate large language models’ capability to place stress marks accurately in poetic texts.

Conclusion: These resources provide valuable tools for researchers and practitioners in creative generative AI, facilitating advancements in generative poetry system development and evaluation.

Abstract: Generative poetry systems require effective tools for data engineering and automatic evaluation, particularly to assess how well a poem adheres to versification rules, such as the correct alternation of stressed and unstressed syllables and the presence of rhymes. In this work, we introduce the Russian Poetry Scansion Tool library designed for stress mark placement in Russian-language syllabo-tonic poetry, rhyme detection, and identification of defects of poeticness. Additionally, we release RIFMA – a dataset of poem fragments spanning various genres and forms, annotated with stress marks. This dataset can be used to evaluate the capability of modern large language models to accurately place stress marks in poetic texts. The published resources provide valuable tools for researchers and practitioners in the field of creative generative AI, facilitating advancements in the development and evaluation of generative poetry systems.

[126] GEM: Empowering MLLM for Grounded ECG Understanding with Time Series and Images

Xiang Lan, Feng Wu, Kai He, Qinghao Zhao, Shenda Hong, Mengling Feng

Main category: cs.CL

TL;DR: GEM is a multimodal large language model that unifies ECG time series, 12-lead ECG images, and text for grounded and clinician-aligned ECG interpretation, addressing limitations in multimodal synergy and explainability.

DetailsMotivation: Recent MLLMs for ECG interpretation face insufficient multimodal synergy between time series signals and visual ECG representations, and limited explainability in linking diagnoses to granular waveform evidence.

Method: Uses a dual-encoder framework extracting complementary time series and image features, cross-modal alignment for multimodal understanding, and knowledge-guided instruction generation for generating high-granularity grounding data (ECG-Grounding) linking diagnoses to measurable parameters.

Result: Significantly improves predictive performance (CSN 7.4% ↑), explainability (22.7% ↑), and grounding (24.8% ↑) on both existing and proposed benchmarks.

Conclusion: GEM makes ECG interpretation more suitable for real-world clinical applications by enhancing multimodal understanding and providing evidence-driven reasoning similar to clinicians.

Abstract: While recent multimodal large language models (MLLMs) have advanced automated ECG interpretation, they still face two key limitations: (1) insufficient multimodal synergy between time series signals and visual ECG representations, and (2) limited explainability in linking diagnoses to granular waveform evidence. We introduce GEM, the first MLLM unifying ECG time series, 12-lead ECG images and text for grounded and clinician-aligned ECG interpretation. GEM enables feature-grounded analysis, evidence-driven reasoning, and a clinician-like diagnostic process through three core innovations: a dual-encoder framework extracting complementary time series and image features, cross-modal alignment for effective multimodal understanding, and knowledge-guided instruction generation for generating high-granularity grounding data (ECG-Grounding) linking diagnoses to measurable parameters ($e.g.$, QRS/PR Intervals). Additionally, we propose the Grounded ECG Understanding task, a clinically motivated benchmark designed to comprehensively assess the MLLM’s capability in grounded ECG understanding. Experimental results on both existing and our proposed benchmarks show GEM significantly improves predictive performance (CSN $7.4% \uparrow$), explainability ($22.7% \uparrow$), and grounding ($24.8% \uparrow$), making it more suitable for real-world clinical applications. GitHub repository: https://github.com/lanxiang1017/GEM.git

[127] Reassessing Active Learning Adoption in Contemporary NLP: A Community Survey

Julia Romberg, Christopher Schröder, Julius Gonsior, Katrin Tomanek, Fredrik Olsson

Main category: cs.CL

TL;DR: Survey of NLP community reveals active learning remains relevant with LLMs, but faces persistent challenges in setup complexity, uncertain cost reduction, and tooling.

DetailsMotivation: To understand how advances in active learning, especially with LLMs, have translated into real-world applications and identify barriers to adoption.

Method: Conducted an online survey in the NLP community to collect insights on implementation practices, obstacles, and future prospects of active learning.

Result: Data annotation expected to remain important; active learning stays relevant with LLMs; three key challenges persist from 15 years ago: setup complexity, uncertain cost reduction, and tooling.

Conclusion: Active learning continues to face adoption barriers despite LLM advances; proposed strategies to alleviate persistent challenges; dataset published for further research.

Abstract: Supervised learning relies on data annotation which usually is time-consuming and therefore expensive. A longstanding strategy to reduce annotation costs is active learning, an iterative process, in which a human annotates only data instances deemed informative by a model. Research in active learning has made considerable progress, especially with the rise of large language models (LLMs). However, we still know little about how these remarkable advances have translated into real-world applications, or contributed to removing key barriers to active learning adoption. To fill in this gap, we conduct an online survey in the NLP community to collect previously intangible insights on current implementation practices, common obstacles in application, and future prospects in active learning. We also reassess the perceived relevance of data annotation and active learning as fundamental assumptions. Our findings show that data annotation is expected to remain important and active learning to stay relevant while benefiting from LLMs. Consistent with a community survey from over 15 years ago, three key challenges yet persist – setup complexity, uncertain cost reduction, and tooling – for which we propose alleviation strategies. We publish an anonymized version of the dataset.

[128] When To Solve, When To Verify: Compute-Optimal Problem Solving and Generative Verification for LLM Reasoning

Nishad Singhi, Hritik Bansal, Arian Hosseini, Aditya Grover, Kai-Wei Chang, Marcus Rohrbach, Anna Rohrbach

Main category: cs.CL

TL;DR: Self-Consistency (SC) is more compute-efficient than Generative Reward Models (GenRM) for most practical inference budgets. GenRM requires up to 8x more compute to match SC performance and even more to outperform it.

DetailsMotivation: To address the trade-off between scaling solution generation (via Self-Consistency) and verification (via Generative Reward Models) under limited inference budgets for LLM reasoning tasks.

Method: Evaluated GenRM against SC under fixed inference budgets across diverse models and datasets, and derived inference scaling laws for the GenRM paradigm.

Result: SC outperforms GenRM for most practical compute budgets. GenRM first matches SC after consuming up to 8x more inference compute and requires significantly more compute to outperform SC.

Conclusion: Compute-optimal inference favors scaling solution generation more aggressively than verification. SC remains the more efficient approach for most practical scenarios.

Abstract: Scaling test-time compute has emerged as a key strategy for enhancing the reasoning capabilities of large language models (LLMs), particularly in tasks like mathematical problem-solving. A traditional approach, Self-Consistency (SC), generates multiple solutions to a problem and selects the most common answer via majority voting. Another common method involves scoring each solution with a reward model (verifier) and choosing the best one. Recent advancements in Generative Reward Models (GenRM) reframe verification as a next-token prediction task, enabling inference-time scaling along a new axis. Specifically, GenRM generates multiple verification chains-of-thought to score each solution. Under a limited inference budget, this introduces a fundamental trade-off: should you spend the budget on scaling solutions via SC or generate fewer solutions and allocate compute to verification via GenRM? To address this, we evaluate GenRM against SC under a fixed inference budget. Interestingly, we find that SC is more compute-efficient than GenRM for most practical inference budgets across diverse models and datasets. For instance, GenRM first matches SC after consuming up to 8x the inference compute and requires significantly more compute to outperform it. Furthermore, we derive inference scaling laws for the GenRM paradigm, revealing that compute-optimal inference favors scaling solution generation more aggressively than scaling the number of verifications. Our work provides practical guidance on optimizing test-time scaling by balancing solution generation and verification. The code is available at https://github.com/nishadsinghi/sc-genrm-scaling.

[129] Leveraging Robust Optimization for LLM Alignment under Distribution Shifts

Mingye Zhu, Yi Liu, Zheren Fu, Yongdong Zhang, Zhendong Mao

Main category: cs.CL

TL;DR: Proposes a distribution-aware optimization framework for preference alignment in LLMs that uses calibration values and robust optimization to mitigate distribution shifts from synthetic training data.

DetailsMotivation: Current preference alignment methods rely on synthetic LLM-generated data, which causes distribution shifts that undermine accurate representation of human preferences and lead to undesirable outputs.

Method: Uses well-learned classifiers to assign calibration values to training samples based on alignment with target human-preferred distribution, then incorporates these into a robust optimization objective that minimizes worst-case loss over preference-relevant data regions.

Result: The approach mitigates distributional mismatch and improves generation of responses that better reflect intended human values.

Conclusion: Distribution-aware optimization effectively addresses distribution shifts in preference alignment, enabling LLMs to produce outputs more consistent with human preferences despite reliance on synthetic training data.

Abstract: Preference alignment methods are increasingly critical for steering large language models (LLMs) to generate outputs consistent with human values. While recent approaches often rely on synthetic data generated by LLMs for scalability and cost-efficiency reasons, this reliance can introduce distribution shifts that undermine the nuanced representation of human preferences needed for desirable outputs. In this paper, we propose a novel distribution-aware optimization framework that improves preference alignment despite such shifts. Our approach first leverages well-learned classifiers to assign a calibration value to each training sample, quantifying its alignment with the target human-preferred distribution. These values are then incorporated into a robust optimization objective that minimizes the worst-case loss over regions of the data space most relevant to human preferences. By explicitly focusing optimization on the target distribution, our approach mitigates the impact of distributional mismatch and improves the generation of responses that better reflect intended values.

[130] Thinking Out Loud: Do Reasoning Models Know When They’re Right?

Qingcheng Zeng, Weihao Xuan, Leyang Cui, Rob Voigt

Main category: cs.CL

TL;DR: Large reasoning models show improved verbalized confidence through training but exhibit reduced awareness of knowledge boundaries, leading to fewer “I don’t know” responses and potential overconfidence despite reasoning capabilities.

DetailsMotivation: To investigate how self-reflection ability in large reasoning models interacts with other behaviors, particularly verbalized confidence and knowledge boundary awareness.

Method: Analyzed verbalized confidence as a lens into self-reflection, examining effects of supervised fine-tuning on reasoning traces (distillation) and reinforcement learning on verbalized calibration in reasoning-intensive settings.

Result: Training improved verbalized calibration progressively but reduced “I don’t know” response rates on factuality benchmarks. Models expressed higher confidence with shorter reasoning chains, indicating diminished knowledge boundary awareness.

Conclusion: Reasoning-oriented training enhances reasoning performance but incurs a “reasoning tax” - reduced ability to recognize knowledge limits, compromising model faithfulness through overconfidence without proper abstention awareness.

Abstract: Large reasoning models (LRMs) have recently demonstrated impressive capabilities in complex reasoning tasks by leveraging increased test-time computation and exhibiting behaviors reminiscent of human-like self-reflection. While LRMs show a clear capacity for valuable self-reflection, how this ability interacts with other model behaviors remains underexplored. We investigate this connection by analyzing verbalized confidence, how models articulate their certainty, as a lens into the nature of self-reflection in LRMs. We find that supervised fine-tuning on reasoning traces (i.e., distillation) and reinforcement learning can improve verbalized calibration in reasoning-intensive settings in a progressive, laddered fashion. However, our results also indicate that reasoning models may possess a diminished awareness of their own knowledge boundaries, as evidenced by significantly lower “I don’t know” response rates on factuality benchmarks. Moreover, we examine the relationship between verbalized confidence and reasoning chains, finding that models tend to express higher confidence when providing shorter or less elaborate reasoning. Our findings highlight how reasoning-oriented training can enhance performance in reasoning-centric tasks while potentially incurring a “reasoning tax,” a cost reflected in the model’s reduced ability to accurately recognize the limits of its own knowledge in small-scale models. More broadly, our work showcases how this erosion of knowledge boundaries can compromise model faithfulness, as models grow more confident without a commensurate understanding of when they should abstain.

[131] Forecasting Clinical Risk from Textual Time Series: Structuring Narratives for Temporal AI in Healthcare

Shahriar Noroozizadeh, Sayantan Kumar, Jeremy C. Weiss

Main category: cs.CL

TL;DR: This paper introduces forecasting from textual time series using LLM-extracted clinical findings, showing encoder-based models excel at event prediction while decoder models perform better in survival analysis, highlighting the importance of time ordering over text ordering.

DetailsMotivation: Clinical case reports contain valuable temporal patient trajectories that are underutilized by traditional machine learning methods relying on structured data.

Method: Used LLM-assisted annotation pipeline to extract timestamped clinical findings, then systematically evaluated decoder-based LLMs and encoder-based transformers on event prediction, temporal ordering, and survival analysis tasks.

Result: Encoder-based models achieved higher F1 scores and better temporal concordance for event forecasting, while fine-tuned masking improved ranking performance. Instruction-tuned decoder models showed relative advantage in survival analysis, especially for early prognosis.

Conclusion: Time ordering in clinical time series provides additional benefits beyond text ordering, with important implications for temporal tasks in the LLM era.

Abstract: Clinical case reports encode temporal patient trajectories that are often underexploited by traditional machine learning methods relying on structured data. In this work, we introduce the forecasting problem from textual time series, where timestamped clinical findings – extracted via an LLM-assisted annotation pipeline – serve as the primary input for prediction. We systematically evaluate a diverse suite of models, including fine-tuned decoder-based large language models and encoder-based transformers, on tasks of event occurrence prediction, temporal ordering, and survival analysis. Our experiments reveal that encoder-based models consistently achieve higher F1 scores and superior temporal concordance for short- and long-horizon event forecasting, while fine-tuned masking approaches enhance ranking performance. In contrast, instruction-tuned decoder models demonstrate a relative advantage in survival analysis, especially in early prognosis settings. Our sensitivity analyses further demonstrate the importance of time ordering, which requires clinical time series construction, as compared to text ordering, the format of the text inputs that LLMs are classically trained on. This highlights the additional benefit that can be ascertained from time-ordered corpora, with implications for temporal tasks in the era of widespread LLM use.

[132] Understanding LLMs’ Cross-Lingual Context Retrieval: How Good It Is And Where It Comes From

Changjiang Gao, Hankun Lin, Xin Huang, Xue Han, Junlan Feng, Chao Deng, Jiajun Chen, Shujian Huang

Main category: cs.CL

TL;DR: LLMs show strong cross-lingual context retrieval ability after post-training, comparable to GPT-4o. The process involves two phases: question encoding (pre-training) and answer retrieval (post-training), with performance bottleneck in the second phase.

DetailsMotivation: To evaluate the performance and mechanism of cross-lingual context retrieval in LLMs, which is fundamental for cross-lingual alignment but remains unclear.

Method: Evaluated over 40 LLMs across 12 languages using cross-lingual machine reading comprehension (xMRC) as a representative scenario, analyzing mechanism through two-phase process.

Result: Post-trained open LLMs show strong cross-lingual context retrieval comparable to GPT-4o. Oracle performances greatly improve after post-training. xMRC bottleneck lies at last model layers in answer retrieval phase.

Conclusion: Larger-scale pretraining alone cannot improve xMRC performance; LLMs need multilingual post-training to fully unlock cross-lingual context retrieval potential.

Abstract: Cross-lingual context retrieval (extracting contextual information in one language based on requests in another) is a fundamental aspect of cross-lingual alignment, but the performance and mechanism of it for large language models (LLMs) remains unclear. In this paper, we evaluate the cross-lingual context retrieval of over 40 LLMs across 12 languages, using cross-lingual machine reading comprehension (xMRC) as a representative scenario. Our results show that post-trained open LLMs show strong cross-lingual context retrieval ability, comparable to closed-source LLMs such as GPT-4o, and their estimated oracle performances greatly improve after post-training. Our mechanism analysis shows that the cross-lingual context retrieval process can be divided into two main phases: question encoding and answer retrieval, which are formed in pre-training and post-training respectively. The phasing stability correlates with xMRC performance, and the xMRC bottleneck lies at the last model layers in the second phase, where the effect of post-training can be evidently observed. Our results also indicate that larger-scale pretraining cannot improve the xMRC performance. Instead, larger LLMs need further multilingual post-training to fully unlock their cross-lingual context retrieval potential.

[133] LLMTaxo: Leveraging Large Language Models for Constructing Taxonomy of Factual Claims from Social Media

Haiqi Zhang, Zhengyuan Zhu, Zeyu Zhang, Chengkai Li

Main category: cs.CL

TL;DR: LLMTaxo is a framework that uses large language models to automatically build hierarchical taxonomies of factual claims from social media, reducing redundancy and improving information organization.

DetailsMotivation: With the rapid expansion of social media content, analyzing online discourse has become increasingly complex, requiring better methods to organize and comprehend factual claims.

Method: The framework leverages large language models to generate topics at multiple levels of granularity, creating hierarchical taxonomies with dedicated evaluation metrics for comprehensive assessment.

Result: Evaluations on three diverse datasets show LLMTaxo effectively produces clear, coherent, and comprehensive taxonomies, with GPT-4o mini consistently outperforming other models across most metrics.

Conclusion: LLMTaxo demonstrates flexibility and low reliance on manual intervention, highlighting its potential for broad applicability in organizing social media discourse.

Abstract: With the rapid expansion of content on social media platforms, analyzing and comprehending online discourse has become increasingly complex. This paper introduces LLMTaxo, a novel framework leveraging large language models for the automated construction of taxonomies of factual claims from social media by generating topics at multiple levels of granularity. The resulting hierarchical structure significantly reduces redundancy and improves information accessibility. We also propose dedicated taxonomy evaluation metrics to enable comprehensive assessment. Evaluations conducted on three diverse datasets demonstrate LLMTaxo’s effectiveness in producing clear, coherent, and comprehensive taxonomies. Among the evaluated models, GPT-4o mini consistently outperforms others across most metrics. The framework’s flexibility and low reliance on manual intervention underscore its potential for broad applicability.

[134] HCR-Reasoner: Synergizing Large Language Models and Theory for Human-like Causal Reasoning

Yanxi Zhang, Xin Cong, Zhong Zhang, Xiao Liu, Dongyan Zhao, Yesai Wu

Main category: cs.CL

TL;DR: HCR-Reasoner integrates actual causality theory and causal judgment factors into LLMs for human-like causal reasoning, significantly improving alignment with human reasoning.

DetailsMotivation: Human causal reasoning involves identifying causal chain membership first, then considering modulatory factors like morality and intention. Current AI lacks systematic integration of actual causality formalisms and psychological modulators studied in cognitive science.

Method: HCR-Reasoner framework uses actual causality formalisms to filter candidate causes, then applies causal judgment factors to determine psychologically selected causes. Evaluated on HCR-Bench with 1,093 annotated instances.

Result: HCR-Reasoner consistently and significantly improves LLMs’ causal alignment with humans. Explicit integration of theory-guided reasoning is highly effective for achieving faithful human-like causal reasoning.

Conclusion: Systematic integration of actual causality theory and causal judgment factors into LLMs enables more human-like causal reasoning, demonstrating the value of theory-guided approaches for strong AI.

Abstract: Genuine human-like causal reasoning is fundamental for strong artificial intelligence. Humans typically identify whether an event is part of the causal chain first, and then influenced by modulatory factors such as morality, normality, and intention to make the final judgment. These two stages naturally map to the fields of 1) actual causality that provides formalisms for causal chain membership and 2) causal judgment from cognitive science that studies psychological modulators that influence causal selection. However, these two domains have largely been studied in isolation, leaving a gap for a systematic method based on LLMs. Therefore, we introduce HCR-Reasoner, a framework that systematically integrates the theory of actual causality and causal judgment into LLMs for human-like causal reasoning. It simulates humans by using actual causality formalisms to filter for structurally necessary candidate causes and causal judgment factors to determine the psychologically selected cause. For fine-grained evaluation, we introduce HCR-Bench, a challenging benchmark with 1,093 annotated instances with detailed reasoning steps. Results show HCR-Reasoner consistently and significantly improves LLMs’ causal alignment with humans, and that explicitly integrating theory-guided reasoning into LLMs is highly effective for achieving faithful human-like causal reasoning.

[135] Intrinsic Self-Correction in LLMs: Towards Explainable Prompting via Mechanistic Interpretability

Yu-Ting Lee, Fu-Chieh Chang, Hui-Ying Shih, Pei-Yuan Wu

Main category: cs.CL

TL;DR: Intrinsic self-correction in LLMs works by steering hidden representations along interpretable latent directions, as revealed by analyzing prompt-induced shifts in text detoxification and toxification tasks.

DetailsMotivation: To understand the internal mechanisms of intrinsic self-correction in language models, which improves performance through prompting alone but remains poorly understood at the representation level.

Method: Analyzed prompt-induced shifts (changes in hidden representations caused by self-correction prompts) across 5 open-source LLMs, comparing shifts in text detoxification and toxification with latent directions from contrastive pairs.

Result: Prompt-induced shifts align with interpretable latent directions: in detoxification they align with non-toxic direction, in toxification with toxic direction, showing self-correction functions as representation steering beyond standard metrics.

Conclusion: Intrinsic self-correction operates through representation steering along interpretable latent directions, providing an interpretability-based account that advances systematic understanding of LLM prompting mechanisms.

Abstract: Intrinsic self-correction refers to the phenomenon where a language model refines its own outputs purely through prompting, without external feedback or parameter updates. While this approach improves performance across diverse tasks, its internal mechanism remains poorly understood. We analyze intrinsic self-correction from a representation-level perspective. We formalize and introduce the notion of a prompt-induced shift, which is the change in hidden representations caused by a self-correction prompt. Across 5 open-source LLMs, prompt-induced shifts in text detoxification and text toxification align with latent directions constructed from contrastive pairs. In detoxification, the shifts align with the non-toxic direction; in toxification, they align with the toxic direction. These results suggest that intrinsic self-correction functions as representation steering along interpretable latent directions, beyond what standard metrics such as task scores or model confidence capture. Our analysis offers an interpretability-based account of intrinsic self-correction and contributes to a more systematic understanding of LLM prompting.

[136] PsyMem: Fine-grained psychological alignment and Explicit Memory Control for Advanced Role-Playing LLMs

Xilong Cheng, Yunxiao Qin, Yuting Tan, Zhengnan Li, Ye Wang, Hongjiang Xiao, Yuan Zhang

Main category: cs.CL

TL;DR: PsyMem is a framework that enhances role-playing LLMs by incorporating fine-grained psychological attributes and explicit memory control, addressing limitations in character modeling and memory consistency.

DetailsMotivation: Existing LLM-based role-playing methods inadequately model character dimensions and lack explicit memory alignment, compromising reliability in applications like social simulation.

Method: PsyMem supplements textual descriptions with 26 psychological indicators and implements memory alignment training to explicitly align character responses with memory, using a dataset of 5,414 characters and 38,962 dialogues from novels.

Result: The resulting PsyMem-Qwen model (trained on Qwen2.5-7B-Instruct) outperforms baseline models in role-playing, achieving the best performance in human-likeness and character fidelity.

Conclusion: PsyMem successfully addresses key limitations in role-playing LLMs through psychological attribute modeling and explicit memory control, enhancing reliability and performance.

Abstract: Existing LLM-based role-playing methods often rely on superficial textual descriptions or simplistic metrics, inadequately modeling both intrinsic and extrinsic character dimensions. Additionally, they typically simulate character memory with implicit model knowledge or basic retrieval augment generation without explicit memory alignment, compromising memory consistency. The two issues weaken reliability of role-playing LLMs in several applications, such as trustworthy social simulation. To address these limitations, we propose PsyMem, a novel framework integrating fine-grained psychological attributes and explicit memory control for role-playing. PsyMem supplements textual descriptions with 26 psychological indicators to detailed model character. Additionally, PsyMem implements memory alignment training, explicitly trains the model to align character’s response with memory, thereby enabling dynamic memory-controlled responding during inference. By training Qwen2.5-7B-Instruct on our specially designed dataset (including 5,414 characters and 38,962 dialogues extracted from novels), the resulting model, termed as PsyMem-Qwen, outperforms baseline models in role-playing, achieving the best performance in human-likeness and character fidelity.

[137] TemplateRL: Structured Template-Guided Reinforcement Learning for LLM Reasoning

Jinyang Wu, Chonghua Liao, Mingkuan Feng, Shuai Zhang, Zhengqi Wen, Haoran Luo, Ling Yang, Huazhe Xu, Jianhua Tao

Main category: cs.CL

TL;DR: TemplateRL is a structured template-guided RL framework that improves reasoning by using explicit templates to guide policy optimization, achieving significant performance gains over existing methods.

DetailsMotivation: Existing RL methods like GRPO rely on unstructured self-sampling with scalar rewards, producing inefficient rollouts that fail to capture transferable problem-solving strategies.

Method: Constructs a problem-solving template library via MCTS on a small seed set, then integrates this high-level structured guidance into RL training to align rollout generation with proven template structures.

Result: Outperforms GRPO by 99% on AIME and 41% on AMC, with superior stability on weak models and remarkable cross-domain generalization.

Conclusion: TemplateRL’s structure-guided design effectively steers policies toward validated strategic patterns, stabilizing training and enhancing sampling efficiency while maintaining interpretability and editability.

Abstract: Reinforcement learning (RL) has emerged as an effective paradigm for enhancing model reasoning. However, existing RL methods like GRPO often rely on unstructured self-sampling to fit scalar rewards, often producing inefficient rollouts that fail to capture transferable problem-solving strategies. To address these limitations, we propose TemplateRL, a structured template-guided RL framework that augments policy optimization with explicit template guidance. Our approach first constructs a problem-solving template library via MCTS on a small seed set, then seamlessly integrates this high-level structured guidance into RL training. By guiding rollout generation to align with proven template structures, TemplateRL significantly improves high-quality trajectory hit rates while reducing ineffective exploration. This structure-guided design steers the policy toward validated strategic patterns, stabilizing training dynamics, and enhancing RL sampling efficiency. Notably, the explicit template library is interpretable, editable, and supports online updates-enabling continuous updates during both training and inference. Extensive experiments demonstrate that TemplateRL outperforms GRPO by 99% on AIME and 41% on AMC, with superior stability on weak models and remarkable cross-domain generalization, highlighting its potential for broader tasks.

[138] Towards Evaluating Proactive Risk Awareness of Multimodal Language Models

Youliang Yuan, Wenxiang Jiao, Yuejin Xie, Chihao Shen, Menghan Tian, Wenxuan Wang, Jen-tse Huang, Pinjia He

Main category: cs.CL

TL;DR: Proactive Safety Bench (PaSBench) evaluates AI systems’ ability to detect safety risks proactively through 416 multimodal scenarios, revealing that top models still miss 45-55% of risks due to unstable reasoning rather than knowledge gaps.

DetailsMotivation: Human safety awareness gaps prevent timely risk recognition, and proactive AI systems that actively monitor behavior and environment would be more effective than reactive ones that only respond to user questions.

Method: Developed PaSBench with 416 multimodal scenarios (128 image sequences, 288 text logs) across 5 safety-critical domains, and evaluated 36 advanced AI models on their proactive risk detection capabilities.

Result: Top performers like Gemini-2.5-pro achieved 71% image and 64% text accuracy, but missed 45-55% of risks in repeated trials. Failure analysis identified unstable proactive reasoning as the primary limitation rather than knowledge deficits.

Conclusion: This work establishes a proactive safety benchmark, provides systematic evidence of model limitations, and identifies critical directions for developing reliable protective AI that actively prevents harm rather than merely responding to requests.

Abstract: Human safety awareness gaps often prevent the timely recognition of everyday risks. In solving this problem, a proactive safety artificial intelligence (AI) system would work better than a reactive one. Instead of just reacting to users’ questions, it would actively watch people’s behavior and their environment to detect potential dangers in advance. Our Proactive Safety Bench (PaSBench) evaluates this capability through 416 multimodal scenarios (128 image sequences, 288 text logs) spanning 5 safety-critical domains. Evaluation of 36 advanced models reveals fundamental limitations: Top performers like Gemini-2.5-pro achieve 71% image and 64% text accuracy, but miss 45-55% risks in repeated trials. Through failure analysis, we identify unstable proactive reasoning rather than knowledge deficits as the primary limitation. This work establishes (1) a proactive safety benchmark, (2) systematic evidence of model limitations, and (3) critical directions for developing reliable protective AI. We believe our dataset and findings can promote the development of safer AI assistants that actively prevent harm rather than merely respond to requests. Our dataset can be found at https://huggingface.co/datasets/Youliang/PaSBench.

[139] MedScore: Generalizable Factuality Evaluation of Free-Form Medical Answers by Domain-adapted Claim Decomposition and Verification

Heyuan Huang, Alexandra DeLucia, Vijay Murari Tiyyala, Mark Dredze

Main category: cs.CL

TL;DR: MedScore is a new pipeline for evaluating factuality of medical LLM responses by decomposing them into condition-aware valid facts and verifying against medical corpora, outperforming existing methods.

DetailsMotivation: Existing factuality evaluation systems are poorly suited for medical domain due to their focus on objective, entity-centric texts, while medical answers are condition-dependent, conversational, and structurally diverse.

Method: Proposes a pipeline that decomposes medical answers into condition-aware valid facts and verifies them against in-domain medical corpora, extracting more facts while reducing hallucinations and vague references.

Result: Extracts up to 3 times more valid facts than existing methods, reduces hallucination and vague references, and retains condition-dependency in facts. Factuality scores vary significantly by decomposition method, verification corpus, and LLM backbone.

Conclusion: Customizing each evaluation step is crucial for reliable factuality assessment in medical domain, and the proposed modular pipeline enables effective domain adaptation.

Abstract: While Large Language Models (LLMs) can generate fluent and convincing responses, they are not necessarily correct. This is especially apparent in the popular decompose-then-verify factuality evaluation pipeline, where LLMs evaluate generations by decomposing the generations into individual, valid claims. Factuality evaluation is especially important for medical answers, since incorrect medical information could seriously harm the patient. However, existing factuality systems are a poor match for the medical domain, as they are typically only evaluated on objective, entity-centric, formulaic texts such as biographies and historical topics. This differs from condition-dependent, conversational, hypothetical, sentence-structure diverse, and subjective medical answers, which makes decomposition into valid facts challenging. We propose MedScore, a new pipeline to decompose medical answers into condition-aware valid facts and verify against in-domain corpora. Our method extracts up to three times more valid facts than existing methods, reducing hallucination and vague references, and retaining condition-dependency in facts. The resulting factuality score substantially varies by decomposition method, verification corpus, and used backbone LLM, highlighting the importance of customizing each step for reliable factuality evaluation by using our generalizable and modularized pipeline for domain adaptation.

[140] Flex-Judge: Text-Only Reasoning Unleashes Zero-Shot Multimodal Evaluators

Jongwoo Ko, Sungnyun Kim, Sungwoo Cho, Se-Young Yun

Main category: cs.CL

TL;DR: Flex-Judge is a reasoning-guided multimodal judge model that uses minimal textual reasoning data to generalize across multiple modalities and evaluation formats, achieving competitive performance with less training data than traditional approaches.

DetailsMotivation: Current LLM-as-a-Judge approaches require extensive modality-specific training data and fail to generalize well across diverse multimodal tasks, making human-aligned reward signals costly and limited in scalability.

Method: Leverages structured textual reasoning explanations that inherently encode generalizable decision-making patterns, enabling effective transfer to multimodal judgments with images, videos, and other modalities using minimal text data.

Result: Achieves competitive or superior performance compared to state-of-the-art commercial APIs and extensively trained multimodal evaluators, despite significantly less training data. Shows broad impact in resource-constrained domains like molecule evaluation.

Conclusion: Reasoning-based text supervision provides a powerful, cost-effective alternative to traditional annotation-intensive approaches, substantially advancing scalable multimodal model-as-a-judge frameworks.

Abstract: Human-generated reward signals are critical for aligning generative models with human preferences, guiding both training and inference-time evaluations. While large language models (LLMs) employed as proxy evaluators, i.e., LLM-as-a-Judge, significantly reduce the costs associated with manual annotations, they typically require extensive modality-specific training data and fail to generalize well across diverse multimodal tasks. In this paper, we propose Flex-Judge, a reasoning-guided multimodal judge model that leverages minimal textual reasoning data to robustly generalize across multiple modalities and evaluation formats. Our core intuition is that structured textual reasoning explanations inherently encode generalizable decision-making patterns, enabling an effective transfer to multimodal judgments, e.g., with images or videos. Empirical results demonstrate that Flex-Judge, despite being trained on significantly fewer text data, achieves competitive or superior performance compared to state-of-the-art commercial APIs and extensively trained multimodal evaluators. Notably, Flex-Judge presents broad impact in modalities like molecule, where comprehensive evaluation benchmarks are scarce, underscoring its practical value in resource-constrained domains. Our framework highlights reasoning-based text supervision as a powerful, cost-effective alternative to traditional annotation-intensive approaches, substantially advancing scalable multimodal model-as-a-judge.

[141] Unifying Attention Heads and Task Vectors via Hidden State Geometry in In-Context Learning

Haolin Yang, Hakaze Cho, Yiqiao Zhong, Naoya Inoue

Main category: cs.CL

TL;DR: The paper proposes a unified framework for understanding in-context learning in classification tasks by analyzing two geometric factors: separability and alignment of query hidden states, revealing a two-stage mechanism across layers.

DetailsMotivation: Prior work on in-context learning mechanisms has focused on either attention heads or task vectors in isolation, lacking a unified framework that connects these components to the evolution of hidden states across layers that produce model outputs.

Method: Analyzed geometric factors (separability and alignment of query hidden states) in classification tasks, conducted fine-grained analysis of layer-wise dynamics, and performed ablation studies on attention heads and task vectors.

Result: Revealed a two-stage mechanism: separability emerges in early layers, while alignment develops in later layers. Previous Token Heads drive separability, while Induction Heads and task vectors enhance alignment.

Conclusion: The findings bridge the gap between attention heads and task vectors, offering a unified account of in-context learning’s underlying mechanisms.

Abstract: The unusual properties of in-context learning (ICL) have prompted investigations into the internal mechanisms of large language models. Prior work typically focuses on either special attention heads or task vectors at specific layers, but lacks a unified framework linking these components to the evolution of hidden states across layers that ultimately produce the model’s output. In this paper, we propose such a framework for ICL in classification tasks by analyzing two geometric factors that govern performance: the separability and alignment of query hidden states. A fine-grained analysis of layer-wise dynamics reveals a striking two-stage mechanism: separability emerges in early layers, while alignment develops in later layers. Ablation studies further show that Previous Token Heads drive separability, while Induction Heads and task vectors enhance alignment. Our findings thus bridge the gap between attention heads and task vectors, offering a unified account of ICL’s underlying mechanisms.

[142] Grounding Language with Vision: A Conditional Mutual Information Calibrated Decoding Strategy for Reducing Hallucinations in LVLMs

Hao Fang, Changle Zhou, Jiawei Kong, Kuofeng Gao, Bin Chen, Tao Liang, Guojun Ma, Shu-Tao Xia

Main category: cs.CL

TL;DR: A novel C-PMI decoding strategy that reduces hallucinations in Large Vision-Language Models by strengthening mutual dependency between generated text and input images through bi-level optimization and token purification.

DetailsMotivation: LVLMs suffer from hallucinations where responses seem plausible but lack relevance to input images, primarily due to over-reliance on language priors while ignoring visual information during decoding.

Method: Proposes Conditional Pointwise Mutual Information (C-PMI) calibrated decoding that jointly models visual and textual token contributions, formulated as bi-level optimization with token purification mechanism to dynamically regulate decoding.

Result: Extensive experiments show significant reduction in hallucinations across various benchmarks while preserving decoding efficiency.

Conclusion: The C-PMI method effectively mitigates hallucinations in LVLMs by adaptively strengthening visual-textual mutual dependency through joint optimization of image and text token contributions.

Abstract: Large Vision-Language Models (LVLMs) are susceptible to hallucinations, where generated responses seem semantically plausible yet exhibit little or no relevance to the input image. Previous studies reveal that this issue primarily stems from LVLMs’ over-reliance on language priors while disregarding the visual information during decoding. To alleviate this issue, we introduce a novel Conditional Pointwise Mutual Information (C-PMI) calibrated decoding strategy, which adaptively strengthens the mutual dependency between generated texts and input images to mitigate hallucinations. Unlike existing methods solely focusing on text token sampling, we propose to jointly model the contributions of visual and textual tokens to C-PMI, formulating hallucination mitigation as a bi-level optimization problem aimed at maximizing mutual information. To solve it, we design a token purification mechanism that dynamically regulates the decoding process by sampling text tokens remaining maximally relevant to the given image, while simultaneously refining image tokens most pertinent to the generated response. Extensive experiments across various benchmarks reveal that the proposed method significantly reduces hallucinations in LVLMs while preserving decoding efficiency.

[143] Leveraging Importance Sampling to Detach Alignment Modules from Large Language Models

Yi Liu, Dianqing Liu, Mingye Zhu, Junbo Guo, Yongdong Zhang, Zhendong Mao

Main category: cs.CL

TL;DR: Proposes Residual Alignment Model (RAM) that treats alignment as importance sampling, enabling flexible adaptation of LLMs without retraining large models.

DetailsMotivation: Traditional alignment methods require retraining large pretrained models, making it difficult to quickly adapt LLMs for diverse applications.

Method: Frames alignment as importance sampling with upstream model as proposal distribution and alignment module as importance weight estimator. Uses sequence-level training and iterative token-level decoding to address latency.

Result: Experimental evaluations on open-source LLMs across instruction following, domain adaptation, and preference optimization tasks show consistent outperformance over baselines.

Conclusion: RAM provides a flexible and scalable approach to LLM alignment that improves performance while addressing practical deployment challenges like latency.

Abstract: The widespread adoption of large language models (LLMs) across industries has increased the demand for high-quality and customizable outputs. However, traditional alignment methods often require retraining large pretrained models, making it difficult to quickly adapt and optimize LLMs for diverse applications. To address this limitation, we propose a novel \textit{Residual Alignment Model} (\textit{RAM}) that formalizes the alignment process as a type of importance sampling. In this framework, the unaligned upstream model serves as the proposal distribution, while the alignment process is framed as secondary sampling based on an autoregressive alignment module that acts as an estimator of the importance weights. This design enables a natural detachment of the alignment module from the target aligned model, improving flexibility and scalability. Based on this model, we derive an efficient sequence-level training strategy for the alignment module, which operates independently of the proposal module. Additionally, we develop a resampling algorithm with iterative token-level decoding to address the common first-token latency issue in comparable methods. Experimental evaluations on two leading open-source LLMs across diverse tasks, including instruction following, domain adaptation, and preference optimization, demonstrate that our approach consistently outperforms baseline models.

[144] ToMAP: Training Opponent-Aware LLM Persuaders with Theory of Mind

Peixuan Han, Zijia Liu, Jiaxuan You

Main category: cs.CL

TL;DR: ToMAP is a 3B-parameter LLM persuader that incorporates Theory of Mind reasoning to model opponent’s mental states, outperforming much larger models like GPT-4o by 39.4% through enhanced opponent awareness and diverse argument generation.

DetailsMotivation: Current LLMs struggle with Theory of Mind reasoning, limiting their ability to model opponents' thoughts dynamically, which results in limited diversity and opponent awareness in persuasion tasks.

Method: Incorporates two theory of mind modules: prompting to consider possible objections, and using a text encoder with MLP classifier to predict opponent’s stance on counterclaims. Uses reinforcement learning to learn how to analyze opponent information and generate effective arguments.

Result: ToMAP outperforms much larger baselines like GPT-4o with 39.4% relative gain across multiple persuadee models and diverse corpora. Exhibits complex reasoning chains, reduced repetition, and more diverse/effective arguments. Suitable for long conversations with logical, opponent-aware strategies.

Conclusion: ToMAP effectively addresses LLMs’ Theory of Mind limitations in persuasion, demonstrating superior performance and highlighting potential for developing more persuasive language agents through enhanced opponent modeling.

Abstract: Large language models (LLMs) have shown promising potential in persuasion, but existing works on training LLM persuaders are still preliminary. Notably, while humans are skilled in modeling their opponent’s thoughts and opinions proactively and dynamically, current LLMs struggle with such Theory of Mind (ToM) reasoning, resulting in limited diversity and opponent awareness. To address this limitation, we introduce Theory of Mind Augmented Persuader (ToMAP), a novel approach for building more flexible persuader agents by incorporating two theory of mind modules that enhance the persuader’s awareness and analysis of the opponent’s mental state. Specifically, we begin by prompting the persuader to consider possible objections to the target central claim, and then use a text encoder paired with a trained MLP classifier to predict the opponent’s current stance on these counterclaims. Our carefully designed reinforcement learning schema enables the persuader learns how to analyze opponent-related information and utilize it to generate more effective arguments. Experiments show that the ToMAP persuader, while containing only 3B parameters, outperforms much larger baselines, like GPT-4o, with a relative gain of 39.4% across multiple persuadee models and diverse corpora. Notably, ToMAP exhibits complex reasoning chains and reduced repetition during training, which leads to more diverse and effective arguments. The opponent-aware feature of ToMAP also makes it suitable for long conversations and enables it to employ more logical and opponent-aware strategies. These results underscore our method’s effectiveness and highlight its potential for developing more persuasive language agents. Code is available at: https://github.com/ulab-uiuc/ToMAP.

[145] A*-Thought: Efficient Reasoning via Bidirectional Compression for Low-Resource Settings

Xiaoang Xu, Shuo Wang, Xu Han, Zhenghao Liu, Huijia Wu, Peipei Li, Zhiyuan Liu, Maosong Sun, Zhaofeng He

Main category: cs.CL

TL;DR: A*-Thought is an efficient tree search framework that compresses lengthy reasoning chains in Large Reasoning Models by using A* search with a cost function to find high-density, low-cost reasoning paths, improving performance while reducing output length.

DetailsMotivation: Large Reasoning Models suffer from reduced efficiency due to lengthy thinking trajectories. Existing compression methods often degrade performance, creating a need for an approach that can efficiently identify essential thoughts without sacrificing accuracy.

Method: Formulates reasoning as a search tree where nodes represent reasoning spans. Uses A* search algorithm with a cost function for reasoning paths and a bidirectional importance estimation mechanism to efficiently compress chains of thought.

Result: Improves QwQ-32B performance by 2.39× with low budget and reduces output token length by nearly 50% with high budget. Shows compatibility with multiple LRMs and effective balance between performance and efficiency on math tasks.

Conclusion: A*-Thought provides an effective framework for compressing reasoning chains in Large Reasoning Models, achieving better performance-efficiency trade-offs than existing methods while maintaining generalization across different models.

Abstract: Large Reasoning Models (LRMs) achieve superior performance by extending the thought length. However, a lengthy thinking trajectory leads to reduced efficiency. Most of the existing methods are stuck in the assumption of overthinking and attempt to reason efficiently by compressing the Chain-of-Thought, but this often leads to performance degradation. To address this problem, we introduce A*-Thought, an efficient tree search-based unified framework designed to identify and isolate the most essential thoughts from the extensive reasoning chains produced by these models. It formulates the reasoning process of LRMs as a search tree, where each node represents a reasoning span in the giant reasoning space. By combining the A* search algorithm with a cost function specific to the reasoning path, it can efficiently compress the chain of thought and determine a reasoning path with high information density and low cost. In addition, we also propose a bidirectional importance estimation mechanism, which further refines this search process and enhances its efficiency beyond uniform sampling. Extensive experiments on several advanced math tasks show that A*-Thought effectively balances performance and efficiency over a huge search space. Specifically, A*-Thought can improve the performance of QwQ-32B by 2.39$\times$ with low-budget and reduce the length of the output token by nearly 50% with high-budget. The proposed method is also compatible with several other LRMs, demonstrating its generalization capability. The code can be accessed at: https://github.com/AI9Stars/AStar-Thought.

[146] SATA-BENCH: Select All That Apply Benchmark for Multiple Choice Questions

Weijie Xu, Shixian Cui, Xi Fang, Chi Xue, Stephanie Eckman, Chandan K. Reddy

Main category: cs.CL

TL;DR: SATA-BENCH is the first benchmark for evaluating LLMs on Select All That Apply questions, revealing significant performance gaps (best model: 41.8% exact match) due to selection and count biases. Choice Funnel decoding strategy improves performance by 29% while reducing inference cost by 64%.

DetailsMotivation: Real-world problems often require identifying all correct answers from multiple options, but current LLM evaluations focus mainly on single-answer multiple-choice tasks, leaving multi-answer reasoning underexplored.

Method: Created SATA-BENCH benchmark for SATA questions across diverse domains. Evaluated 27 models, identified selection bias and count bias as core challenges. Proposed Choice Funnel decoding strategy combining token debiasing with adaptive thresholding.

Result: Even the strongest model achieved only 41.8% exact match on SATA questions. Choice Funnel achieved up to 29% higher exact match than baselines while reducing inference cost by over 64%.

Conclusion: Current LLMs have fundamental limitations in multi-answer reasoning. Choice Funnel provides an effective framework for improving performance on SATA tasks while reducing computational costs.

Abstract: Large language models (LLMs) are increasingly evaluated on single-answer multiple-choice tasks, yet many real-world problems require identifying all correct answers from a set of options. This capability remains underexplored. We introduce SATA-BENCH, the first dedicated benchmark for evaluating LLMs on Select All That Apply (SATA) questions across diverse domains, including reading comprehension, law, and biomedicine. Our evaluation of 27 open-source and proprietary models reveals a significant gap: even the strongest model achieves only 41.8% exact match, exposing LLMs’ inability to reliably identify all correct answers. We find that this weakness stems from two core challenges: selection bias - models favor certain choices regardless of content, and count bias - models fail to predict the correct number of answers. To address these issues, we propose Choice Funnel, a decoding strategy that combines token debiasing with adaptive thresholding to guide models toward complete and accurate selections. Choice Funnel achieves up to 29% higher exact match than competitive baselines while reducing inference cost by over 64%. Our findings expose fundamental limitations in current LLMs and introduce a new framework for diagnosing and improving multi-answer reasoning. We release SATA-BENCH and Choice Funnel to promote LLM development for robust decision-making in realistic, multi-answer applications.

[147] KG-TRACES: Enhancing Large Language Models with Knowledge Graph-constrained Trajectory Reasoning and Attribution Supervision

Rong Wu, Pinlong Cai, Jianbiao Mei, Licheng Wen, Tao Hu, Xuemeng Yang, Daocheng Fu, Botian Shi

Main category: cs.CL

TL;DR: KG-TRACES is a framework that enhances LLM reasoning by supervising reasoning paths and processes using knowledge graphs, improving explainability and performance on complex reasoning tasks.

DetailsMotivation: LLMs struggle with explainability and trustworthiness in complex reasoning, often producing hallucinations or unattributable reasoning, limiting their practical application.

Method: KG-TRACES jointly supervises LLMs to predict symbolic relation paths, full triple-level reasoning paths, and generate attribution-aware reasoning processes grounded in knowledge graphs.

Result: Significant performance improvements: 1.6% Hits@1 and 4.7% F1 on WebQSP; 4.8% Hits@1 and 2.1% F1 on CWQ. Also shows transferability to domains like medicine and produces more stable reasoning processes.

Conclusion: KG-TRACES enables explainable, source-attributable reasoning by explicitly supervising reasoning paths, outperforming SOTA methods and providing more transparent reasoning processes.

Abstract: Large language models (LLMs) have made remarkable strides in various natural language processing tasks, but their performance on complex reasoning problems remains hindered by a lack of explainability and trustworthiness. This issue, often manifesting as hallucinations or unattributable reasoning processes, limits their applicability in complex reasoning scenarios. To address this, we propose Knowledge Graph-constrained Trajectory Reasoning Attribution and Chain Explanation Supervision (KG-TRACES), a novel framework that enhances the reasoning ability of LLMs through explicit supervision over reasoning paths and processes. KG-TRACES jointly supervises the model to: (1) predict symbolic relation paths, (2) predict full triple-level reasoning paths, and (3) generate attribution-aware reasoning processes grounded in the reasoning paths. At inference phase, the model adapts to both KG-available and KG-unavailable scenarios, retrieving reasoning paths from a KG when possible or predicting plausible reasoning paths with only intrinsic knowledge when not. This design enables the model to reason in an explainable and source-attributable pattern. Through extensive experiments on complex reasoning tasks, we demonstrate that KG-TRACES significantly outperforms existing SOTA: it improves Hits@1 by 1.6% and F1 by 4.7% on WebQSP, and achieves improvements of 4.8% in Hits@1 and 2.1% in F1 on CWQ. Moreover, we show its transferability to specialized domains such as medicine. By visualizing the intermediate steps of reasoning processes, we further show that the explicit supervision introduced by KG-TRACES leads to more stable and goal-directed reasoning processes, aligning closely with correct answers. Code is available at https://github.com/Edaizi/KG-TRACES.

[148] A Controllable Examination for Long-Context Language Models

Yijun Yang, Zeyu Huang, Wenhao Zhu, Zihan Qiu, Fei Yuan, Jeff Z. Pan, Ivan Titov

Main category: cs.CL

TL;DR: LongBioBench is a new benchmark for evaluating long-context language models using artificially generated biographies, addressing limitations of existing evaluation frameworks by providing controlled, interpretable testing of understanding, reasoning, and trustworthiness.

DetailsMotivation: Existing long-context evaluation frameworks have limitations: real-world tasks are complex and suffer from data contamination, while synthetic tasks lack meaningful coherence between target information and context, undermining their validity as realistic proxies.

Method: The study introduces LongBioBench, which uses artificially generated biographies as a controlled environment for assessing LCLMs. It focuses on three essential features: seamless context, controllable setting, and sound evaluation across dimensions of understanding, reasoning, and trustworthiness.

Result: Evaluation of 18 LCLMs shows most models still exhibit deficiencies in semantic understanding and elementary reasoning over retrieved results, and become less trustworthy as context length increases. Analysis reveals existing synthetic benchmarks’ design choices make them vulnerable for testing long-context capabilities.

Conclusion: LongBioBench achieves a better trade-off between mirroring authentic language tasks and maintaining controllability compared to previous synthetic benchmarks, and is highly interpretable and configurable.

Abstract: Existing frameworks for evaluating long-context language models (LCLM) can be broadly categorized into real-world applications (e.g, document summarization) and synthetic tasks (e.g, needle-in-a-haystack). Despite their utility, both approaches are accompanied by certain intrinsic limitations. Real-world tasks often involve complexity that makes interpretation challenging and suffer from data contamination, whereas synthetic tasks frequently lack meaningful coherence between the target information (needle) and its surrounding context (haystack), undermining their validity as proxies for realistic applications. In response to these challenges, we posit that an ideal long-context evaluation framework should be characterized by three essential features: 1) seamless context 2) controllable setting and 3) sound evaluation. This study introduces $\textbf{LongBioBench}$, a benchmark that utilizes artificially generated biographies as a controlled environment for assessing LCLMs across dimensions of understanding, reasoning, and trustworthiness. Our experimental evaluation, which includes 18 LCLMs in total, demonstrates that most models still exhibit deficiencies in semantic understanding and elementary reasoning over retrieved results and are less trustworthy as context length increases. Our further analysis indicates some design choices employed by existing synthetic benchmarks, such as contextual non-coherence, numerical needles, and the absence of distractors, rendering them vulnerable to test the model’s long-context capabilities. To sum up, compared to previous synthetic benchmarks, LongBioBench achieves a better trade-off between mirroring authentic language tasks and maintaining controllability, and is highly interpretable and configurable.

[149] KG-Infused RAG: Augmenting Corpus-Based RAG with External Knowledge Graphs

Dingjun Wu, Yukun Yan, Zhenghao Liu, Zhiyuan Liu, Maosong Sun

Main category: cs.CL

TL;DR: KG-Infused RAG enhances traditional RAG by incorporating pre-existing knowledge graphs using spreading activation for better retrieval and generation, achieving significant performance improvements over vanilla RAG and other KG-based methods.

DetailsMotivation: Existing RAG methods either rely solely on text corpora (neglecting structural knowledge) or build ad-hoc knowledge graphs at high cost and low reliability, creating a need for more efficient structured knowledge integration.

Method: Proposes KG-Infused RAG framework that performs spreading activation over external knowledge graphs to retrieve structured knowledge, expands queries with this knowledge, integrates it with corpus passages, and uses preference learning on key pipeline stages.

Result: Outperforms vanilla RAG by 3.9% to 17.8% on five QA benchmarks, achieves superior performance compared to GraphRAG and LightRAG at lower cost, and shows further gains when integrated with Self-RAG and DeepNote.

Conclusion: KG-Infused RAG is an effective and versatile plug-and-play enhancement module that improves factual accuracy through structured knowledge integration while maintaining cost efficiency.

Abstract: Retrieval-Augmented Generation (RAG) improves factual accuracy by grounding responses in external knowledge. However, existing RAG methods either rely solely on text corpora and neglect structural knowledge, or build ad-hoc knowledge graphs (KGs) at high cost and low reliability. To address these issues, we propose KG-Infused RAG, a framework that incorporates pre-existing large-scale KGs into RAG and applies spreading activation to enhance both retrieval and generation. KG-Infused RAG directly performs spreading activation over external KGs to retrieve relevant structured knowledge, which is then used to expand queries and integrated with corpus passages, enabling interpretable and semantically grounded multi-source retrieval. We further improve KG-Infused RAG through preference learning on sampled key stages of the pipeline. Experiments on five QA benchmarks show that KG-Infused RAG consistently outperforms vanilla RAG (by 3.9% to 17.8%). Compared with KG-based approaches such as GraphRAG and LightRAG, our method obtains structured knowledge at lower cost while achieving superior performance. Additionally, integrating KG-Infused RAG with Self-RAG and DeepNote yields further gains, demonstrating its effectiveness and versatility as a plug-and-play enhancement module for corpus-based RAG methods.

[150] Code Execution as Grounded Supervision for LLM Reasoning

Dongwon Jung, Wenxuan Zhou, Muhao Chen

Main category: cs.CL

TL;DR: A scalable method for generating high-quality chain-of-thought supervision by extracting verifiable reasoning traces from code execution and converting them to natural language reasoning.

DetailsMotivation: Obtaining reliable and accurate reasoning supervision for LLMs is challenging, as existing methods rely on costly human annotations or error-prone LLM-generated chain-of-thought.

Method: Leverage the determinism of program execution to extract verifiable step-by-step reasoning traces from code and transform them into natural language chain-of-thought reasoning.

Result: Experiments show the method effectively equips LLMs with transferable reasoning abilities across diverse tasks, produces highly accurate reasoning data, and reduces token length during inference by eliminating meaningless repetition.

Conclusion: The proposed approach provides a scalable way to generate high-quality chain-of-thought supervision that enhances LLM reasoning capabilities while reducing inference costs.

Abstract: Training large language models (LLMs) with chain-of-thought (CoT) supervision has proven effective for enhancing their reasoning abilities. However, obtaining reliable and accurate reasoning supervision remains a significant challenge. We propose a scalable method for generating a high-quality CoT supervision dataset by leveraging the determinism of program execution. Unlike existing reasoning dataset generation methods that rely on costly human annotations or error-prone LLM-generated CoT, our approach extracts verifiable, step-by-step reasoning traces from code execution and transforms them into a natural language CoT reasoning. Experiments on reasoning benchmarks across various domains show that our method effectively equips LLMs with transferable reasoning abilities across diverse tasks. Furthermore, the ablation studies validate that our method produces highly accurate reasoning data and reduces overall token length during inference by reducing meaningless repetition and overthinking.

[151] From Multimodal Perception to Strategic Reasoning: A Survey on AI-Generated Game Commentary

Qirui Zheng, Xingbo Wang, Keyuan Cheng, Muhammad Asif Ali, Yunlong Lu, Wenxin Li

Main category: cs.CL

TL;DR: This survey paper provides a unified framework for AI-Generated Game Commentary (AI-GGC) research, introducing a novel taxonomy based on commentator capabilities and commentary types, while reviewing current methods, datasets, and evaluation metrics.

DetailsMotivation: Current AI-GGC research is fragmented and lacks a comprehensive survey to systematically unify existing efforts in this rapidly expanding field.

Method: The survey introduces a unified framework with a taxonomy focused on three core commentator capabilities (Live Observation, Strategic Analysis, Historical Recall) and three commentary types (Descriptive, Analytical, Background). It provides an in-depth review of state-of-the-art methods, datasets, and evaluation metrics.

Result: The paper organizes the fragmented AI-GGC landscape into a systematic framework and identifies key challenges and future research directions.

Conclusion: The survey bridges the gap in AI-GGC research by providing a comprehensive framework and highlighting challenges like real-time reasoning, multimodal integration, and evaluation bottlenecks for future development.

Abstract: The advent of artificial intelligence has propelled AI-Generated Game Commentary (AI-GGC) into a rapidly expanding field, offering benefits such as unlimited availability and personalized narration. However, current researches in this area remain fragmented, and a comprehensive survey that systematically unifies existing efforts is still missing. To bridge this gap, our survey introduces a unified framework that systematically organizes the AI-GGC landscape. We present a novel taxonomy focused on three core commentator capabilities: Live Observation, Strategic Analysis, and Historical Recall. Commentary is further categorized into three functional types: Descriptive, Analytical, and Background. Building on this structure, we provide an in-depth review of state-of-the-art methods, datasets, and evaluation metrics across various game genres. Finally, we highlight key challenges such as real-time reasoning, multimodal integration, and evaluation bottlenecks, and outline promising directions for future research and system development in AI-GGC.

[152] AnTKV: Anchor Token-Aware Sub-Bit Vector Quantization for KV Cache in Large Language Models

Zeyu Li, Chuanfu Xiao, Yang Wang, Xiang Liu, Zhenheng Tang, Baotong Lu, Mao Yang, Xinyu Chen, Xiaowen Chu

Main category: cs.CL

TL;DR: AnTKV is a dual-stage framework that uses anchor token-aware vector quantization to compress KV cache in LLMs, achieving ultra-low-bit quantization (down to 1-bit) while maintaining accuracy through selective preservation of sensitive tokens.

DetailsMotivation: To reduce memory footprint of KV cache in Large Language Models through quantization while minimizing accuracy degradation, especially in ultra-low-bit regimes where scalar quantization is limited.

Method: Proposes AnTKV framework with offline token-aware centroids learning and online anchor token selection using vector quantization. Introduces anchor score to measure token sensitivity and preserves top 1% most sensitive tokens.

Result: Achieves 1-bit quantization on Mistral-7B with perplexity of 6.32 (vs 7.25 for CQ and 15.36 for KVQuant). Enables LLaMA3-8B to scale to 840K tokens on single 80GB A100 with 3.5× higher decoding throughput than FP16 baseline.

Conclusion: AnTKV effectively balances compression and accuracy through anchor token-aware vector quantization, significantly outperforming prior methods in ultra-low-bit regimes while maintaining deployment efficiency.

Abstract: Quantization has emerged as an effective and lightweight solution to reduce the memory footprint of the KV cache in Large Language Models. Nevertheless, minimizing the accuracy degradation caused by ultra-low-bit KV cache quantization remains a significant challenge. While scalar quantization is constrained by 1-bit bound, vector quantization exploits intra-vector correlations and enables sub-bit regimes, making it more suitable for ultra-low-bit quantization. To further mitigate quantization-induced degradation, we reveal that the degradation is highly uneven across tokens in attention quality. To investigate this unevenness, we introduce anchor score to measure each token’s sensitivity to quantization. Our analysis and experiments show that preserving a small subset (1%) of tokens with the highest Anchor Score significantly mitigates accuracy loss under aggressive quantization. We propose AnTKV, a dual-stage framework that leverages anchor token-aware vector quantization to compress the KV cache. It combines offline token-aware centroids learning and online anchor token selection to balance compression and accuracy. To enable efficient deployment, we design an online anchor token selection kernel compatible with FlashAttention. It allows LLaMA3-8B to scale to 840K tokens on a single 80GB A100, while delivering up to $3.5\times$ higher decoding throughput over the FP16 baseline. Experiments demonstrate that AnTKV matches or surpasses prior methods at 4-bit, and significantly reduce perplexity under ultra-low-bit quantization, achieving 6.32 at 1-bit on Mistral-7B, compared to 7.25 for CQ and 15.36 for KVQuant.

[153] Compressed and Smooth Latent Space for Text Diffusion Modeling

Viacheslav Meshchaninov, Egor Chimbulatov, Alexander Shabalin, Aleksandr Abramov, Dmitry Vetrov

Main category: cs.CL

TL;DR: Cosmos is a novel text generation approach using diffusion models in a compressed latent space, achieving comparable quality to token-level diffusion models with 8x compression and 2x faster inference than autoregressive models.

DetailsMotivation: Autoregressive language models have limitations in slow decoding and global coherence, while diffusion models face challenges with high-dimensional token representations in text generation.

Method: Uses an autoencoder to learn a compressed latent space trained for token-level reconstruction and alignment with pretrained language encoder activations, enabling diffusion-based text generation.

Result: Achieves 8x compression while maintaining generation quality comparable to token-level diffusion models, and surpasses both diffusion and autoregressive baselines with longer latent sequences.

Conclusion: Cosmos demonstrates that compressed latent spaces enable efficient diffusion-based text generation with comparable or superior quality and significantly faster inference across multiple tasks.

Abstract: Autoregressive language models dominate modern text generation, yet their sequential nature introduces fundamental limitations: decoding is slow, and maintaining global coherence remains challenging. Diffusion models offer a promising alternative by enabling parallel generation and flexible control; however, their application to text generation is hindered by the high dimensionality of token-level representations. We introduce Cosmos, a novel approach to text generation that operates entirely in a compressed, smooth latent space tailored specifically for diffusion. This space is learned using an autoencoder trained simultaneously for token-level reconstruction and alignment with frozen activations from a pretrained language encoder, providing robust semantic grounding and enabling effective perturbation-based augmentations. Empirically, we demonstrate that text representations can be compressed by $8\times$ while maintaining generation quality comparable to token-level diffusion models. Furthermore, increasing the latent sequence length allows Cosmos to surpass both diffusion-based and autoregressive baselines. We evaluate Cosmos on four diverse generative tasks including story generation, question generation, summarization, and detoxification and compare it with various generative paradigms. Cosmos achieves comparable or superior generation quality while offering more than $2\times$ faster inference. Code is released at \href{https://github.com/MeshchaninovViacheslav/cosmos}{GitHub}

[154] Don’t Trust Generative Agents to Mimic Communication on Social Networks Unless You Benchmarked their Empirical Realism

Simon Münker, Nils Schwager, Achim Rettinger

Main category: cs.CL

TL;DR: This paper examines the use of LLMs to simulate social network user behavior, finding that social simulations need empirical validation and arguing for more rigor in generative-agent-based modeling.

DetailsMotivation: To better understand the conflicting research findings about using AI agents for human behavior studies, specifically focusing on social network user communication simulation.

Method: Developed a formal framework for social network simulation and empirically tested different approaches to imitate user behavior on X platform in both English and German languages.

Result: Findings suggest that social simulations should be validated by their empirical realism measured in the setting where simulation components were fitted.

Conclusion: The paper argues for more rigor when applying generative-agent-based modeling for social simulation, emphasizing the need for proper validation of simulated behaviors.

Abstract: The ability of Large Language Models (LLMs) to mimic human behavior triggered a plethora of computational social science research, assuming that empirical studies of humans can be conducted with AI agents instead. Since there have been conflicting research findings on whether and when this hypothesis holds, there is a need to better understand the differences in their experimental designs. We focus on replicating the behavior of social network users with the use of LLMs for the analysis of communication on social networks. First, we provide a formal framework for the simulation of social networks, before focusing on the sub-task of imitating user communication. We empirically test different approaches to imitate user behavior on X in English and German. Our findings suggest that social simulations should be validated by their empirical realism measured in the setting in which the simulation components were fitted. With this paper, we argue for more rigor when applying generative-agent-based modeling for social simulation.

[155] Lost at the Beginning of Reasoning

Baohao Liao, Xinyi Chen, Sara Rajaee, Yuhui Xu, Christian Herold, Anders Søgaard, Maarten de Rijke, Christof Monz

Main category: cs.CL

TL;DR: The first reasoning step in chain-of-thought reasoning has disproportionate influence on final predictions, and a reward model-based sampling strategy can reduce inference costs by 70% without accuracy loss.

DetailsMotivation: Self-correction abilities of LLMs during long chain-of-thought reasoning remain underexplored, and models often engage in unnecessarily redundant reasoning (overthinking).

Method: Propose an efficient sampling strategy that uses a reward model to identify and retain high-quality first reasoning steps while discarding suboptimal ones.

Result: Achieved up to 70% reduction in inference cost without sacrificing any accuracy, consistently observed across various state-of-the-art reasoning models.

Conclusion: The first reasoning step plays a central role in generating high-quality reasoning trajectories, enabling significantly efficient sampling.

Abstract: Recent advancements in large language models (LLMs) have significantly advanced complex reasoning capabilities, particularly through extended chain-of-thought (CoT) reasoning that incorporates mechanisms such as backtracking, self-reflection, and self-correction. Despite these developments, the self-correction abilities of LLMs during long CoT reasoning remain underexplored. And recent findings on overthinking suggest that such models often engage in unnecessarily redundant reasoning. In this work, we empirically show that the first reasoning step exerts a disproportionately large influence on the final prediction. I.e., errors introduced at this stage can substantially degrade subsequent reasoning quality. This phenomenon is consistently observed across various state-of-the-art open- and closed-source reasoning models. Leveraging this insight, we propose an efficient sampling strategy that leverages a reward model to identify and retain high-quality first reasoning steps while discarding suboptimal ones, achieving up to a 70% reduction in inference cost without sacrificing any accuracy. Our work highlights the central role of the first reasoning step in generating a high-quality reasoning trajectory, and thus enabling significantly efficient sampling.

[156] DP-Fusion: Token-Level Differentially Private Inference for Large Language Models

Rushil Thareja, Preslav Nakov, Praneeth Vepakomma, Nils Lukas

Main category: cs.CL

TL;DR: DP-Fusion is a differentially private inference mechanism for LLMs that provably bounds the influence of sensitive tokens in context, enabling document privatization while maintaining text quality.

DetailsMotivation: LLMs can inadvertently reveal sensitive information from their context during inference, creating privacy risks when augmented with tools or databases containing private data. Existing privacy methods lack provable guarantees or have poor utility/privacy trade-offs.

Method: Four-step process: (1) label sensitive tokens, (2) infer LLM without sensitive tokens for baseline, (3) infer LLM with sensitive tokens, (4) blend distributions to bound distance from baseline. Privacy/utility controlled by epsilon parameter.

Result: Achieves substantially improved theoretical and empirical privacy with 6× lower perplexity than related differentially private inference methods, creating token-level provably privatized documents.

Conclusion: DP-Fusion provides a practical solution for document privatization that balances privacy protection with text quality through provable differential privacy guarantees.

Abstract: Large language models (LLMs) do not preserve privacy at inference-time. The LLM’s outputs can inadvertently reveal information about the model’s context, which presents a privacy challenge when the LLM is augmented via tools or databases containing sensitive information. Existing privacy-preserving methods at inference-time have significant limitations since they (i) lack provable guarantees or (ii) have a poor utility/privacy trade-off. We propose DP-Fusion, a Differentially Private Inference (DPI) mechanism for LLMs that provably bounds the influence a set of tokens in the context can have on the LLM’s output. DP-Fusion works as follows: (1) label a subset of sensitive tokens, (2) infer the LLM without any sensitive tokens to obtain a baseline, (3) infer the LLM with the sensitive tokens, and (4) blend distributions so that the final output remains within a bounded distance of the baseline distribution. While this per-token influence bound also mitigates jailbreak-style prompt injection, we focus on \emph{document privatization}, where the goal is to paraphrase a document containing sensitive tokens, e.g., personally identifiable information, so that no attacker can reliably infer them from the paraphrased document while preserving high text quality. The privacy/utility trade-off is controlled by $\epsilon$, where $\epsilon=0$ hides sensitive tokens entirely, while higher values trade off privacy for improved text quality. We show that our method creates token-level provably privatized documents with substantially improved theoretical and empirical privacy, achieving $6\times$ lower perplexity than related DPI methods.

[157] Controlling What You Share: Assessing Language Model Adherence to Privacy Preferences

Guillem Ramírez, Alexandra Birch, Ivan Titov

Main category: cs.CL

TL;DR: A framework using privacy profiles (natural language instructions) to rewrite user queries before sending to external LLMs, balancing privacy and performance while keeping data under user control.

DetailsMotivation: Users need to expose data to commercial LLM APIs, compromising privacy. Privacy profiles allow users to control what information is revealed while maintaining API functionality.

Method: Use lightweight local LLMs fine-tuned with privacy profiles to rewrite queries, hiding sensitive content before sending to external models. Built PEEP dataset for training.

Result: Fine-tuned lightweight LLMs achieve better privacy preservation and match/exceed performance of larger zero-shot models, but still struggle with fully adhering to user instructions.

Conclusion: The approach effectively balances privacy and performance, but better understanding of user-defined privacy preferences is needed for full instruction adherence.

Abstract: Large language models (LLMs) are primarily accessed via commercial APIs, but this often requires users to expose their data to service providers. In this paper, we explore how users can stay in control of their data by using privacy profiles: simple natural language instructions that say what should and should not be revealed. We build a framework where a local model uses these instructions to rewrite queries, only hiding details deemed sensitive by the user, before sending them to an external model, thus balancing privacy with performance. To support this research, we introduce PEEP, a multilingual dataset of real user queries annotated to mark private content and paired with synthetic privacy profiles. Experiments with lightweight local LLMs show that, after fine-tuning, they not only achieve markedly better privacy preservation but also match or exceed the performance of much larger zero-shot models. At the same time, the system still faces challenges in fully adhering to user instructions, underscoring the need for models with a better understanding of user-defined privacy preferences.

[158] From Sequence to Structure: Uncovering Substructure Reasoning in Transformers

Xinnan Dai, Kai Yang, Jay Revolinsky, Kai Guo, Aoran Wang, Bohang Zhang, Jiliang Tang

Main category: cs.CL

TL;DR: LLMs can understand graph structures from text through a process called Induced Substructure Filtration (ISF), where transformers identify and extract substructures from graph data embedded in sequences.

DetailsMotivation: To understand how decoder-only Transformer architectures can comprehend underlying graph structures when graphs are represented as text, particularly focusing on substructure extraction tasks.

Method: Proposed Induced Substructure Filtration (ISF) perspective through empirical results and theoretical analysis, analyzing multi-layer transformers’ internal mechanisms and input query impact. Also introduced ’thinking in substructures’ concept for complex pattern extraction.

Result: Validated ISF process in LLMs, revealing consistent internal dynamics across layers. Demonstrated that decoder-only Transformers can successfully extract substructures from attributed graphs like molecular graphs.

Conclusion: Sequence-based Transformers can perform substructure extraction over graph data through the ISF process, providing new insights into how they understand graph structures from textual representations.

Abstract: Recent studies suggest that large language models (LLMs) possess the capability to solve graph reasoning tasks. Notably, even when graph structures are embedded within textual descriptions, LLMs can still effectively answer related questions. This raises a fundamental question: How can a decoder-only Transformer architecture understand underlying graph structures? To address this, we start with the substructure extraction task, interpreting the inner mechanisms inside the transformers and analyzing the impact of the input queries. Specifically, through both empirical results and theoretical analysis, we present Induced Substructure Filtration (ISF), a perspective that captures the substructure identification in the multi-layer transformers. We further validate the ISF process in LLMs, revealing consistent internal dynamics across layers. Building on these insights, we explore the broader capabilities of Transformers in handling diverse graph types. Specifically, we introduce the concept of thinking in substructures to efficiently extract complex composite patterns, and demonstrate that decoder-only Transformers can successfully extract substructures from attributed graphs, such as molecular graphs. Together, our findings offer a new insight on how sequence-based Transformers perform the substructure extraction task over graph data.

[159] FinResearchBench: A Logic Tree based Agent-as-a-Judge Evaluation Framework for Financial Research Agents

Rui Sun, Zuo Bai, Wentao Zhang, Yuxiang Zhang, Li Zhao, Shan Sun, Zhengwen Qiu

Main category: cs.CL

TL;DR: FinResearchBench is the first logic tree-based Agent-as-a-Judge system for evaluating financial research agents across 7 key task types, addressing the lack of systematic benchmarks for AI research agents in finance.

DetailsMotivation: There are few evaluation frameworks and benchmarks that systematically and automatically investigate the capabilities of deep research agents, particularly in the financial domain which has distinct complexity and subtlety.

Method: Proposes a logic tree-based Agent-as-a-Judge system that extracts the logic tree of research outcomes as intermediate information for comprehensive, reliable, and robust evaluation.

Result: FinResearchBench covers 70 typical financial research questions across 7 frequently encountered task types in the financial research domain.

Conclusion: The work fills the gap in evaluating financial research agents by providing the first innovative Agent-as-a-Judge system specifically designed for the financial domain with comprehensive coverage of typical research tasks.

Abstract: Recently, AI agents are rapidly evolving in intelligence and widely used in professional research applications, such as STEM, software development, and finance. Among these AI agents, deep research agent is a key category as it can perform long-horizon tasks and solve problems of greater complexity. However, there are few evaluation frameworks and benchmarks that systematically and automatically investigate the capabilities of these research agents. In addition, financial research problems have distinct complexity and subtlety. To fill in the gap, we propose FinResearchBench, which is a logic tree-based Agent-as-a-Judge and targets specifically for the financial research agents. It provides a comprehensive and automatic assessment of the research agents across 7 key types of tasks in the financial research domain. The contributions of this work are two-folded: (1) the first and innovative Agent-as-a-Judge system that extracts the logic tree of the research outcome and uses it as the intermediate information to present a comprehensive, reliable, and robust evaluation; (2) finance-oriented that it covers 70 typical financial research questions, spreading across 7 frequently encountered types of task in the domain.

[160] Geometric-Mean Policy Optimization

Yuzhong Zhao, Yue Liu, Junpeng Liu, Jingye Chen, Xun Wu, Yaru Hao, Tengchao Lv, Shaohan Huang, Lei Cui, Qixiang Ye, Fang Wan, Furu Wei

Main category: cs.CL

TL;DR: GMPO improves GRPO by using geometric mean instead of arithmetic mean for token-level rewards, reducing sensitivity to outliers and stabilizing policy updates in language model training.

DetailsMotivation: GRPO suffers from unstable policy updates due to outlier token rewards that cause extreme importance sampling ratios during training.

Method: Replace GRPO’s arithmetic mean with geometric mean of token-level rewards, which is inherently less sensitive to outliers and maintains more stable importance sampling ratios.

Result: GMPO-7B improves average Pass@1 by up to 4.1% over GRPO on multiple mathematical reasoning benchmarks, outperforming state-of-the-art approaches.

Conclusion: GMPO is a plug-and-play improvement over GRPO that provides more stable policy optimization through geometric mean aggregation, leading to better reasoning performance.

Abstract: Group Relative Policy Optimization (GRPO) has significantly enhanced the reasoning capability of large language models by optimizing the arithmetic mean of token-level rewards. Unfortunately, GRPO is observed to suffer from unstable policy updates when facing tokens with outlier importance-weighted rewards, which manifest as extreme importance sampling ratios during training. In this study, we propose Geometric-Mean Policy Optimization (GMPO), with the aim to improve the stability of GRPO through suppressing token reward outliers. Instead of optimizing the arithmetic mean, GMPO maximizes the geometric mean of token-level rewards, which is inherently less sensitive to outliers and maintains a more stable range of importance sampling ratio. GMPO is plug-and-play-simply replacing GRPO’s arithmetic mean with the geometric mean of token-level rewards, as the latter is inherently less sensitive to outliers. GMPO is theoretically plausible-analysis reveals that both GMPO and GRPO are weighted forms of the policy gradient while the former enjoys more stable weights, which consequently benefits policy optimization and performance. Experiments on multiple mathematical reasoning benchmarks show that GMPO-7B improves the average Pass@1 of GRPO by up to 4.1%, outperforming many state-of-the-art approaches. Code is available at https://github.com/callsys/GMPO.

[161] CultureGuard: Towards Culturally-Aware Dataset and Guard Model for Multilingual Safety Applications

Raviraj Joshi, Rakesh Paul, Kanishk Singla, Anusha Kamath, Michael Evans, Katherine Luna, Shaona Ghosh, Utkarsh Vaidya, Eileen Long, Sanjay Singh Chauhan, Niranjan Wartikar

Main category: cs.CL

TL;DR: CultureGuard introduces a pipeline to create multilingual safety datasets for LLMs, addressing the lack of culturally aligned non-English safety data through synthetic generation and filtering.

DetailsMotivation: Non-English languages lack robust safety guard models for LLMs due to high costs of collecting culturally aligned datasets, while English content safety is well-studied.

Method: Four-stage pipeline: cultural data segregation, cultural data adaptation, machine translation, and quality filtering to expand English safety dataset to 8 languages.

Result: Created Nemotron-Safety-Guard-Dataset-v3 with 386,661 samples in 9 languages; trained Llama-3.1-Nemotron-Safety-Guard-8B-v3 model achieves SOTA performance on multilingual benchmarks with strong cross-lingual transfer.

Conclusion: This work advances multilingual LLM safety by enabling development of culturally aware safety guard models, showing current LLMs are more prone to unsafe responses in non-English languages.

Abstract: The increasing use of Large Language Models (LLMs) in agentic applications highlights the need for robust safety guard models. While content safety in English is well-studied, non-English languages lack similar advancements due to the high cost of collecting culturally aligned labeled datasets. We present CultureGuard, a novel solution for curating culturally aligned, high-quality safety datasets across multiple languages. Our approach introduces a four-stage synthetic data generation and filtering pipeline: cultural data segregation, cultural data adaptation, machine translation, and quality filtering. This pipeline enables the conversion and expansion of the Nemotron-Content-Safety-Dataset-V2 English safety dataset into eight distinct languages: Arabic, German, Spanish, French, Hindi, Japanese, Thai, and Chinese. The resulting dataset, Nemotron-Safety-Guard-Dataset-v3, comprises 386,661 samples in 9 languages and facilitates the training of Llama-3.1-Nemotron-Safety-Guard-8B-v3 via LoRA-based fine-tuning. The final model achieves state-of-the-art performance on several multilingual content safety benchmarks. Furthermore, we show our moderately multilingual fine-tuning enables robust cross-lingual transfer and strong zero-shot generalization to unseen languages. We also benchmark the latest open LLMs on multilingual safety and observe that these LLMs are more prone to give unsafe responses when prompted in non-English languages. This work advances multilingual LLM safety by enabling the development of culturally aware safety guard models.

[162] “Mirror” Language AI Models of Depression are Criterion-Contaminated

Tong Li, Rasiq Hussain, Mehak Gupta, Joshua R. Oltmanns

Main category: cs.CL

TL;DR: The study compares ‘Mirror’ models (using language from depression assessments to predict scores) with ‘Non-Mirror’ models (using external language) and finds Mirror models suffer from criterion contamination, while Non-Mirror models show strong predictive power and may provide more valid clinical applications.

DetailsMotivation: To address the problem of criterion contamination in existing language-based depression prediction models that use assessment responses to predict assessment scores, which artificially inflates prediction accuracy.

Method: Compared Mirror vs Non-Mirror models using 110 participants who completed both structured diagnostic interviews (Mirror condition) and life history interviews (Non-Mirror condition), with LLMs prompted to predict depression scores.

Result: Mirror models showed near-perfect prediction (R² = .70), but Non-Mirror models also displayed large prediction sizes. Both model types correlated similarly with other depression questionnaires, suggesting bias in Mirror models. Topic modeling revealed different theme structures across model types.

Conclusion: Incorporating Non-Mirror approaches may support more valid and clinically useful language-based AI applications in psychological assessment, as they avoid criterion contamination while maintaining predictive power.

Abstract: Recent studies show near-perfect language-based predictions of depression scores (R2 = .70), but these “Mirror” models rely on language responses directly from depression assessments to predict depression assessment scores. These methods suffer from criterion contamination that inflate prediction estimates. We compare “Mirror” models to “Non-Mirror” models, which use other external language to predict depression scores. 110 participants completed both structured diagnostic (Mirror condition) and life history (Non-Mirror condition) interviews. LLMs were prompted to predict diagnostic depression scores. As expected, Mirror models were near-perfect. However, Non-Mirror models also displayed prediction sizes considered large in psychology. Further, both Mirror and Non-Mirror predictions correlated with other questionnaire-based depression symptoms at similar sizes, suggesting bias in Mirror models. Topic modeling revealed different theme structures across model types. As language models for depression continue to evolve, incorporating Non-Mirror approaches may support more valid and clinically useful language-based AI applications in psychological assessment.

[163] CorrSteer: Generation-Time LLM Steering via Correlated Sparse Autoencoder Features

Seonglae Cho, Zekun Wu, Adriano Koshiyama

Main category: cs.CL

TL;DR: CorrSteer: A correlation-based method for selecting sparse autoencoder features using inference-time activations to improve LLM steering without contrastive datasets or large storage.

DetailsMotivation: Current SAE methods for LLM steering require contrastive datasets or large activation storage, limiting their practical effectiveness in downstream tasks.

Method: Correlate sample correctness with SAE activations from generated tokens at inference time, using only inference-time activations to extract relevant features and obtain steering coefficients from average activations.

Result: Improved performance on QA, bias mitigation, jailbreaking prevention, and reasoning benchmarks on Gemma-2 2B and LLaMA-3.1 8B, with +3.3% MMLU improvement (4000 samples) and +27.2% HarmBench improvement (108 samples). Selected features show semantically meaningful patterns.

Conclusion: Correlation-based selection is an effective and scalable approach for automated SAE steering across language model applications.

Abstract: Sparse Autoencoders (SAEs) can extract interpretable features from large language models (LLMs) without supervision. However, their effectiveness in downstream steering tasks is limited by the requirement for contrastive datasets or large activation storage. To address these limitations, we propose CorrSteer, which selects features by correlating sample correctness with SAE activations from generated tokens at inference time. This approach uses only inference-time activations to extract more relevant features, thereby reducing spurious correlations. It also obtains steering coefficients from average activations, automating the entire pipeline. Our method shows improved task performance on QA, bias mitigation, jailbreaking prevention, and reasoning benchmarks on Gemma-2 2B and LLaMA-3.1 8B, notably achieving a +3.3% improvement in MMLU performance with 4000 samples and a +27.2% improvement in HarmBench with only 108 samples. Selected features demonstrate semantically meaningful patterns aligned with each task’s requirements, revealing the underlying capabilities that drive performance. Our work establishes correlation-based selection as an effective and scalable approach for automated SAE steering across language model applications.

[164] A Survey of Scientific Large Language Models: From Data Foundations to Agent Frontiers

Ming Hu, Chenglong Ma, Wei Li, Wanghan Xu, Jiamin Wu, Jucheng Hu, Tianbin Li, Guohang Zhuang, Jiaqi Liu, Yingzhou Lu, Ying Chen, Chaoyang Zhang, Cheng Tan, Jie Ying, Guocheng Wu, Shujian Gao, Pengcheng Chen, Jiashi Lin, Haitao Wu, Lulu Chen, Fengxiang Wang, Yuanyuan Zhang, Xiangyu Zhao, Feilong Tang, Encheng Su, Junzhi Ning, Xinyao Liu, Ye Du, Changkai Ji, Pengfei Jiang, Cheng Tang, Ziyan Huang, Jiyao Liu, Jiaqi Wei, Yuejin Yang, Xiang Zhang, Guangshuai Wang, Yue Yang, Huihui Xu, Ziyang Chen, Yizhou Wang, Chen Tang, Jianyu Wu, Yuchen Ren, Siyuan Yan, Zhonghua Wang, Zhongxing Xu, Shiyan Su, Shangquan Sun, Runkai Zhao, Zhisheng Zhang, Dingkang Yang, Jinjie Wei, Jiaqi Wang, Jiahao Xu, Jiangtao Yan, Wenhao Tang, Hongze Zhu, Yu Liu, Fudi Wang, Yiqing Shen, Yuanfeng Ji, Yanzhou Su, Tong Xie, Hongming Shan, Chun-Mei Feng, Zhi Hou, Diping Song, Lihao Liu, Yanyan Huang, Lequan Yu, Bin Fu, Shujun Wang, Xiaomeng Li, Xiaowei Hu, Yun Gu, Ben Fei, Benyou Wang, Yuewen Cao, Minjie Shen, Jie Xu, Haodong Duan, Fang Yan, Hongxia Hao, Jielan Li, Jiajun Du, Yanbo Wang, Imran Razzak, Zhongying Deng, Chi Zhang, Lijun Wu, Conghui He, Zhaohui Lu, Jinhai Huang, Wenqi Shao, Yihao Liu, Siqi Luo, Yi Xin, Xiaohong Liu, Fenghua Ling, Yuqiang Li, Aoran Wang, Siqi Sun, Qihao Zheng, Nanqing Dong, Tianfan Fu, Dongzhan Zhou, Yan Lu, Wenlong Zhang, Jin Ye, Jianfei Cai, Yirong Chen, Wanli Ouyang, Yu Qiao, Zongyuan Ge, Shixiang Tang, Junjun He, Chunfeng Song, Lei Bai, Bowen Zhou

Main category: cs.CL

TL;DR: This survey reframes Sci-LLM development as co-evolution between models and data, presenting unified taxonomies for scientific data and knowledge, analyzing over 270 datasets and 190 benchmarks, and outlining a shift toward closed-loop AI systems for scientific discovery.

DetailsMotivation: Scientific LLMs face unique challenges due to the complex nature of scientific data - multimodal, cross-scale, and domain-specific - requiring different approaches than general NLP datasets.

Method: Comprehensive data-centric synthesis with unified taxonomies of scientific data and knowledge, systematic review of Sci-LLMs across disciplines, analysis of 270+ datasets, and examination of 190+ benchmark datasets.

Result: Identifies distinct demands of Sci-LLMs including heterogeneous, multi-scale, uncertainty-laden corpora requiring domain-invariant representations and cross-modal reasoning. Shows shift from static to process-oriented evaluation.

Conclusion: Outlines paradigm shift toward closed-loop systems where autonomous Sci-LLM agents actively experiment and contribute to evolving knowledge bases, providing roadmap for trustworthy AI partners in scientific discovery.

Abstract: Scientific Large Language Models (Sci-LLMs) are transforming how knowledge is represented, integrated, and applied in scientific research, yet their progress is shaped by the complex nature of scientific data. This survey presents a comprehensive, data-centric synthesis that reframes the development of Sci-LLMs as a co-evolution between models and their underlying data substrate. We formulate a unified taxonomy of scientific data and a hierarchical model of scientific knowledge, emphasizing the multimodal, cross-scale, and domain-specific challenges that differentiate scientific corpora from general natural language processing datasets. We systematically review recent Sci-LLMs, from general-purpose foundations to specialized models across diverse scientific disciplines, alongside an extensive analysis of over 270 pre-/post-training datasets, showing why Sci-LLMs pose distinct demands – heterogeneous, multi-scale, uncertainty-laden corpora that require representations preserving domain invariance and enabling cross-modal reasoning. On evaluation, we examine over 190 benchmark datasets and trace a shift from static exams toward process- and discovery-oriented assessments with advanced evaluation protocols. These data-centric analyses highlight persistent issues in scientific data development and discuss emerging solutions involving semi-automated annotation pipelines and expert validation. Finally, we outline a paradigm shift toward closed-loop systems where autonomous agents based on Sci-LLMs actively experiment, validate, and contribute to a living, evolving knowledge base. Collectively, this work provides a roadmap for building trustworthy, continually evolving artificial intelligence (AI) systems that function as a true partner in accelerating scientific discovery.

[165] BED-LLM: Intelligent Information Gathering with LLMs and Bayesian Experimental Design

Deepro Choudhury, Sinead Williamson, Adam Goliński, Ning Miao, Freddie Bickford Smith, Michael Kirchhof, Yizhe Zhang, Tom Rainforth

Main category: cs.CL

TL;DR: BED-LLM improves LLMs’ ability to gather information through sequential Bayesian experimental design, enabling intelligent multi-turn conversations by selecting questions that maximize expected information gain.

DetailsMotivation: To enhance LLMs' capability as conversational agents that can intelligently and adaptively gather information from users or external sources, rather than relying on direct prompting.

Method: Uses sequential Bayesian experimental design framework where LLMs iteratively choose questions that maximize expected information gain about the task, based on probabilistic models derived from the LLM’s predictive distributions.

Result: Achieves substantial performance gains across 20 questions game and user preference inference tasks compared to direct prompting and other adaptive strategies.

Conclusion: BED-LLM provides a principled approach for LLMs to act as effective multi-turn conversational agents by intelligently gathering information through Bayesian experimental design.

Abstract: We propose a general-purpose approach for improving the ability of Large Language Models (LLMs) to intelligently and adaptively gather information from a user or other external source using the framework of sequential Bayesian experimental design (BED). This enables LLMs to act as effective multi-turn conversational agents and interactively interface with external environments. Our approach, which we call BED-LLM (Bayesian Experimental Design with Large Language Models), is based on iteratively choosing questions or queries that maximize the expected information gain (EIG) about the task of interest given the responses gathered previously. We show how this EIG can be formulated (and then estimated) in a principled way using a probabilistic model derived from the LLM’s predictive distributions and provide detailed insights into key decisions in its construction and updating procedure. We find that BED-LLM achieves substantial gains in performance across a wide range of tests based on the 20 questions game and using the LLM to actively infer user preferences, compared to direct prompting of the LLM and other adaptive design strategies.

[166] Supervised In-Context Fine-Tuning for Generative Sequence Labeling

David Dukić, Goran Glavaš, Jan Šnajder

Main category: cs.CL

TL;DR: Proposes SIFT (supervised in-context fine-tuning) for generative sequence labeling, combining ICL with supervised fine-tuning to outperform existing methods on standard SL tasks.

DetailsMotivation: Sequence labeling tasks are typically handled by encoder-only models, but causal LLMs are expected to outperform them due to rapid scaling. Less work has focused on generative SL, which is more natural for causal LLMs.

Method: SIFT casts SL tasks as constrained response generation, combining in-context learning from demonstrations with supervised fine-tuning of causal LLMs.

Result: SIFT considerably outperforms both ICL and decoder-as-encoder fine-tuning baselines on standard SL tasks. Removing instructions improves performance as they’re largely unnecessary for strong SL performance with SIFT.

Conclusion: Response-based generative task formulation is crucial for effective sequence labeling with LLMs, highlighting both strengths and limitations of SL with LLMs.

Abstract: Sequence labeling (SL) tasks, where labels are assigned to tokens, are abundant in NLP (e.g., named entity recognition and aspect-based sentiment analysis). Owing to the intuition that they require bidirectional context, SL tasks are commonly tackled with encoder-only models. Recent work also shows that removing the causal mask in fine-tuning enables decoder-based LLMs to become effective token classifiers. Less work, however, focused on (supervised) generative SL, a more natural setting for causal LLMs. Due to their rapid scaling, causal LLMs applied to SL are expected to outperform encoders, whose own development has stagnated. In this work, we propose supervised in-context fine-tuning (SIFT) for generative SL. SIFT casts SL tasks as constrained response generation, natural to LLMs, combining in-context learning (ICL) from demonstrations with supervised fine-tuning. SIFT considerably outperforms both ICL and decoder-as-encoder fine-tuning baselines on a range of standard SL tasks. We further find that although long context hinders the performance of generative SL in both ICL and SIFT, this deficiency can be mitigated by removing the instruction, as instructions are shown to be largely unnecessary for achieving strong SL performance with SIFT. Our findings highlight strengths and limitations of SL with LLMs, underscoring the importance of a response-based generative task formulation for effective SL performance.

[167] From Scarcity to Efficiency: Investigating the Effects of Data Augmentation on African Machine Translation

Mardiyyah Oduwole, Oluwatosin Olajide, Jamiu Suleiman, Faith Hunja, Busayo Awobade, Fatimo Adebanjo, Comfort Akanni, Chinonyelum Igwe, Peace Ododo, Promise Omoigui, Abraham Owodunni, Steven Kolawole

Main category: cs.CL

TL;DR: Data augmentation techniques (sentence concatenation with back translation and switch-out) significantly improve machine translation performance for low-resource African languages, with minimum 25% BLEU score increase across six languages.

DetailsMotivation: Linguistic diversity across Africa presents challenges for machine translation, especially for low-resource languages that lack sufficient training data.

Method: Applied two data augmentation techniques: sentence concatenation with back translation and switch-out, tested across six African languages.

Result: Significant improvements in machine translation performance with minimum 25% BLEU score increase across all six languages tested.

Conclusion: Data augmentation techniques show strong potential to improve machine translation systems for low-resource languages, contributing to more robust translation systems for under-resourced languages.

Abstract: The linguistic diversity across the African continent presents different challenges and opportunities for machine translation. This study explores the effects of data augmentation techniques in improving translation systems in low-resource African languages. We focus on two data augmentation techniques: sentence concatenation with back translation and switch-out, applying them across six African languages. Our experiments show significant improvements in machine translation performance, with a minimum increase of 25% in BLEU score across all six languages. We provide a comprehensive analysis and highlight the potential of these techniques to improve machine translation systems for low-resource languages, contributing to the development of more robust translation systems for under-resourced languages.

[168] Creativity Benchmark: A benchmark for marketing creativity for large language models

Ninad Bhat, Kieran Browne, Pip Bingemann

Main category: cs.CL

TL;DR: Creativity Benchmark evaluates LLMs in marketing creativity across 100 brands and 3 prompt types, showing tightly clustered performance with no dominant model and highlighting limitations of automated evaluation.

DetailsMotivation: To systematically evaluate large language models' creative capabilities in marketing contexts, addressing the need for domain-specific creativity assessment beyond conventional tests.

Method: Human pairwise preferences from 678 creatives analyzed with Bradley-Terry models, covering 100 brands across 12 categories and three prompt types (Insights, Ideas, Wild Ideas), plus analysis of model diversity and LLM-as-judge setups.

Result: Models show tightly clustered performance (Δθ≈0.45, head-to-head win probability 0.61), no model dominates across brands or prompt types, weak correlations between LLM judges and human rankings, and partial transfer of conventional creativity tests.

Conclusion: Expert human evaluation remains essential for marketing creativity assessment, automated judges cannot substitute humans, and diversity-aware workflows are needed due to limited model dominance and inconsistent performance.

Abstract: We introduce Creativity Benchmark, an evaluation framework for large language models (LLMs) in marketing creativity. The benchmark covers 100 brands (12 categories) and three prompt types (Insights, Ideas, Wild Ideas). Human pairwise preferences from 678 practising creatives over 11,012 anonymised comparisons, analysed with Bradley-Terry models, show tightly clustered performance with no model dominating across brands or prompt types: the top-bottom spread is $\Delta\theta \approx 0.45$, which implies a head-to-head win probability of $0.61$; the highest-rated model beats the lowest only about $61%$ of the time. We also analyse model diversity using cosine distances to capture intra- and inter-model variation and sensitivity to prompt reframing. Comparing three LLM-as-judge setups with human rankings reveals weak, inconsistent correlations and judge-specific biases, underscoring that automated judges cannot substitute for human evaluation. Conventional creativity tests also transfer only partially to brand-constrained tasks. Overall, the results highlight the need for expert human evaluation and diversity-aware workflows.

[169] RPG: A Repository Planning Graph for Unified and Scalable Codebase Generation

Jane Luo, Xin Zhang, Steven Liu, Jie Wu, Jianfeng Liu, Yiming Huang, Yangyu Huang, Chengyu Yin, Ying Xin, Yuefeng Zhan, Hao Sun, Qi Chen, Scarlett Li, Mao Yang

Main category: cs.CL

TL;DR: ZeroRepo introduces Repository Planning Graph (RPG) to enable structured repository generation, achieving 3.9x larger code output and significant improvements in test coverage and accuracy compared to baselines.

DetailsMotivation: Current approaches rely on ambiguous natural language planning for repository generation, leading to unclear specifications, misaligned components, and brittle designs. There's a need for structured planning to generate complete repositories from scratch.

Method: ZeroRepo uses a graph-driven framework with three stages: proposal-level planning, implementation-level construction, and graph-guided code generation with test validation. It employs Repository Planning Graph (RPG) as a structured blueprint encoding capabilities, file structures, data flows, and functions.

Result: On RepoCraft benchmark (6 real-world projects, 1,052 tasks), ZeroRepo generated nearly 36K Code Lines and 445K Code Tokens, 3.9x larger than Claude Code baseline. Achieved 81.5% coverage and 69.7% test accuracy, improving over Claude Code by 27.3 and 35.8 points.

Conclusion: RPG enables consistent long-horizon planning for repository generation, models complex dependencies, allows near-linear scaling for sophisticated planning, and improves agent understanding of repositories for faster localization.

Abstract: Large language models excel at generating individual functions or single files of code, yet generating complete repositories from scratch remains a fundamental challenge. This capability is key to building coherent software systems from high-level specifications and realizing the full potential of automated code generation. The process requires planning at two levels: deciding what features and modules to build (proposal stage) and defining their implementation details (implementation stage). Current approaches rely on natural language planning, which often produces unclear specifications, misaligned components, and brittle designs due to its inherent ambiguity and lack of structure. To address these limitations, we introduce the Repository Planning Graph (RPG), a structured representation that encodes capabilities, file structures, data flows, and functions in a unified graph. By replacing free-form natural language with an explicit blueprint, RPG enables consistent long-horizon planning for repository generation. Building on RPG, we develop ZeroRepo, a graph-driven framework that operates in three stages: proposal-level planning, implementation-level construction, and graph-guided code generation with test validation. To evaluate, we construct RepoCraft, a benchmark of six real-world projects with 1,052 tasks. On RepoCraft, ZeroRepo produces nearly 36K Code Lines and 445K Code Tokens, on average 3.9$\times$ larger than the strongest baseline (Claude Code), and 68$\times$ larger than other baselines. It achieves 81.5% coverage and 69.7% test accuracy, improving over Claude Code by 27.3 and 35.8 points. Further analysis shows that RPG models complex dependencies, enables more sophisticated planning through near-linear scaling, and improves agent understanding of repositories, thus accelerating localization.

[170] Semantic Representation Attack against Aligned Large Language Models

Jiawei Lian, Jianhong Pan, Lefan Wang, Yi Wang, Shaohui Mei, Lap-Pui Chau

Main category: cs.CL

TL;DR: Semantic Representation Attack is a new method that exploits semantic representation space to generate adversarial prompts that bypass LLM safety alignment, achieving high success rates while maintaining naturalness and efficiency.

DetailsMotivation: Current adversarial attack methods on aligned LLMs suffer from limited convergence, unnatural prompts, and high computational costs by targeting exact textual patterns rather than semantic meaning.

Method: The approach uses Semantic Representation Heuristic Search to generate adversarial prompts in semantic space, maintaining interpretability during incremental expansion and targeting diverse responses with equivalent harmful meanings.

Result: Achieves unprecedented attack success rates (89.41% averaged across 18 LLMs, including 100% on 11 models) while maintaining stealthiness and efficiency.

Conclusion: Semantic Representation Attack fundamentally resolves the trade-off between attack efficacy and prompt naturalness, demonstrating overall superiority over existing methods.

Abstract: Large Language Models (LLMs) increasingly employ alignment techniques to prevent harmful outputs. Despite these safeguards, attackers can circumvent them by crafting prompts that induce LLMs to generate harmful content. Current methods typically target exact affirmative responses, such as ``Sure, here is…’’, suffering from limited convergence, unnatural prompts, and high computational costs. We introduce Semantic Representation Attack, a novel paradigm that fundamentally reconceptualizes adversarial objectives against aligned LLMs. Rather than targeting exact textual patterns, our approach exploits the semantic representation space comprising diverse responses with equivalent harmful meanings. This innovation resolves the inherent trade-off between attack efficacy and prompt naturalness that plagues existing methods. The Semantic Representation Heuristic Search algorithm is proposed to efficiently generate semantically coherent and concise adversarial prompts by maintaining interpretability during incremental expansion. We establish rigorous theoretical guarantees for semantic convergence and demonstrate that our method achieves unprecedented attack success rates (89.41% averaged across 18 LLMs, including 100% on 11 models) while maintaining stealthiness and efficiency. Comprehensive experimental results confirm the overall superiority of our Semantic Representation Attack. The code will be publicly available.

[171] Concise and Sufficient Sub-Sentence Citations for Retrieval-Augmented Generation

Guo Chen, Qiuyuan Li, Qiuxian Li, Hongliang Dai, Xiang Chen, Piji Li

Main category: cs.CL

TL;DR: The paper proposes generating sub-sentence citations in RAG systems to improve verifiability and reduce user effort in confirming LLM outputs.

DetailsMotivation: Existing citation methods in RAG systems have two problems: sentence/paragraph-level citations include irrelevant content, and they may omit essential verification information, forcing users to read surrounding context.

Method: Developed annotation guidelines for sub-sentence citations, constructed a dataset, and proposed an attribution framework using LLMs to generate fine-tuning data with a credit model to filter low-quality examples.

Result: Experiments on the constructed dataset demonstrate that the proposed approach can generate high-quality and more readable citations.

Conclusion: Sub-sentence citations provide concise and sufficient attribution, reducing user effort in verifying LLM outputs while maintaining verifiability.

Abstract: In retrieval-augmented generation (RAG) question answering systems, generating citations for large language model (LLM) outputs enhances verifiability and helps users identify potential hallucinations. However, we observe two problems in the citations produced by existing attribution methods. First, the citations are typically provided at the sentence or even paragraph level. Long sentences or paragraphs may include a substantial amount of irrelevant content. Second, sentence-level citations may omit information that is essential for verifying the output, forcing users to read the surrounding context. In this paper, we propose generating sub-sentence citations that are both concise and sufficient, thereby reducing the effort required by users to confirm the correctness of the generated output. To this end, we first develop annotation guidelines for such citations and construct a corresponding dataset. Then, we propose an attribution framework for generating citations that adhere to our standards. This framework leverages LLMs to automatically generate fine-tuning data for our task and employs a credit model to filter out low-quality examples. Our experiments on the constructed dataset demonstrate that the propose approach can generate high-quality and more readable citations.

[172] A Comparison of Independent and Joint Fine-tuning Strategies for Retrieval-Augmented Generation

Neal Gregory Lawton, Alfy Samuel, Anoop Kumar, Daben Liu

Main category: cs.CL

TL;DR: Comparison of fine-tuning strategies for Retrieval-Augmented Generation (RAG) systems, showing that independent, joint, and two-phase fine-tuning achieve similar performance improvements but have different computational costs.

DetailsMotivation: RAG systems use two LLMs (embedding and generator) that can be fine-tuned to improve performance on new tasks, but different fine-tuning strategies have varying costs and benefits that need evaluation.

Method: Evaluated and compared several RAG fine-tuning strategies including independent, joint, and two-phase fine-tuning approaches.

Result: All fine-tuning strategies achieved approximately equal improvement in EM and F1 generation quality metrics, despite having significantly different computational costs.

Conclusion: The optimal fine-tuning strategy depends on whether the training dataset includes context labels and whether grid search over learning rates for both embedding and generator models is required.

Abstract: A Comparison of Independent and Joint Fine-tuning Strategies for Retrieval-Augmented Generation Download PDF Neal Gregory Lawton, Alfy Samuel, Anoop Kumar, Daben Liu Published: 20 Aug 2025, Retrieval augmented generation (RAG) is a popular framework for question answering that is powered by two large language models (LLMs): an embedding model that retrieves context documents from a database that are relevant to a given question, and a generator model that uses the retrieved context to generate an answer to the question. Both the embedding and generator models can be fine-tuned to increase performance of a RAG pipeline on a new task, but multiple fine-tuning strategies exist with different costs and benefits. In this paper, we evaluate and compare several RAG fine-tuning strategies, including independent, joint, and two-phase fine-tuning. In our experiments, we observe that all of these strategies achieve about equal improvement in EM and F1 generation quality metrics, although they have significantly different computational costs. We conclude the optimal fine-tuning strategy to use depends on whether the training dataset includes context labels and whether a grid search over the learning rates for the embedding and generator models is required.

[173] mini-vec2vec: Scaling Universal Geometry Alignment with Linear Transformations

Guy Dar

Main category: cs.CL

TL;DR: mini-vec2vec is an efficient and robust linear alternative to vec2vec for aligning text embedding spaces without parallel data, offering orders of magnitude improvement in computational efficiency while maintaining or improving performance.

DetailsMotivation: The original vec2vec method for aligning text embedding spaces without parallel data was expensive and unstable, motivating the development of a more efficient and robust alternative.

Method: The method uses three main stages: tentative matching of pseudo-parallel embedding vectors, transformation fitting, and iterative refinement to learn a linear transformation between embedding spaces.

Result: mini-vec2vec exceeds vec2vec by orders of magnitude in efficiency while matching or exceeding its results, with improved stability and interpretability.

Conclusion: The method’s stability, efficiency, and interpretable steps enable scaling and adoption in new domains and fields for embedding space alignment.

Abstract: We build upon vec2vec, a procedure designed to align text embedding spaces without parallel data. vec2vec finds a near-perfect alignment, but it is expensive and unstable. We present mini-vec2vec, a simple and efficient alternative that requires substantially lower computational cost and is highly robust. Moreover, the learned mapping is a linear transformation. Our method consists of three main stages: a tentative matching of pseudo-parallel embedding vectors, transformation fitting, and iterative refinement. Our linear alternative exceeds the original instantiation of vec2vec by orders of magnitude in efficiency, while matching or exceeding their results. The method’s stability and interpretable algorithmic steps facilitate scaling and unlock new opportunities for adoption in new domains and fields.

[174] Customer-R1: Personalized Simulation of Human Behaviors via RL-based LLM Agent in Online Shopping

Ziyi Wang, Yuxuan Lu, Yimeng Zhang, Jing Huang, Dakuo Wang

Main category: cs.CL

TL;DR: Customer-R1 is an RL-based method that enables personalized user behavior simulation in online shopping by conditioning on explicit personas and optimizing next-step rationale and action generation.

DetailsMotivation: Prior methods for simulating step-wise human behavior with LLMs learn population-level policies without considering individual personas, resulting in generic rather than personalized simulations.

Method: Uses reinforcement learning with policy conditioned on explicit persona, optimizing next-step rationale and action generation via action correctness reward signals.

Result: Significantly outperforms prompting and SFT-based baselines in next-action prediction tasks and better matches users’ action distribution, indicating higher fidelity in personalized behavior simulation.

Conclusion: Customer-R1 demonstrates superior capability for personalized step-wise user behavior simulation compared to existing methods.

Abstract: Simulating step-wise human behavior with Large Language Models (LLMs) has become an emerging research direction, enabling applications in various practical domains. While prior methods, including prompting, supervised fine-tuning (SFT), and reinforcement learning (RL), have shown promise in modeling step-wise behavior, they primarily learn a population-level policy without conditioning on a user’s persona, yielding generic rather than personalized simulations. In this work, we pose a critical question: how can LLM agents better simulate personalized user behavior? We introduce Customer-R1, an RL-based method for personalized, step-wise user behavior simulation in online shopping environments. Our policy is conditioned on an explicit persona, and we optimize next-step rationale and action generation via action correctness reward signals. Experiments on the OPeRA dataset emonstrate that Customer-R1 not only significantly outperforms prompting and SFT-based baselines in next-action prediction tasks, but also better matches users’ action distribution, indicating higher fidelity in personalized behavior simulation.

[175] Autoencoding-Free Context Compression for LLMs via Contextual Semantic Anchors

Xin Liu, Runsong Zhao, Pengcheng Huang, Xinyu Liu, Junyi Xiao, Chunyang Xiao, Tong Xiao, Shengxiang Gao, Zhengtao Yu, Jingbo Zhu

Main category: cs.CL

TL;DR: SAC is a novel context compression method that replaces autoencoding-based training with direct selection of anchor tokens and contextual information aggregation into their KV representations, eliminating the need for compression training.

DetailsMotivation: Current context compression methods using autoencoding tasks create a mismatch between reconstruction optimization and actual downstream tasks, weakening features beneficial for real-world usage.

Method: SAC directly selects anchor tokens from original context and aggregates contextual information into their KV representations using anchor embeddings and bidirectional attention modification.

Result: SAC consistently outperforms existing context compression methods across various compression ratios, achieving 1 EM improvement at 5x compression on MRQA with increasing advantages at higher ratios.

Conclusion: SAC provides a more effective approach to context compression by eliminating autoencoding training and directly leveraging contextual tokens through anchor-based architecture.

Abstract: Context compression presents a promising approach for accelerating large language model (LLM) inference by compressing long contexts into compact representations. Current context compression methods predominantly rely on autoencoding tasks to train context-agnostic compression tokens to compress contextual semantics. While autoencoding tasks enable compression tokens to acquire compression capabilities, compression via autoencoding tasks creates a fundamental mismatch: the models are optimized for reconstruction that diverge from actual downstream tasks, thereby weakening the features more beneficial for real-world usage. We propose Semantic-Anchor Compression (SAC), a novel method that shifts from autoencoding task based compression to an architecture that is equipped with this compression capability \textit{a priori}. Instead of training models to compress contexts through autoencoding tasks, SAC directly selects so-called anchor tokens from the original context and aggregates contextual information into their key-value (KV) representations. By deriving representations directly from the contextual tokens, SAC eliminates the need for autoencoding training. To ensure compression performance while directly leveraging anchor tokens, SAC incorporates two key designs: (1) anchor embeddings that enable the compressor to identify critical tokens, and (2) bidirectional attention modification that allows anchor tokens to capture information from the entire context. Experimental results demonstrate that SAC consistently outperforms existing context compression methods across various compression ratios. On out-of-distribution evaluation using MRQA, SAC achieves 1 EM improvement at 5x compression over strong baselines, with increasing advantages at higher compression ratios.

[176] DICE: Structured Reasoning in LLMs through SLM-Guided Chain-of-Thought Correction

Yiqi Li, Yusheng Liao, Zhe Chen, Yanfeng Wang, Yu Wang

Main category: cs.CL

TL;DR: DICE is a lightweight framework that uses small language models (SLMs) to refine LLM outputs through chain-of-thought correction, improving format accuracy and content correctness without expensive LLM fine-tuning.

DetailsMotivation: LLMs often prioritize reasoning over following detailed output format instructions, and fine-tuning LLMs is computationally expensive with limited parameter access.

Method: DICE decouples the process: LLMs generate natural language responses, then trained SLMs analyze and refine these outputs using chain-of-thought correction. It constructs structured CoT datasets via two-stage method and applies dual-tuning strategy to fine-tune SLMs.

Result: DICE improves average format accuracy by 35.4% and content correctness by 29.4%, achieving state-of-the-art performance over competitive baselines.

Conclusion: DICE effectively preserves LLMs’ knowledge and reasoning while ensuring outputs meet structured specifications, providing a practical alternative to expensive LLM fine-tuning.

Abstract: When performing reasoning tasks with user-specific requirements, such as strict output formats, large language models (LLMs) often prioritize reasoning over adherence to detailed instructions. Fine-tuning LLMs on supervised datasets to address this is impractical due to high computational costs and limited parameter access. To tackle this, we propose DICE, a lightweight framework that guides small language models (SLMs) to refine LLMs’ outputs through chain-of-thought (CoT) correction. DICE decouples the process by first prompting LLMs to generate natural language responses, then using trained SLMs to analyze and refine these outputs to meet structured output specifications. This framework preserves LLMs’ broad knowledge and reasoning capabilities while ensuring the outputs conform to user demands. Specifically, DICE first constructs structured CoT adaptation datasets via a two-stage method and subsequently applies a dual-tuning strategy to fine-tune SLMs for generating structured outputs in an analyze-then-answer pattern. Experiments demonstrate that DICE improves the average format accuracy and content correctness of LLM outputs by 35.4% and 29.4%, respectively, achieving state-of-the-art (SOTA) performance over other competitive baselines.

[177] ShiZhi: A Chinese Lightweight Large Language Model for Court View Generation

Zhitian Hou, Kun Zeng

Main category: cs.CL

TL;DR: ShiZhi is the first LLM for criminal court view generation, achieving 70.00 ROUGE-1 and 67.85 BLEU-1 scores on generating court views from case facts using a 110K Chinese dataset.

DetailsMotivation: Court View Generation is challenging due to case diversity and complexity, and direct generation from raw facts limits performance, requiring specialized legal AI models.

Method: Developed ShiZhi LLM specifically for court view generation using a Chinese Court View Generation dataset (CCVG) of 110K+ cases with fact descriptions paired with court views.

Result: Achieved 70.00 ROUGE-1 and 67.85 BLEU-1 on court view generation, and 86.48% accuracy with 92.75% macro F1 on charge prediction.

Conclusion: Even small LLMs can generate reasonable and legally coherent court views when trained on high-quality domain-specific legal data.

Abstract: Criminal Court View Generation (CVG) is a fundamental task in legal artificial intelligence, aiming to automatically generate the “Court View” section of a legal case document. Generating court views is challenging due to the diversity and complexity of case facts, and directly generating from raw facts may limit performance. In this paper, we present ShiZhi, the first large language model (LLM) specifically designed for court view generation. We construct a Chinese Court View Generation dataset, CCVG, of more than 110K cases, each containing fact descriptions paired with corresponding court views. Based on this dataset, ShiZhi achieving 70.00 ROUGE-1 and 67.85 BLEU-1 on court view generation, as well as 86.48% accuracy with 92.75% macro F1 on charge prediction. Experimental results demonstrate that even a small LLM can generate reasonable and legally coherent court views when trained on high-quality domain-specific data. Our model and dataset are available at \href{https://github.com/ZhitianHou/ShiZhi}{https://github.com/ZhitianHou/ShiZhi}.

[178] HUME: Measuring the Human-Model Performance Gap in Text Embedding Tasks

Adnan El Assadi, Isaac Chung, Roman Solomatin, Niklas Muennighoff, Kenneth Enevoldsen

Main category: cs.CL

TL;DR: HUME is a human evaluation framework for text embeddings that measures human performance across 16 MTEB datasets, revealing that humans achieve 77.6% average performance compared to 80.1% for the best embedding model, with substantial variation across tasks and languages.

DetailsMotivation: To enable meaningful comparison between human and model performance on embedding tasks, as current frameworks like MTEB lack reliable human performance estimates, limiting interpretability of model scores.

Method: Developed HUME framework to measure human performance across 16 MTEB datasets spanning reranking, classification, clustering, and semantic textual similarity across high- and low-resource languages.

Result: Humans achieved 77.6% average performance vs 80.1% for best embedding model, with models reaching near-ceiling performance on some datasets but struggling on others, particularly in low-resource languages.

Conclusion: HUME provides human performance baselines, insights into task difficulty patterns, and an extensible framework for more meaningful model evaluation and benchmark development.

Abstract: Comparing human and model performance offers a valuable perspective for understanding the strengths and limitations of embedding models, highlighting where they succeed and where they fail to capture meaning and nuance. However, such comparisons are rarely made, as human performance on embedding tasks is difficult to measure. To fill this gap, we introduce HUME: Human Evaluation Framework for Text Embeddings. While frameworks like MTEB provide broad model evaluation, they lack reliable estimates of human performance, limiting the interpretability of model scores. We measure human performance across 16 MTEB datasets spanning reranking, classification, clustering, and semantic textual similarity across linguistically diverse high- and low-resource languages. Humans achieve an average performance of 77.6% compared to 80.1% for the best embedding model, although variation is substantial: models reach near-ceiling performance on some datasets while struggling on others, suggesting dataset issues and revealing shortcomings in low-resource languages. We provide human performance baselines, insight into task difficulty patterns, and an extensible evaluation framework that enables a more meaningful interpretation of the model and informs the development of both models and benchmarks. Our code, dataset, and leaderboard are publicly available at https://github.com/embeddings-benchmark/mteb.

[179] Audit-of-Understanding: Posterior-Constrained Inference for Mathematical Reasoning in Language Models

Samir Abdaljalil, Erchin Serpedin, Khalid Qaraqe, Hasan Kurban

Main category: cs.CL

TL;DR: AoU is a framework that prevents reasoning hallucinations in LLMs by decomposing queries into assumptions, validating them, and only using supported premises for inference.

DetailsMotivation: LLMs often generate reasoning traces that appear coherent but rely on unsupported assumptions, leading to hallucinated conclusions. Prior work mainly addresses factual hallucinations or uses post-hoc verification, leaving reasoning-induced hallucinations unaddressed.

Method: Audit-of-Understanding (AoU) framework with three phases: (1) decomposing queries into candidate assumptions, (2) auditing their support, and (3) conditioning inference only on validated premises. Formally, it’s posterior-constrained inference connected to selective prediction and rejection learning.

Result: AoU improves both accuracy and faithfulness on GSM8K, MultiArith, and SVAMP benchmarks, achieving up to +30% gains on GSM8K, +45% on MultiArith, and consistent +20-28% improvements on SVAMP over Chain-of-Thought, Self-Consistency, and CoT-Decoding.

Conclusion: AoU effectively addresses reasoning-induced hallucinations in LLMs by constraining inference to validated premises, providing theoretical guarantees and empirical improvements across multiple reasoning benchmarks.

Abstract: Large language models (LLMs) often generate reasoning traces that appear coherent but rest on unsupported assumptions, leading to hallucinated conclusions. Prior work mainly addresses factual hallucinations or relies on post-hoc verification, leaving reasoning-induced hallucinations largely unaddressed. We propose Audit-of-Understanding (AoU), a framework that constrains inference to validated premises through three phases: (1) decomposing a query into candidate assumptions, (2) auditing their support, and (3) conditioning inference only on the validated subset. Formally, AoU is \emph{posterior-constrained inference}, connecting to selective prediction and rejection learning. Our contributions are threefold: (i) theoretical guarantees under perfect validation, (ii) excess-risk bounds under imperfect audits, and (iii) tractability analysis. Empirically, AoU improves both accuracy and faithfulness on GSM8K, MultiArith, and SVAMP, achieving up to +30% gains on GSM8K, +45% on MultiArith, and consistent +20–28% improvements on SVAMP over Chain-of-Thought, Self-Consistency, and CoT-Decoding. Code is available at https://anonymous.4open.science/r/audit-of-understanding-E28B.

[180] The Curious Case of Factual (Mis)Alignment between LLMs’ Short- and Long-Form Answers

Saad Obaid ul Islam, Anne Lauscher, Goran Glavaš

Main category: cs.CL

TL;DR: SLAQ framework reveals systematic inconsistencies in LLMs’ factual knowledge across query complexities - models answer simple questions correctly but fail on the same facts in complex queries, challenging current evaluation practices.

DetailsMotivation: LLMs show impressive accuracy on simple factual QA benchmarks but exhibit reliability gaps between simple and complex queries, eroding trustworthiness. The fundamental inconsistency in how models access factual knowledge across task complexities remains poorly understood.

Method: Introduced SLAQ framework comparing LLMs’ answers to same factual questions asked in isolation (short) vs. integrated into complex queries (long). Evaluated 16 LLMs across 600 queries, conducted mechanistic analysis to examine model internals.

Result: Found systematic misalignment between short and long query answers, position-dependent accuracy loss, and momentum effects. Aligned facts activate overlapping model internals, and mechanistic similarity metrics can predict short-long answer alignment with up to 78% accuracy.

Conclusion: Factual consistency over query complexity is crucial for LLMs’ trustworthiness. Current evaluation practices are flawed as they assume good performance on simple queries implies reliability in complex knowledge-seeking tasks.

Abstract: Large language models (LLMs) can correctly answer “When was Einstein born?” yet fail to provide the same date when writing about Einstein’s life revealing a fundamental inconsistency in how models access factual knowledge across task complexities. While models display impressive accuracy on factual question-answering benchmarks, the reliability gap between simple and complex queries remains poorly understood, eroding their trustworthiness. In this work, we introduce Short-Long Form Alignment for Factual Question Answering (SLAQ), a controlled evaluation framework that compares LLMs’ answers to the same factual questions asked (a) in isolation (short) vs. (b) integrated into complex queries (long). Looking at 16 LLMs across 600 queries, we find a systematic misalignment of answers to the corresponding short and long queries. We further uncover position-dependent accuracy loss and momentum effects where consecutive correct or incorrect answers create self-reinforcing patterns. Through mechanistic analysis, we find that aligned facts activate overlapping model internals, and that metrics based on mechanistic similarity can predict short-long answer alignment with up to 78% accuracy. Our work establishes factual consistency over query complexity as an important aspect of LLMs’ trustworthiness and challenges current evaluation practices, which implicitly assume that good performance for simple factual queries implies reliability in more complex knowledge-seeking tasks too.

[181] Valid Survey Simulations with Limited Human Data: The Roles of Prompting, Fine-Tuning, and Rectification

Stefan Krsteski, Giuseppe Russo, Serina Chang, Robert West, Kristina Gligorić

Main category: cs.CL

TL;DR: LLMs can substitute for human survey respondents but introduce bias. Combining synthesis (LLM-generated responses) with rectification (debiasing methods) reduces bias below 5% and increases effective sample size by up to 14%. Optimal allocation under fixed budget favors most human responses for rectification rather than fine-tuning.

DetailsMotivation: Traditional surveys are costly and slow, while LLMs offer scalable alternatives but produce biased estimates. Need to find optimal combination of LLM synthesis and debiasing methods.

Method: Study synthesis methods (using LLMs to generate survey responses) combined with rectification methods (debiasing population estimates). Tested on panel surveys covering nutrition, politics, and economics. Explored optimal allocation of human responses between synthesis and rectification under fixed budget.

Result: Synthesis alone introduces substantial bias (24-86%). Combining synthesis with rectification reduces bias below 5% and increases effective sample size by up to 14%. Allocating most human responses to rectification rather than fine-tuning provides more effective estimation.

Conclusion: LLMs can effectively substitute for human survey respondents when combined with proper debiasing methods. Optimal strategy allocates majority of human responses to rectification rather than fine-tuning, challenging common practice.

Abstract: Surveys provide valuable insights into public opinion and behavior, but their execution is costly and slow. Large language models (LLMs) have been proposed as a scalable, low-cost substitute for human respondents, but their outputs are often biased and yield invalid estimates. We study the interplay between synthesis methods that use LLMs to generate survey responses and rectification methods that debias population estimates, and explore how human responses are best allocated between them. Using two panel surveys with questions on nutrition, politics, and economics, we find that synthesis alone introduces substantial bias (24-86%), whereas combining it with rectification reduces bias below 5% and increases effective sample size by up to 14%. Overall, we challenge the common practice of using all human responses for fine-tuning, showing that under a fixed budget, allocating most to rectification results in far more effective estimation.

[182] Evaluating Retrieval-Augmented Generation Systems on Unanswerable, Uncheatable, Realistic, Multi-hop Queries

Gabrielle Kaili-May Liu, Bryan Li, Arman Cohan, William Gantt Walden, Eugene Yang

Main category: cs.CL

TL;DR: The paper introduces CRUMQs, a pipeline for creating challenging, uncheatable, realistic, unanswerable, and multi-hop queries to better evaluate RAG systems.

DetailsMotivation: Existing RAG benchmarks fail to reflect realistic task complexity, allowing systems to cheat via disconnected reasoning and lacking proper evaluation of unanswerable queries and multi-hop reasoning failures.

Method: Developed an automatic pipeline for difficulty-controlled creation of CRUMQs (uncheatable, realistic, unanswerable, multi-hop queries) adaptable to any corpus and domain.

Result: CRUMQs significantly challenge RAG systems, achieving up to 81.0% reduction in cheatability scores compared to prior benchmarks when tested on leading retrieval-augmented LLMs.

Conclusion: The CRUMQs pipeline effectively enhances benchmark difficulty and realism, providing a better framework to drive development of more capable RAG systems.

Abstract: Real-world use cases often present RAG systems with complex queries for which relevant information is missing from the corpus or is incomplete. In these settings, RAG systems must be able to reject unanswerable, out-of-scope queries and identify failures of retrieval and multi-hop reasoning. Despite this, existing RAG benchmarks rarely reflect realistic task complexity for multi-hop or out-of-scope questions, which often can be cheated via disconnected reasoning (i.e., solved without genuine multi-hop inference) or require only simple factual recall. This limits the ability for such benchmarks to uncover limitations of existing RAG systems. To address this gap, we present the first pipeline for automatic, difficulty-controlled creation of un$\underline{c}$heatable, $\underline{r}$ealistic, $\underline{u}$nanswerable, and $\underline{m}$ulti-hop $\underline{q}$uerie$\underline{s}$ (CRUMQs), adaptable to any corpus and domain. We use our pipeline to create CRUMQs over two popular RAG datasets and demonstrate its effectiveness via benchmark experiments on leading retrieval-augmented LLMs. Results show that compared to prior RAG benchmarks, CRUMQs are highly challenging for RAG systems and achieve up to 81.0% reduction in cheatability scores. More broadly, our pipeline offers a simple way to enhance benchmark difficulty and realism and drive development of more capable RAG systems.

[183] The Curious Case of Curiosity across Human Cultures and LLMs

Angana Borah, Zhijing Jin, Rada Mihalcea

Main category: cs.CL

TL;DR: LLMs flatten cross-cultural diversity in curiosity expression, aligning more with Western patterns. Fine-tuning can reduce this human-model alignment gap by up to 50%, demonstrating curiosity’s importance for LLM adaptability across cultures.

DetailsMotivation: Curiosity is a central driver of inquiry but remains underexplored in LLMs across cultural contexts, despite their expanding role in human interaction.

Method: Introduced CUEST framework to evaluate human-model alignment in curiosity using Yahoo! Answers dataset across multiple countries, analyzing linguistic style and topic preferences, grounded in social science constructs.

Result: LLMs across both open- and closed-source models flatten cross-cultural diversity in curiosity, aligning more closely with Western expression patterns. Fine-tuning strategies narrowed the human-model alignment gap by up to 50%.

Conclusion: Curiosity has practical value for LLM adaptability across cultures and is important for future NLP research.

Abstract: Recent advances in Large Language Models (LLMs) have expanded their role in human interaction, yet curiosity – a central driver of inquiry – remains underexplored in these systems, particularly across cultural contexts. In this work, we investigate cultural variation in curiosity using Yahoo! Answers, a real-world multi-country dataset spanning diverse topics. We introduce CUEST (CUriosity Evaluation across SocieTies), an evaluation framework that measures human-model alignment in curiosity through linguistic (style), topic preference (content) analysis and grounding insights in social science constructs. Across open- and closed-source models, we find that LLMs flatten cross-cultural diversity, aligning more closely with how curiosity is expressed in Western countries. We then explore fine-tuning strategies to induce curiosity in LLMs, narrowing the human-model alignment gap by up to 50%. Finally, we demonstrate the practical value of curiosity for LLM adaptability across cultures, showing its importance for future NLP research.

[184] ConsintBench: Evaluating Language Models on Real-World Consumer Intent Understanding

Xiaozhe Li, TianYi Lyu, Siyi Yang, Yuxi Gong, Yizhao Yang, Jinxuan Huang, Ligao Zhang, Zhuoyi Huang, Qingwen Liu

Main category: cs.CL

TL;DR: The paper introduces \bench, the first dynamic, live evaluation benchmark for intent understanding in real-world public discussions, specifically in the consumer domain, addressing the lack of large-scale benchmarks for this complex task.

DetailsMotivation: Current LLMs struggle with understanding human intent in real-world public discussions which involve interwoven perspectives, conflicting views, emotional tendencies, and implicit assumptions. No existing benchmark evaluates LLMs on this complex task due to challenges in collecting real-world data and building evaluation pipelines.

Method: The authors developed \bench, a dynamic and live evaluation benchmark that supports real-time updates and prevents data contamination through an automated curation pipeline. It focuses on consumer domain discussions.

Result: \bench is presented as the largest and most diverse benchmark of its kind, specifically designed for intent understanding in real-world public discussions with dynamic, live evaluation capabilities.

Conclusion: The paper bridges a critical gap in LLM evaluation by providing the first specialized benchmark for understanding human intent in complex, multi-participant public discussions, enabling better assessment of LLMs’ analytical reasoning and contextual interpretation abilities.

Abstract: Understanding human intent is a complex, high-level task for large language models (LLMs), requiring analytical reasoning, contextual interpretation, dynamic information aggregation, and decision-making under uncertainty. Real-world public discussions, such as consumer product discussions, are rarely linear or involve a single user. Instead, they are characterized by interwoven and often conflicting perspectives, divergent concerns, goals, emotional tendencies, as well as implicit assumptions and background knowledge about usage scenarios. To accurately understand such explicit public intent, an LLM must go beyond parsing individual sentences; it must integrate multi-source signals, reason over inconsistencies, and adapt to evolving discourse, similar to how experts in fields like politics, economics, or finance approach complex, uncertain environments. Despite the importance of this capability, no large-scale benchmark currently exists for evaluating LLMs on real-world human intent understanding, primarily due to the challenges of collecting real-world public discussion data and constructing a robust evaluation pipeline. To bridge this gap, we introduce \bench, the first dynamic, live evaluation benchmark specifically designed for intent understanding, particularly in the consumer domain. \bench is the largest and most diverse benchmark of its kind, supporting real-time updates while preventing data contamination through an automated curation pipeline.

[185] Deflanderization for Game Dialogue: Balancing Character Authenticity with Task Execution in LLM-based NPCs

Pasin Buakhaw, Kun Kerdthaisong, Phuree Phenhiran, Pitikorn Khlaisamniang, Supasate Vorathammathorn, Piyalitt Ittichaiwong, Nutchanon Yongsatianchot

Main category: cs.CL

TL;DR: The paper presents an approach for creating dynamic NPCs using LLMs, combining lightweight prompting techniques and fine-tuned models to achieve top rankings in the CPDC 2025 competition.

DetailsMotivation: To leverage large language models for creating dynamic non-player characters in gaming environments that can perform functional tasks and generate persona-consistent dialogues.

Method: Combines two strategies: (1) lightweight prompting techniques including Deflanderization to reduce excessive role-play and improve task fidelity in API track, and (2) fine-tuned Qwen3-14B models using supervised fine-tuning and LoRA in GPU track.

Result: Achieved 2nd place on Task 1, 2nd place on Task 3 (API track), and 4th place on Task 3 (GPU track) in the Commonsense Persona-Grounded Dialogue Challenge 2025 Round 2.

Conclusion: The combination of prompting techniques and fine-tuned models effectively enables both task execution and persona-consistent dialogue generation for dynamic NPCs in gaming environments.

Abstract: The emergence of large language models (LLMs) has opened new opportunities for cre- ating dynamic non-player characters (NPCs) in gaming environments, enabling both func- tional task execution and persona-consistent dialogue generation. In this paper, we (Tu_Character_lab) report our participation in the Commonsense Persona-Grounded Dialogue Challenge (CPDC) 2025 Round 2, which eval- uates agents across three tracks: task-oriented dialogue, context-aware dialogue, and their integration. Our approach combines two complementary strategies: (i) lightweight prompting techniques in the API track, including a Deflanderization prompting method to suppress excessive role-play and improve task fidelity, and (ii) fine-tuned large models in the GPU track, leveraging Qwen3-14B with supervisedfinetuning (SFT) and Low-Rank Adaptation(LoRA). Our best submissions ranked 2nd on Task 1, 2nd on Task 3 (API track), and 4th on Task 3 (GPU track).

[186] Beyond One World: Benchmarking Super Heros in Role-Playing Across Multiversal Contexts

Perapard Ngokpol, Kun Kerdthaisong, Pasin Buakhaw, Pitikorn Khlaisamniang, Supasate Vorathammathorn, Piyalitt Ittichaiwong, Nutchanon Yongsatianchot

Main category: cs.CL

TL;DR: The paper introduces Beyond One World, a benchmark for evaluating LLMs’ ability to faithfully portray version-specific characters across multiple universes, focusing on canonical accuracy and reasoning fidelity.

DetailsMotivation: To address the underexplored capacity of LLMs to consistently portray version-specific characters across different storytelling universes, using superhero canons as a rich testbed.

Method: Created a benchmark with 30 iconic heroes and 90 canon-specific versions, featuring two tasks: Canon Events (factual recall) and Moral Dilemmas (ethical scenarios). Proposed Think-Act Matching metric to quantify alignment between reasoning and actions.

Result: Chain-of-thought improves coherence in weaker models but reduces accuracy in stronger ones; cross-version generalization remains challenging; models rarely excel at both thinking and acting simultaneously.

Conclusion: Beyond One World exposes critical gaps in multiversal consistency and reasoning alignment, providing a challenging evaluation framework for role-playing LLMs.

Abstract: Large language models (LLMs) are increasingly used as role-playing agents, yet their capacity to faithfully and consistently portray version-specific characters – for example, superheroes across comic and cinematic universes – remains underexplored. Superhero canons such as Marvel and DC provide a rich testbed: decades of storytelling yield multiple incarnations of the same character with distinct histories, values, and moral codes. To study this problem, we introduce Beyond One World, a benchmark for character-grounded roleplay spanning 30 iconic heroes and 90 canon-specific versions. The benchmark comprises two tasks: (i) Canon Events, which probes factual recall of pivotal life stages, and (ii) Moral Dilemmas, which confronts models with ethically charged scenarios. We score responses for canonical accuracy and reasoning fidelity under a framework that separates internal deliberation (“thinking”) from outward decisions (“acting”). We further propose Think-Act Matching, a metric that quantifies alignment between reasons and actions and serves as a proxy for model trustworthiness. Experiments across reasoning- and non-reasoning-oriented models yield three findings: (1) chain-of-thought prompting improves narrative coherence in weaker models but can reduce canonical accuracy in stronger ones; (2) cross-version generalization within a character remains a major obstacle; and (3) models often excel at either thinking or acting, but rarely both. Beyond One World exposes critical gaps in multiversal consistency and reasoning alignment, offering a challenging evaluation for role-playing LLMs.

[187] Your Next Token Prediction: A Multilingual Benchmark for Personalized Response Generation

Shiyao Ding, Takayuki Ito

Main category: cs.CL

TL;DR: The paper introduces ‘Your Next Token Prediction (YNTP)’ task to model individual communication styles through controlled human-agent conversations, addressing privacy concerns in collecting real SNS/email data.

DetailsMotivation: LLMs struggle to generate responses that reflect how individuals truly communicate in daily interactions like emails or social messages, and real communication data is difficult to collect due to privacy concerns.

Method: Built a multilingual benchmark of 100 dialogue sessions across English, Japanese, and Chinese, where users interact for five days with psychologically grounded NPCs based on MBTI dimensions to capture natural communication patterns.

Result: Established the first benchmark for YNTP and evaluated prompt-based and fine-tuning-based personalization methods, providing a foundation for user-aligned language modeling.

Conclusion: The YNTP task and dataset enable analysis of users’ internal communication models and advance personalized language generation while addressing privacy limitations of real communication data collection.

Abstract: Large language models (LLMs) excel at general next-token prediction but still struggle to generate responses that reflect how individuals truly communicate, such as replying to emails or social messages in their own style. However, real SNS or email histories are difficult to collect due to privacy concerns. To address this, we propose the task of “Your Next Token Prediction (YNTP)”, which models a user’s precise word choices through controlled human-agent conversations. We build a multilingual benchmark of 100 dialogue sessions across English, Japanese, and Chinese, where users interact for five days with psychologically grounded NPCs based on MBTI dimensions. This setup captures natural, daily-life communication patterns and enables analysis of users’ internal models. We evaluate prompt-based and fine-tuning-based personalization methods, establishing the first benchmark for YNTP and a foundation for user-aligned language modeling. The dataset is available at: https://github.com/AnonymousHub4Submissions/your-next-token-prediction-dataset-100

[188] MedTrust-RAG: Evidence Verification and Trust Alignment for Biomedical Question Answering

Yingpeng Ning, Yuanyuan Sun, Ling Luo, Yanhua Wang, Yuchen Pan, Hongfei Lin

Main category: cs.CL

TL;DR: MedTrust-Guided Iterative RAG framework enhances biomedical QA by reducing hallucinations through citation-aware reasoning, iterative retrieval-verification, and preference optimization.

DetailsMotivation: RAG systems in biomedical QA suffer from hallucinations due to post-retrieval noise and insufficient evidence verification, undermining response reliability.

Method: Three innovations: citation-aware reasoning with Negative Knowledge Assertions, iterative retrieval-verification with Medical Gap Analysis, and MedTrust-Align Module using Direct Preference Optimization.

Result: Enhanced factual consistency and reduced hallucinations in medical question answering.

Conclusion: The proposed framework significantly improves reliability of biomedical QA systems by systematically addressing hallucination issues through structured verification and preference optimization.

Abstract: Biomedical question answering (QA) requires accurate interpretation of complex medical knowledge. Large language models (LLMs) have shown promising capabilities in this domain, with retrieval-augmented generation (RAG) systems enhancing performance by incorporating external medical literature. However, RAG-based approaches in biomedical QA suffer from hallucinations due to post-retrieval noise and insufficient verification of retrieved evidence, undermining response reliability. We propose MedTrust-Guided Iterative RAG, a framework designed to enhance factual consistency and mitigate hallucinations in medical QA. Our method introduces three key innovations. First, it enforces citation-aware reasoning by requiring all generated content to be explicitly grounded in retrieved medical documents, with structured Negative Knowledge Assertions used when evidence is insufficient. Second, it employs an iterative retrieval-verification process, where a verification agent assesses evidence adequacy and refines queries through Medical Gap Analysis until reliable information is obtained. Third, it integrates the MedTrust-Align Module (MTAM) that combines verified positive examples with hallucination-aware negative samples, leveraging Direct Preference Optimization to reinforce citation-grounded reasoning while penalizing hallucination-prone response patterns.

[189] Planner and Executor: Collaboration between Discrete Diffusion And Autoregressive Models in Reasoning

Lina Berrayana, Ahmed Heakl, Muhammad Abdullah Sohail, Thomas Hofmann, Salman Khan, Wei Chen

Main category: cs.CL

TL;DR: Hybrid architectures combining discrete diffusion language models (DDLMs) with autoregressive models (ARMs) achieve better performance and computational efficiency by shifting communication from text space to latent space.

DetailsMotivation: Current ARMs are accurate but computationally expensive due to long token sequences, while DDLMs offer parallel generation and strong reasoning capabilities. The study explores whether combining both models can yield complementary benefits.

Method: Examined hybrid architectures with DDLM-ARM collaboration in both text space (one model plans, another executes) and latent space (using learned projector to map DDLM latents to ARM embedding space).

Result: Latent-space communication significantly improved accuracy (27.0% to 54.0% on DART-5, 0.0% to 14.0% on AIME24). The hybrid approach provided substantial computational savings - using only 69 tokens total (64 for planning, 5 for execution) outperformed Qwen3.1-7B which used 44x more tokens.

Conclusion: DDLM-ARM hybrid architectures offer new insights into reasoning and highlight DDLMs’ potential in efficient hybrid systems, achieving better performance with significantly reduced computational costs.

Abstract: Current autoregressive language models (ARMs) achieve high accuracy but require long token sequences, making them costly. Discrete diffusion language models (DDLMs) enable parallel and flexible generation within a fixed number of steps and have recently emerged for their strong performance in complex reasoning and long-term planning tasks. We present a study exploring hybrid architectures that couple DDLMs with ARMs to assess whether their collaboration can yield complementary benefits. We first examine collaboration in text space, where one model plans the reasoning process and another executes the final answer based on that plan. We then extend this setup to latent-space communication, introducing a learned projector that maps DDLM latents into the ARM’s embedding space, potentially bypassing some of the text-generation limitations of diffusion models. We find that shifting DDLM –> ARM communication from text space to latent space yields significant accuracy gains, for example increasing from 27.0% to 54.0% on DART-5 and from 0.0% to 14.0% on AIME24. We also find that combining a DDLM planner with an ARM executor can provide substantial computational savings with little to no impact on accuracy. For example, the latent-space pipeline, using 64 tokens for planning and roughly 5 for execution, surpasses Qwen3.1-7B on DART-5 and AIME, despite Qwen using 44 times more tokens. Overall, our study offers new insights into reasoning with DDLMs and highlights their potential in hybrid architectures.

[190] Accelerating Mobile Language Model via Speculative Decoding and NPU-Coordinated Execution

Zhiyang Chen, Daliang Xu, Haiyang Shen, Mengwei Xu, Shangguang Wang, Yun Ma

Main category: cs.CL

TL;DR: sd.npu is a mobile inference framework that accelerates context-aware text generation on mobile devices using speculative decoding with dynamic hardware scheduling, achieving up to 3.8x speedup and 4.7x energy efficiency improvements.

DetailsMotivation: On-device LLMs with local contextual information enable personalized applications, but token-by-token generation suffers from high latency and limited hardware utilization due to memory-bound characteristics.

Method: Three synergistic components: adaptive execution scheduling (dynamic compute graph balancing), context-aligned drafting (lightweight online calibration), and hardware-efficient draft extension (reusing and expanding intermediate sequences).

Result: Experiments show consistent improvements of up to 3.8x in generation speed and 4.7x in energy efficiency compared to existing mobile inference solutions across multiple smartphones and workloads.

Conclusion: The framework successfully accelerates context-aware text generation on mobile devices through speculative decoding and dynamic hardware scheduling, with component-level analysis validating each optimization’s contribution.

Abstract: Enhancing on-device large language models (LLMs) with contextual information from local data enables personalized and task-aware generation, powering use cases such as intelligent assistants and UI agents. While recent developments in neural processors have substantially improved the efficiency of prefill on mobile devices, the token-by-token generation process still suffers from high latency and limited hardware utilization due to its inherently memory-bound characteristics. This work presents sd.npu, a mobile inference framework that integrates speculative decoding with dynamic hardware scheduling to accelerate context-aware text generation on mobile devices. The framework introduces three synergistic components: (1) adaptive execution scheduling, which dynamically balances compute graphs between prefill and decoding phases; (2) context-aligned drafting, which improves speculative efficiency through lightweight online calibration to current tasks; and (3) hardware-efficient draft extension, which reuses and expands intermediate sequences to improve processing parallelism and reduce verification cost. Experiments on multiple smartphones and representative workloads show consistent improvements of up to 3.8x in generation speed and 4.7x in energy efficiency compared with existing mobile inference solutions. Component-level analysis further validates the contribution of each optimization.

[191] AutoGraph-R1: End-to-End Reinforcement Learning for Knowledge Graph Construction

Hong Ting Tsang, Jiaxin Bai, Haoyu Huang, Qiao Xiao, Tianshi Zheng, Baixuan Xu, Shujie Liu, Yangqiu Song

Main category: cs.CL

TL;DR: AutoGraph-R1 is a framework that uses reinforcement learning to optimize knowledge graph construction for RAG-based question answering systems, bridging the gap between KG construction and downstream task performance.

DetailsMotivation: Current knowledge graph construction for RAG systems is decoupled from downstream applications, resulting in suboptimal graph structures that don't maximize task performance.

Method: Uses reinforcement learning to train an LLM constructor, framing graph generation as policy learning with task-aware reward functions based on the graph’s utility in RAG pipelines.

Result: AutoGraph-R1 consistently enables graph RAG methods to achieve significant performance gains over task-agnostic baseline graphs across multiple QA benchmarks.

Conclusion: It’s possible to close the loop between KG construction and application, shifting from building intrinsically ‘good’ graphs to building demonstrably ‘useful’ ones.

Abstract: Building effective knowledge graphs (KGs) for Retrieval-Augmented Generation (RAG) is pivotal for advancing question answering (QA) systems. However, its effectiveness is hindered by a fundamental disconnect: the knowledge graph (KG) construction process is decoupled from its downstream application, yielding suboptimal graph structures. To bridge this gap, we introduce AutoGraph-R1, the first framework to directly optimize KG construction for task performance using Reinforcement Learning (RL). AutoGraph-R1 trains an LLM constructor by framing graph generation as a policy learning problem, where the reward is derived from the graph’s functional utility in a RAG pipeline. We design two novel, task-aware reward functions, one for graphs as knowledge carriers and another as knowledge indices. Across multiple QA benchmarks, AutoGraph-R1 consistently enables graph RAG methods to achieve significant performance gains over using task-agnostic baseline graphs. Our work shows it is possible to close the loop between construction and application, shifting the paradigm from building intrinsically good'' graphs to building demonstrably useful’’ ones.

[192] Infinity Parser: Layout Aware Reinforcement Learning for Scanned Document Parsing

Baode Wang, Biao Wu, Weizhen Li, Meng Fang, Zuming Huang, Jun Huang, Haozhe Wang, Yanjie Liang, Ling Chen, Wei Chu, Yuan Qi

Main category: cs.CL

TL;DR: LayoutRL is a reinforcement learning framework for document parsing that uses composite rewards to improve layout understanding, trained on the Infinity-Doc-400K dataset to create Infinity-Parser model.

DetailsMotivation: Existing supervised fine-tuning methods struggle with generalization across diverse document types and have limited training data for layout-aware parsing tasks.

Method: Reinforcement learning framework with composite rewards (normalized edit distance, paragraph count accuracy, reading order preservation) trained on Infinity-Doc-400K dataset.

Result: Infinity-Parser achieves state-of-the-art performance across multiple benchmarks (OmniDocBench, olmOCR-Bench, PubTabNet, FinTabNet) for various document types, languages, and complexities.

Conclusion: The approach demonstrates robust generalization and substantially outperforms existing document parsing systems and general-purpose vision-language models.

Abstract: Document parsing from scanned images into structured formats remains a significant challenge due to its complexly intertwined elements such as text paragraphs, figures, formulas, and tables. Existing supervised fine-tuning methods often struggle to generalize across diverse document types, leading to poor performance, particularly on out-of-distribution data. This issue is further exacerbated by the limited availability of high-quality training data for layout-aware parsing tasks. To address these challenges, we introduce LayoutRL, a reinforcement learning framework that optimizes layout understanding through composite rewards integrating normalized edit distance, paragraph count accuracy, and reading order preservation. To support this training, we construct the Infinity-Doc-400K dataset, which we use to train Infinity-Parser, a vision-language model demonstrating robust generalization across various domains. Extensive evaluations on benchmarks including OmniDocBench, olmOCR-Bench, PubTabNet, and FinTabNet show that Infinity-Parser consistently achieves state-of-the-art performance across a broad range of document types, languages, and structural complexities, substantially outperforming both specialized document parsing systems and general-purpose vision-language models. We will release our code, dataset, and model to facilitate reproducible research in document parsing.

[193] Large-scale User Game Lifecycle Representation Learning

Yanjie Gou, Jiangming Liu, Kouying Xue, Yi Hu

Main category: cs.CL

TL;DR: The paper proposes User Game Lifecycle (UGL) to address game sparsity and imbalance issues in game advertising/recommendation systems, with strategies for extracting user interests and handling popularity bias.

DetailsMotivation: Existing recommendation methods fail for games due to game sparsity (only hundreds of games) and game imbalance (user behaviors dominated by few popular games), making traditional large-scale representation learning unsuitable.

Method: Introduces User Game Lifecycle (UGL) to enrich user behaviors, strategies for extracting short/long-term interests, and Inverse Probability Masking for handling game imbalance in UGL representation learning.

Result: Significant improvements: 1.83% AUC offline increase and 21.67% CVR online increase for game advertising; 0.5% AUC offline increase and 0.82% ARPU online increase for in-game item recommendation.

Conclusion: UGL representations effectively address game sparsity and imbalance challenges, demonstrating substantial performance gains in both game advertising and in-game item recommendation systems.

Abstract: The rapid expansion of video game production necessitates the development of effective advertising and recommendation systems for online game platforms. Recommending and advertising games to users hinges on capturing their interest in games. However, existing representation learning methods crafted for handling billions of items in recommendation systems are unsuitable for game advertising and recommendation. This is primarily due to game sparsity, where the mere hundreds of games fall short for large-scale user representation learning, and game imbalance, where user behaviors are overwhelmingly dominated by a handful of popular games. To address the sparsity issue, we introduce the User Game Lifecycle (UGL), designed to enrich user behaviors in games. Additionally, we propose two innovative strategies aimed at manipulating user behaviors to more effectively extract both short and long-term interests. To tackle the game imbalance challenge, we present an Inverse Probability Masking strategy for UGL representation learning. The offline and online experimental results demonstrate that the UGL representations significantly enhance model by achieving a 1.83% AUC offline increase on average and a 21.67% CVR online increase on average for game advertising and a 0.5% AUC offline increase and a 0.82% ARPU online increase for in-game item recommendation.

cs.CV

[194] ESCA: Contextualizing Embodied Agents via Scene-Graph Generation

Jiani Huang, Amish Sethi, Matthew Kuo, Mayank Keoliya, Neelay Velingker, JungHo Jung, Ser-Nam Lim, Ziyang Li, Mayur Naik

Main category: cs.CV

TL;DR: ESC A is a framework that improves multi-modal large language models (MLLMs) for embodied agents through structured spatial-temporal understanding, using a novel SGClip model for scene graph generation without human-labeled annotations.

DetailsMotivation: Current MLLM training lacks fine-grained alignment between pixel-level visual content and textual semantics, limiting their effectiveness as embodied agents.

Method: Propose ESCA framework with SGClip - a CLIP-based, promptable scene graph generation model trained on 87K+ videos using neurosymbolic learning with model-driven self-supervision from video-caption pairs.

Result: SGClip excels in scene graph generation and action localization benchmarks. ESCA consistently improves both open-source and commercial MLLMs, achieving state-of-the-art performance in embodied environments with significant error reduction.

Conclusion: ESCA enables open-source models to surpass proprietary baselines by providing structured spatial-temporal understanding through scene graph contextualization.

Abstract: Multi-modal large language models (MLLMs) are making rapid progress toward general-purpose embodied agents. However, current training pipelines primarily rely on high-level vision-sound-text pairs and lack fine-grained, structured alignment between pixel-level visual content and textual semantics. To overcome this challenge, we propose ESCA, a new framework for contextualizing embodied agents through structured spatial-temporal understanding. At its core is SGClip, a novel CLIP-based, open-domain, and promptable model for generating scene graphs. SGClip is trained on 87K+ open-domain videos via a neurosymbolic learning pipeline, which harnesses model-driven self-supervision from video-caption pairs and structured reasoning, thereby eliminating the need for human-labeled scene graph annotations. We demonstrate that SGClip supports both prompt-based inference and task-specific fine-tuning, excelling in scene graph generation and action localization benchmarks. ESCA with SGClip consistently improves both open-source and commercial MLLMs, achieving state-of-the-art performance across two embodied environments. Notably, it significantly reduces agent perception errors and enables open-source models to surpass proprietary baselines.

[195] RefAtomNet++: Advancing Referring Atomic Video Action Recognition using Semantic Retrieval based Multi-Trajectory Mamba

Kunyu Peng, Di Wen, Jia Fu, Jiamin Wu, Kailun Yang, Junwei Zheng, Ruiping Liu, Yufan Chen, Yuqian Fu, Danda Pani Paudel, Luc Van Gool, Rainer Stiefelhagen

Main category: cs.CV

TL;DR: RAVAR++ introduces RefAVA++ dataset with 2.9M+ frames and 75.1k+ annotated persons, and proposes RefAtomNet++ framework with multi-hierarchical semantic-aligned cross-attention and multi-trajectory Mamba modeling for improved referring atomic video action recognition.

DetailsMotivation: To address the limitations of existing methods in precisely localizing target persons and predicting fine-grained actions in complex multi-person scenarios through language-guided action understanding.

Method: RefAtomNet++ uses multi-hierarchical semantic-aligned cross-attention mechanism with multi-trajectory Mamba modeling at partial-keyword, scene-attribute, and holistic-sentence levels, with dynamic scanning trajectories for visual spatial tokens.

Result: RefAtomNet++ establishes new state-of-the-art results on the RefAVA++ dataset, outperforming previous baselines including RefAtomNet.

Conclusion: The proposed framework effectively overcomes cross-modal alignment limitations and improves performance in referring atomic video action recognition tasks.

Abstract: Referring Atomic Video Action Recognition (RAVAR) aims to recognize fine-grained, atomic-level actions of a specific person of interest conditioned on natural language descriptions. Distinct from conventional action recognition and detection tasks, RAVAR emphasizes precise language-guided action understanding, which is particularly critical for interactive human action analysis in complex multi-person scenarios. In this work, we extend our previously introduced RefAVA dataset to RefAVA++, which comprises >2.9 million frames and >75.1k annotated persons in total. We benchmark this dataset using baselines from multiple related domains, including atomic action localization, video question answering, and text-video retrieval, as well as our earlier model, RefAtomNet. Although RefAtomNet surpasses other baselines by incorporating agent attention to highlight salient features, its ability to align and retrieve cross-modal information remains limited, leading to suboptimal performance in localizing the target person and predicting fine-grained actions. To overcome the aforementioned limitations, we introduce RefAtomNet++, a novel framework that advances cross-modal token aggregation through a multi-hierarchical semantic-aligned cross-attention mechanism combined with multi-trajectory Mamba modeling at the partial-keyword, scene-attribute, and holistic-sentence levels. In particular, scanning trajectories are constructed by dynamically selecting the nearest visual spatial tokens at each timestep for both partial-keyword and scene-attribute levels. Moreover, we design a multi-hierarchical semantic-aligned cross-attention strategy, enabling more effective aggregation of spatial and temporal tokens across different semantic hierarchies. Experiments show that RefAtomNet++ establishes new state-of-the-art results. The dataset and code are released at https://github.com/KPeng9510/refAVA2.

[196] CrossRay3D: Geometry and Distribution Guidance for Efficient Multimodal 3D Detection

Huiming Yang

Main category: cs.CV

TL;DR: CrossRay3D is a sparse multi-modal 3D detector that improves token representation quality through Ray-Aware Supervision and Class-Balanced Supervision, achieving state-of-the-art performance on nuScenes while being computationally efficient and robust to missing sensor data.

DetailsMotivation: Existing sparse cross-modality detectors overlook token representation quality, resulting in sub-optimal foreground quality and limited performance. The paper identifies that geometric structure preservation and class distribution are key to improving sparse detector performance.

Method: Proposes Sparse Selector (SS) with two core modules: Ray-Aware Supervision (RAS) to preserve geometric information during training, and Class-Balanced Supervision to adaptively reweight class semantics and retain small object tokens. Also introduces Ray Positional Encoding to address LiDAR-image distribution differences.

Result: Achieves state-of-the-art performance on nuScenes benchmark with 72.4 mAP and 74.7 NDS, while running 1.84× faster than other leading methods. Demonstrates strong robustness even with partially or entirely missing LiDAR or camera data.

Conclusion: CrossRay3D successfully addresses token representation quality issues in sparse detectors through geometric structure preservation and balanced class handling, achieving superior performance and efficiency while maintaining robustness to sensor failures.

Abstract: The sparse cross-modality detector offers more advantages than its counterpart, the Bird’s-Eye-View (BEV) detector, particularly in terms of adaptability for downstream tasks and computational cost savings. However, existing sparse detectors overlook the quality of token representation, leaving it with a sub-optimal foreground quality and limited performance. In this paper, we identify that the geometric structure preserved and the class distribution are the key to improving the performance of the sparse detector, and propose a Sparse Selector (SS). The core module of SS is Ray-Aware Supervision (RAS), which preserves rich geometric information during the training stage, and Class-Balanced Supervision, which adaptively reweights the salience of class semantics, ensuring that tokens associated with small objects are retained during token sampling. Thereby, outperforming other sparse multi-modal detectors in the representation of tokens. Additionally, we design Ray Positional Encoding (Ray PE) to address the distribution differences between the LiDAR modality and the image. Finally, we integrate the aforementioned module into an end-to-end sparse multi-modality detector, dubbed CrossRay3D. Experiments show that, on the challenging nuScenes benchmark, CrossRay3D achieves state-of-the-art performance with 72.4 mAP and 74.7 NDS, while running 1.84 faster than other leading methods. Moreover, CrossRay3D demonstrates strong robustness even in scenarios where LiDAR or camera data are partially or entirely missing.

[197] Enrich and Detect: Video Temporal Grounding with Multimodal LLMs

Shraman Pramanick, Effrosyni Mavroudi, Yale Song, Rama Chellappa, Lorenzo Torresani, Triantafyllos Afouras

Main category: cs.CV

TL;DR: ED-VTG is a two-stage method for video temporal grounding that uses multimodal LLMs to enrich text queries with missing details before localizing them in videos, achieving state-of-the-art performance.

DetailsMotivation: To improve video temporal grounding by leveraging multimodal LLMs to handle complex natural language queries and mitigate noise/hallucinations through query enrichment.

Method: Two-stage approach: 1) Transform language queries into enriched sentences with additional details, 2) Use lightweight decoder to ground enriched queries with multiple-instance-learning training that dynamically selects optimal query versions.

Result: State-of-the-art performance across temporal video grounding benchmarks, significantly outperforming previous LLM-based approaches and comparable to specialized models, with strong zero-shot evaluation capabilities.

Conclusion: ED-VTG demonstrates that query enrichment through multimodal LLMs combined with dynamic training strategies effectively improves video temporal grounding, offering superior performance while maintaining generalization in zero-shot scenarios.

Abstract: We introduce ED-VTG, a method for fine-grained video temporal grounding utilizing multi-modal large language models. Our approach harnesses the capabilities of multimodal LLMs to jointly process text and video, in order to effectively localize natural language queries in videos through a two-stage process. Rather than being directly grounded, language queries are initially transformed into enriched sentences that incorporate missing details and cues to aid in grounding. In the second stage, these enriched queries are grounded, using a lightweight decoder, which specializes at predicting accurate boundaries conditioned on contextualized representations of the enriched queries. To mitigate noise and reduce the impact of hallucinations, our model is trained with a multiple-instance-learning objective that dynamically selects the optimal version of the query for each training sample. We demonstrate state-of-the-art results across various benchmarks in temporal video grounding and paragraph grounding settings. Experiments reveal that our method significantly outperforms all previously proposed LLM-based temporal grounding approaches and is either superior or comparable to specialized models, while maintaining a clear advantage against them in zero-shot evaluation scenarios.

[198] InfraGPT Smart Infrastructure: An End-to-End VLM-Based Framework for Detecting and Managing Urban Defects

Ibrahim Sheikh Mohamed, Abdullah Yahya Abdullah Omaisan

Main category: cs.CV

TL;DR: A pipeline using CCTV streams with YOLO detectors and vision language models for multi-defect detection and structured maintenance planning in smart cities.

DetailsMotivation: Manual infrastructure inspection is costly and hazardous, while existing automatic systems lack comprehensive defect coverage and structured outputs for maintenance crews.

Method: Combines YOLO object detectors for multi-defect detection and segmentation with vision language models for scene-aware summarization, generating structured JSON action plans.

Result: Accurately identifies diverse defects and produces coherent, structured summaries with incident descriptions, recommended tools, dimensions, repair plans, and urgent alerts.

Conclusion: The system shows promise for city-wide deployment but faces challenges in scaling, requiring further development for broader implementation.

Abstract: Infrastructure in smart cities is increasingly monitored by networks of closed circuit television (CCTV) cameras. Roads, bridges and tunnels develop cracks, potholes, and fluid leaks that threaten public safety and require timely repair. Manual inspection is costly and hazardous, and existing automatic systems typically address individual defect types or provide unstructured outputs that cannot directly guide maintenance crews. This paper proposes a comprehensive pipeline that leverages street CCTV streams for multi defect detection and segmentation using the YOLO family of object detectors and passes the detections to a vision language model (VLM) for scene aware summarization. The VLM generates a structured action plan in JSON format that includes incident descriptions, recommended tools, dimensions, repair plans, and urgent alerts. We review literature on pothole, crack and leak detection, highlight recent advances in large vision language models such as QwenVL and LLaVA, and describe the design of our early prototype. Experimental evaluation on public datasets and captured CCTV clips demonstrates that the system accurately identifies diverse defects and produces coherent summaries. We conclude by discussing challenges and directions for scaling the system to city wide deployments.

[199] LongInsightBench: A Comprehensive Benchmark for Evaluating Omni-Modal Models on Human-Centric Long-Video Understanding

ZhaoYang Han, Qihan Lin, Hao Liang, Bowen Chen, Zhou Liu, Wentao Zhang

Main category: cs.CV

TL;DR: LongInsightBench is the first benchmark for evaluating models’ ability to understand long videos across visual, audio, and text modalities, focusing on human language, viewpoints, actions, and contextual elements.

DetailsMotivation: There was no existing benchmark to assess models' understanding of long videos with rich multimodal content, particularly for tasks requiring temporal localization and long-range causal inference.

Method: Created a benchmark with 1,000 long-duration, information-dense videos from FineVideo dataset, designed six challenging task scenarios (Intra-Event and Inter-Event Tasks), and implemented a three-step semi-automated quality assurance pipeline.

Result: Omni-modal models still struggle with precise temporal localization and long-range causal inference tasks. Extended experiments revealed information loss and processing bias in multi-modal fusion.

Conclusion: LongInsightBench provides a comprehensive evaluation framework that reveals current limitations of omni-modal models in handling long video understanding, particularly for temporal and causal reasoning tasks.

Abstract: We introduce \textbf{LongInsightBench}, the first benchmark designed to assess models’ ability to understand long videos, with a focus on human language, viewpoints, actions, and other contextual elements, while integrating \textbf{visual, audio, and text} modalities. Our benchmark excels in three key areas: \textbf{a) Long-Duration, Information-Dense Videos:} We carefully select approximately 1,000 videos from open-source datasets FineVideo based on duration limit and the information density of both visual and audio modalities, focusing on content like lectures, interviews, and vlogs, which contain rich language elements. \textbf{b) Diverse and Challenging Task Scenarios:} We have designed six challenging task scenarios, including both Intra-Event and Inter-Event Tasks. \textbf{c) Rigorous and Comprehensive Quality Assurance Pipelines:} We have developed a three-step, semi-automated data quality assurance pipeline to ensure the difficulty and validity of the synthesized questions and answer options. Based on LongInsightBench, we designed a series of experiments. Experimental results shows that Omni-modal models(OLMs) still face challenge in tasks requiring precise temporal localization (T-Loc) and long-range causal inference (CE-Caus). Extended experiments reveal the information loss and processing bias in multi-modal fusion of OLMs. Our dataset and code is available at https://anonymous.4open.science/r/LongInsightBench-910F/.

[200] IAD-GPT: Advancing Visual Knowledge in Multimodal Large Language Model for Industrial Anomaly Detection

Zewen Li, Zitong Yu, Qilang Ye, Weicheng Xie, Wei Zhuo, Linlin Shen

Main category: cs.CV

TL;DR: IAD-GPT is a novel MLLM-based paradigm for Industrial Anomaly Detection that combines text semantics with image-level and pixel-level information, using abnormal prompts and multi-mask fusion to achieve state-of-the-art performance.

DetailsMotivation: Traditional IAD methods lack multi-turn dialogues and detailed descriptions, while large pre-trained models haven't fully utilized their potential for anomaly detection tasks.

Method: Uses Abnormal Prompt Generator for detailed anomaly prompts, Text-Guided Enhancer for visual grounding, and Multi-Mask Fusion module to incorporate mask as expert knowledge for pixel-level anomaly perception.

Result: Achieves state-of-the-art performance on MVTec-AD and VisA datasets for self-supervised and few-shot anomaly detection and segmentation tasks.

Conclusion: IAD-GPT effectively combines MLLMs’ causal capabilities with visual grounding techniques to advance industrial anomaly detection with detailed descriptions and dialogue capabilities.

Abstract: The robust causal capability of Multimodal Large Language Models (MLLMs) hold the potential of detecting defective objects in Industrial Anomaly Detection (IAD). However, most traditional IAD methods lack the ability to provide multi-turn human-machine dialogues and detailed descriptions, such as the color of objects, the shape of an anomaly, or specific types of anomalies. At the same time, methods based on large pre-trained models have not fully stimulated the ability of large models in anomaly detection tasks. In this paper, we explore the combination of rich text semantics with both image-level and pixel-level information from images and propose IAD-GPT, a novel paradigm based on MLLMs for IAD. We employ Abnormal Prompt Generator (APG) to generate detailed anomaly prompts for specific objects. These specific prompts from the large language model (LLM) are used to activate the detection and segmentation functions of the pre-trained visual-language model (i.e., CLIP). To enhance the visual grounding ability of MLLMs, we propose Text-Guided Enhancer, wherein image features interact with normal and abnormal text prompts to dynamically select enhancement pathways, which enables language models to focus on specific aspects of visual data, enhancing their ability to accurately interpret and respond to anomalies within images. Moreover, we design a Multi-Mask Fusion module to incorporate mask as expert knowledge, which enhances the LLM’s perception of pixel-level anomalies. Extensive experiments on MVTec-AD and VisA datasets demonstrate our state-of-the-art performance on self-supervised and few-shot anomaly detection and segmentation tasks, such as MVTec-AD and VisA datasets. The codes are available at \href{https://github.com/LiZeWen1225/IAD-GPT}{https://github.com/LiZeWen1225/IAD-GPT}.

[201] Effect of Reporting Mode and Clinical Experience on Radiologists’ Gaze and Image Analysis Behavior in Chest Radiography

Mahta Khoobi, Marc Sebastian von der Stueck, Felix Barajas Ordonez, Anca-Maria Iancu, Eric Corban, Julia Nowak, Aleksandar Kargaliev, Valeria Perelygina, Anna-Sophie Schott, Daniel Pinto dos Santos, Christiane Kuhl, Daniel Truhn, Sven Nebelung, Robert Siepmann

Main category: cs.CV

TL;DR: AI-assisted structured reporting (AI-SR) significantly improves diagnostic accuracy, efficiency, and user experience compared to free-text and structured reporting alone for chest radiograph analysis.

DetailsMotivation: To evaluate how structured reporting and AI assistance impact radiologists' image analysis behavior, diagnostic accuracy, efficiency, and user experience in clinical practice.

Method: Prospective study with 8 readers (4 novice, 4 non-novice) analyzing 35 bedside chest radiographs each using three reporting modes: free-text (FT), structured reporting (SR), and AI-assisted structured reporting (AI-SR), with eye-tracking and timing measurements.

Result: AI-SR achieved highest diagnostic accuracy (κ=0.71 vs 0.58-0.60), fastest reporting times (25±9s vs 37±18s vs 88±38s), reduced saccade counts and fixation durations, and was the preferred reporting mode.

Conclusion: Structured reporting improves efficiency by guiding visual attention toward images, and AI-prefilled structured reporting further enhances diagnostic accuracy and user satisfaction.

Abstract: Structured reporting (SR) and artificial intelligence (AI) may transform how radiologists interact with imaging studies. This prospective study (July to December 2024) evaluated the impact of three reporting modes: free-text (FT), structured reporting (SR), and AI-assisted structured reporting (AI-SR), on image analysis behavior, diagnostic accuracy, efficiency, and user experience. Four novice and four non-novice readers (radiologists and medical students) each analyzed 35 bedside chest radiographs per session using a customized viewer and an eye-tracking system. Outcomes included diagnostic accuracy (compared with expert consensus using Cohen’s $\kappa$), reporting time per radiograph, eye-tracking metrics, and questionnaire-based user experience. Statistical analysis used generalized linear mixed models with Bonferroni post-hoc tests with a significance level of ($P \le .01$). Diagnostic accuracy was similar in FT ($\kappa = 0.58$) and SR ($\kappa = 0.60$) but higher in AI-SR ($\kappa = 0.71$, $P < .001$). Reporting times decreased from $88 \pm 38$ s (FT) to $37 \pm 18$ s (SR) and $25 \pm 9$ s (AI-SR) ($P < .001$). Saccade counts for the radiograph field ($205 \pm 135$ (FT), $123 \pm 88$ (SR), $97 \pm 58$ (AI-SR)) and total fixation duration for the report field ($11 \pm 5$ s (FT), $5 \pm 3$ s (SR), $4 \pm 1$ s (AI-SR)) were lower with SR and AI-SR ($P < .001$ each). Novice readers shifted gaze towards the radiograph in SR, while non-novice readers maintained their focus on the radiograph. AI-SR was the preferred mode. In conclusion, SR improves efficiency by guiding visual attention toward the image, and AI-prefilled SR further enhances diagnostic accuracy and user satisfaction.

[202] Data-Driven Analysis of Intersectional Bias in Image Classification: A Framework with Bias-Weighted Augmentation

Farjana Yesmin

Main category: cs.CV

TL;DR: A framework for analyzing and mitigating intersectional biases in image classification through interpretable fairness evaluation and adaptive data augmentation.

DetailsMotivation: Machine learning models trained on imbalanced datasets exhibit systematic errors from interactions of multiple attributes like object class and environmental conditions.

Method: Introduces Intersectional Fairness Evaluation Framework (IFEF) with quantitative metrics and interpretability tools, plus Bias-Weighted Augmentation (BWA) that adapts transformation intensities based on subgroup statistics.

Result: On Open Images V7 dataset, BWA improves accuracy for underrepresented class-environment intersections by up to 24 percentage points, reduces fairness metric disparities by 35%, with statistically significant improvements (p < 0.05).

Conclusion: Provides a replicable methodology for analyzing and addressing intersectional biases in image classification systems.

Abstract: Machine learning models trained on imbalanced datasets often exhibit intersectional biases-systematic errors arising from the interaction of multiple attributes such as object class and environmental conditions. This paper presents a data-driven framework for analyzing and mitigating such biases in image classification. We introduce the Intersectional Fairness Evaluation Framework (IFEF), which combines quantitative fairness metrics with interpretability tools to systematically identify bias patterns in model predictions. Building on this analysis, we propose Bias-Weighted Augmentation (BWA), a novel data augmentation strategy that adapts transformation intensities based on subgroup distribution statistics. Experiments on the Open Images V7 dataset with five object classes demonstrate that BWA improves accuracy for underrepresented class-environment intersections by up to 24 percentage points while reducing fairness metric disparities by 35%. Statistical analysis across multiple independent runs confirms the significance of improvements (p < 0.05). Our methodology provides a replicable approach for analyzing and addressing intersectional biases in image classification systems.

[203] Differentiable, Bit-shifting, and Scalable Quantization without training neural network from scratch

Zia Badar

Main category: cs.CV

TL;DR: This paper presents a differentiable quantization method for neural networks that provides proof of convergence and supports n-bit quantization, achieving near-full precision accuracy with only 15 training epochs.

DetailsMotivation: Previous quantization approaches lacked differentiability and proper convergence guarantees, and struggled with activation quantization alongside weight quantization while maintaining accuracy.

Method: The authors propose a differentiable quantization approach that supports logarithmic quantization of values in the form 2^n, enabling n-bit quantization without requiring higher precision multiplication.

Result: When tested on ImageNet with ResNet18, the method achieved less than 1% accuracy drop compared to full precision with weight-only quantization, and achieved state-of-the-art accuracy with both weight and activation quantization in just 15 training epochs.

Conclusion: The proposed quantization method provides a differentiable, convergent approach that achieves high accuracy with minimal training epochs and reasonable inference costs, addressing key limitations of previous quantization techniques.

Abstract: Quantization of neural networks provides benefits of inference in less compute and memory requirements. Previous work in quantization lack two important aspects which this work provides. First almost all previous work in quantization used a non-differentiable approach and for learning; the derivative is usually set manually in backpropogation which make the learning ability of algorithm questionable, our approach is not just differentiable, we also provide proof of convergence of our approach to the optimal neural network. Second previous work in shift/logrithmic quantization either have avoided activation quantization along with weight quantization or achieved less accuracy. Learning logrithmic quantize values of form $2^n$ requires the quantization function can scale to more than 1 bit quantization which is another benifit of our quantization that it provides $n$ bits quantization as well. Our approach when tested with image classification task using imagenet dataset, resnet18 and weight quantization only achieves less than 1 percent accuracy compared to full precision accuracy while taking only 15 epochs to train using shift bit quantization and achieves comparable to SOTA approaches accuracy in both weight and activation quantization using shift bit quantization in 15 training epochs with slightly higher(only higher cpu instructions) inference cost compared to 1 bit quantization(without logrithmic quantization) and not requiring any higher precision multiplication.

[204] StripRFNet: A Strip Receptive Field and Shape-Aware Network for Road Damage Detection

Jianhan Lin, Yuchu Qin, Shuai Gao, Yikang Rui, Jie Liu, Yanjie Lv

Main category: cs.CV

TL;DR: StripRFNet is a novel deep neural network for road damage detection that addresses challenges in detecting diverse damage shapes, slender cracks, and small-scale damages through three specialized modules, achieving state-of-the-art performance on benchmark datasets.

DetailsMotivation: Road surface damage threatens traffic safety and hinders sustainable urban development. Accurate detection is challenging due to diverse damage shapes, difficulty capturing slender cracks with high aspect ratios, and high error rates in small-scale damage recognition.

Method: StripRFNet comprises three modules: (1) Shape Perception Module (SPM) with large separable kernel attention for shape discrimination; (2) Strip Receptive Field Module (SRFM) with large strip convolutions and pooling for slender cracks; (3) Small-Scale Enhancement Module (SSEM) with high-resolution P2 feature map, dedicated detection head, and dynamic upsampling for small-object detection.

Result: On RDD2022 benchmark: Chinese subset improved F1-score, mAP50, and mAP50:95 by 4.4, 2.9, and 3.4 percentage points over baseline. On full dataset, achieved highest F1-score of 80.33% compared to CRDDC'2022 participants and ORDDC'2024 Phase 2 results, while maintaining competitive inference speed.

Conclusion: StripRFNet achieves state-of-the-art accuracy and real-time efficiency, offering a promising tool for intelligent road maintenance and sustainable infrastructure management.

Abstract: Well-maintained road networks are crucial for achieving Sustainable Development Goal (SDG) 11. Road surface damage not only threatens traffic safety but also hinders sustainable urban development. Accurate detection, however, remains challenging due to the diverse shapes of damages, the difficulty of capturing slender cracks with high aspect ratios, and the high error rates in small-scale damage recognition. To address these issues, we propose StripRFNet, a novel deep neural network comprising three modules: (1) a Shape Perception Module (SPM) that enhances shape discrimination via large separable kernel attention (LSKA) in multi-scale feature aggregation; (2) a Strip Receptive Field Module (SRFM) that employs large strip convolutions and pooling to capture features of slender cracks; and (3) a Small-Scale Enhancement Module (SSEM) that leverages a high-resolution P2 feature map, a dedicated detection head, and dynamic upsampling to improve small-object detection. Experiments on the RDD2022 benchmark show that StripRFNet surpasses existing methods. On the Chinese subset, it improves F1-score, mAP50, and mAP50:95 by 4.4, 2.9, and 3.4 percentage points over the baseline, respectively. On the full dataset, it achieves the highest F1-score of 80.33% compared with CRDDC'2022 participants and ORDDC'2024 Phase 2 results, while maintaining competitive inference speed. These results demonstrate that StripRFNet achieves state-of-the-art accuracy and real-time efficiency, offering a promising tool for intelligent road maintenance and sustainable infrastructure management.

[205] ObjectTransforms for Uncertainty Quantification and Reduction in Vision-Based Perception for Autonomous Vehicles

Nishad Sahu, Shounak Sural, Aditya Satish Patil, Ragunathan, Rajkumar

Main category: cs.CV

TL;DR: ObjectTransforms is a technique that uses object-specific transformations to quantify and reduce uncertainty in vision-based object detection for autonomous driving, improving robustness and filtering false detections.

DetailsMotivation: Vision-based object detectors are vulnerable to uncertainty from data bias and distributional shifts, which is critical for safety in autonomous driving decision making.

Method: Uses object-specific transformations: color space perturbations on individual objects during training, diffusion models to generate diverse pedestrian instances, and applies object perturbations at inference to quantify uncertainty using detection score variance.

Result: Experiments with YOLOv8 on NuImages dataset show notable accuracy improvements, uncertainty reduction across all object classes, and higher uncertainty values for false positives compared to true positives.

Conclusion: ObjectTransforms is a lightweight yet effective mechanism for reducing uncertainty during training and quantifying uncertainty during inference in vision-based perception systems.

Abstract: Reliable perception is fundamental for safety critical decision making in autonomous driving. Yet, vision based object detector neural networks remain vulnerable to uncertainty arising from issues such as data bias and distributional shifts. In this paper, we introduce ObjectTransforms, a technique for quantifying and reducing uncertainty in vision based object detection through object specific transformations at both training and inference times. At training time, ObjectTransforms perform color space perturbations on individual objects, improving robustness to lighting and color variations. ObjectTransforms also uses diffusion models to generate realistic, diverse pedestrian instances. At inference time, object perturbations are applied to detected objects and the variance of detection scores are used to quantify predictive uncertainty in real time. This uncertainty signal is then used to filter out false positives and also recover false negatives, improving the overall precision recall curve. Experiments with YOLOv8 on the NuImages 10K dataset demonstrate that our method yields notable accuracy improvements and uncertainty reduction across all object classes during training, while predicting desirably higher uncertainty values for false positives as compared to true positives during inference. Our results highlight the potential of ObjectTransforms as a lightweight yet effective mechanism for reducing and quantifying uncertainty in vision-based perception during training and inference respectively.

[206] Text-controlled Motion Mamba: Text-Instructed Temporal Grounding of Human Motion

Xinghan Wang, Zixi Kang, Yadong Mu

Main category: cs.CV

TL;DR: This paper introduces TM-Mamba, a model for text-based human motion grounding that localizes temporal segments in untrimmed motion sequences based on text descriptions, addressing computational challenges of Transformers with linear memory cost.

DetailsMotivation: Existing text-motion tasks focus on generation and editing, but there's a need for precise temporal localization of actions in untrimmed motion sequences, which requires capturing global temporal information efficiently.

Method: Proposes TM-Mamba with text-controlled selection mechanism that dynamically incorporates global temporal information based on text queries, enhanced with relational embeddings for spatial graph topology awareness.

Result: Extensive evaluations on the new BABEL-Grounding dataset demonstrate the effectiveness of TM-Mamba for text-based motion grounding tasks.

Conclusion: TM-Mamba successfully addresses the computational challenges of long untrimmed motion sequences while effectively performing temporal localization based on text descriptions.

Abstract: Human motion understanding is a fundamental task with diverse practical applications, facilitated by the availability of large-scale motion capture datasets. Recent studies focus on text-motion tasks, such as text-based motion generation, editing and question answering. In this study, we introduce the novel task of text-based human motion grounding (THMG), aimed at precisely localizing temporal segments corresponding to given textual descriptions within untrimmed motion sequences. Capturing global temporal information is crucial for the THMG task. However, Transformer-based models that rely on global temporal self-attention face challenges when handling long untrimmed sequences due to the quadratic computational cost. We address these challenges by proposing Text-controlled Motion Mamba (TM-Mamba), a unified model that integrates temporal global context, language query control, and spatial graph topology with only linear memory cost. The core of the model is a text-controlled selection mechanism which dynamically incorporates global temporal information based on text query. The model is further enhanced to be topology-aware through the integration of relational embeddings. For evaluation, we introduce BABEL-Grounding, the first text-motion dataset that provides detailed textual descriptions of human actions along with their corresponding temporal segments. Extensive evaluations demonstrate the effectiveness of TM-Mamba on BABEL-Grounding.

[207] Aria Gen 2 Pilot Dataset

Chen Kong, James Fort, Aria Kang, Jonathan Wittmer, Simon Green, Tianwei Shen, Yipu Zhao, Cheng Peng, Gustavo Solaira, Andrew Berkovich, Nikhil Raina, Vijay Baiyya, Evgeniy Oleinik, Eric Huang, Fan Zhang, Julian Straub, Mark Schwesinger, Luis Pesqueira, Xiaqing Pan, Jakob Julian Engel, Carl Ren, Mingfei Yan, Richard Newcombe

Main category: cs.CV

TL;DR: The Aria Gen 2 Pilot Dataset (A2PD) is an incremental multimodal egocentric dataset captured using Aria Gen 2 glasses, featuring daily activities across five scenarios with raw sensor data and perception algorithm outputs.

DetailsMotivation: To provide timely access to state-of-the-art egocentric multimodal data for research, facilitating the study of human-environment interactions and perception algorithms across diverse users and conditions.

Method: Dataset collection using Aria Gen 2 glasses with a primary subject and friends recording daily activities in five scenarios (cleaning, cooking, eating, playing, outdoor walking), providing both raw sensor data and processed outputs from machine perception algorithms.

Result: A comprehensive publicly available dataset at projectaria.com with open-source tools and usage examples, demonstrating robust performance across diverse users and conditions.

Conclusion: A2PD serves as a valuable resource for egocentric AI research, enabling the development and evaluation of perception algorithms for understanding human activities and interactions with the environment.

Abstract: The Aria Gen 2 Pilot Dataset (A2PD) is an egocentric multimodal open dataset captured using the state-of-the-art Aria Gen 2 glasses. To facilitate timely access, A2PD is released incrementally with ongoing dataset enhancements. The initial release features Dia’ane, our primary subject, who records her daily activities alongside friends, each equipped with Aria Gen 2 glasses. It encompasses five primary scenarios: cleaning, cooking, eating, playing, and outdoor walking. In each of the scenarios, we provide comprehensive raw sensor data and output data from various machine perception algorithms. These data illustrate the device’s ability to perceive the wearer, the surrounding environment, and interactions between the wearer and the environment, while maintaining robust performance across diverse users and conditions. The A2PD is publicly available at projectaria.com, with open-source tools and usage examples provided in Project Aria Tools.

[208] Seeing in the Dark: A Teacher-Student Framework for Dark Video Action Recognition via Knowledge Distillation and Contrastive Learning

Sharana Dharshikgan Suresh Dass, Hrishav Bakul Barua, Ganesh Krishnasamy, Raveendran Paramesran, Raphael C. -W. Phan

Main category: cs.CV

TL;DR: ActLumos is a teacher-student framework for action recognition in dark videos that achieves single-stream inference with multi-stream accuracy through knowledge distillation and dynamic feature fusion.

DetailsMotivation: Action recognition in dark or low-light videos is challenging due to visibility degradation that hinders spatiotemporal details, requiring solutions that maintain accuracy while being computationally efficient.

Method: Teacher uses dual streams (original dark frames + retinex-enhanced frames) with weight-shared R(2+1)D-34 backbones and Dynamic Feature Fusion module. Student uses only dark frames and is pre-trained with self-supervision then fine-tuned with knowledge distillation from teacher.

Result: State-of-the-art accuracy: 96.92% (Top-1) on ARID V1.0, 88.27% on ARID V1.5, and 48.96% on Dark48 under single-stream inference.

Conclusion: The framework successfully transfers multi-stream knowledge to single-stream model, with Dynamic Feature Fusion outperforming static fusion, knowledge distillation enabling transfer of gains, and spatio-temporal SSL surpassing spatial/temporal-only variants.

Abstract: Action recognition in dark or low-light (under-exposed) videos is a challenging task due to visibility degradation, which can hinder critical spatiotemporal details. This paper proposes ActLumos, a teacher-student framework that attains single-stream inference while retaining multi-stream level accuracy. The teacher consumes dual stream inputs, which include original dark frames and retinex-enhanced frames, processed by weight-shared R(2+1)D-34 backbones and fused by a Dynamic Feature Fusion (DFF) module, which dynamically re-weights the two streams at each time step, emphasising the most informative temporal segments. The teacher is also included with a supervised contrastive loss (SupCon) that sharpens class margins. The student shares the R(2+1)D-34 backbone but uses only dark frames and no fusion at test time. The student is first pre-trained with self-supervision on dark clips of both datasets without their labels and then fine-tuned with knowledge distillation from the teacher, transferring the teacher’s multi-stream knowledge into a single-stream model. Under single-stream inference, the distilled student attains state-of-the-art accuracy of 96.92% (Top-1) on ARID V1.0, 88.27% on ARID V1.5, and 48.96% on Dark48. Ablation studies further highlight the individual contributions of each component, i.e., DFF in the teacher outperforms single or static fusion, knowledge distillation (KD) transfers these gains to the single-stream student, and two-view spatio-temporal SSL surpasses spatial-only or temporal-only variants without increasing inference cost. The official website of this work is available at: https://github.com/HrishavBakulBarua/ActLumos

[209] Person Re-Identification via Generalized Class Prototypes

Md Ahmed Al Muzaddid, William J. Beksi

Main category: cs.CV

TL;DR: A new method for selecting better class representatives in person re-identification that improves beyond state-of-the-art by using generalized representations beyond class centroids.

DetailsMotivation: Current approaches focus on feature extraction and objective functions, but class representative selection is underexplored. Prior centroid-based methods yield suboptimal results.

Method: Proposed a generalized selection method for class representations that are not limited to centroids, allowing adjustable number of representations per class to meet application requirements.

Result: The approach substantially improves re-identification performance across multiple embeddings, achieving better balance between accuracy and mean average precision.

Conclusion: The generalized representation selection method effectively addresses limitations of prior centroid-based approaches and advances person re-identification performance beyond current state-of-the-art.

Abstract: Advanced feature extraction methods have significantly contributed to enhancing the task of person re-identification. In addition, modifications to objective functions have been developed to further improve performance. Nonetheless, selecting better class representatives is an underexplored area of research that can also lead to advancements in re-identification performance. Although past works have experimented with using the centroid of a gallery image class during training, only a few have investigated alternative representations during the retrieval stage. In this paper, we demonstrate that these prior techniques yield suboptimal results in terms of re-identification metrics. To address the re-identification problem, we propose a generalized selection method that involves choosing representations that are not limited to class centroids. Our approach strikes a balance between accuracy and mean average precision, leading to improvements beyond the state of the art. For example, the actual number of representations per class can be adjusted to meet specific application requirements. We apply our methodology on top of multiple re-identification embeddings, and in all cases it substantially improves upon contemporary results

[210] GuideFlow3D: Optimization-Guided Rectified Flow For Appearance Transfer

Sayan Deb Sarkar, Sinisa Stekovic, Vincent Lepetit, Iro Armeni

Main category: cs.CV

TL;DR: A training-free method for transferring appearance to 3D assets using universal guidance with pretrained rectified flow models, outperforming direct 3D generative approaches.

DetailsMotivation: Current methods fail when geometry between input and appearance objects differs significantly, and direct 3D generative models produce unappealing results.

Method: Uses pretrained rectified flow models conditioned on image/text with periodic guidance addition during sampling, implemented as differentiable loss functions including part-aware appearance and self-similarity losses.

Result: Successfully transfers texture and geometric details to 3D assets, outperforming baselines both qualitatively and quantitatively. Traditional metrics are unsuitable, so evaluation uses GPT-based ranking system.

Conclusion: The method is general and can be extended to different diffusion models and guidance functions, with robust evaluation confirmed by user studies.

Abstract: Transferring appearance to 3D assets using different representations of the appearance object - such as images or text - has garnered interest due to its wide range of applications in industries like gaming, augmented reality, and digital content creation. However, state-of-the-art methods still fail when the geometry between the input and appearance objects is significantly different. A straightforward approach is to directly apply a 3D generative model, but we show that this ultimately fails to produce appealing results. Instead, we propose a principled approach inspired by universal guidance. Given a pretrained rectified flow model conditioned on image or text, our training-free method interacts with the sampling process by periodically adding guidance. This guidance can be modeled as a differentiable loss function, and we experiment with two different types of guidance including part-aware losses for appearance and self-similarity. Our experiments show that our approach successfully transfers texture and geometric details to the input 3D asset, outperforming baselines both qualitatively and quantitatively. We also show that traditional metrics are not suitable for evaluating the task due to their inability of focusing on local details and comparing dissimilar inputs, in absence of ground truth data. We thus evaluate appearance transfer quality with a GPT-based system objectively ranking outputs, ensuring robust and human-like assessment, as further confirmed by our user study. Beyond showcased scenarios, our method is general and could be extended to different types of diffusion models and guidance functions.

[211] MIRROR: Multi-Modal Pathological Self-Supervised Representation Learning via Modality Alignment and Retention

Tianyi Wang, Jianan Fan, Dingxin Zhang, Dongnan Liu, Yong Xia, Heng Huang, Weidong Cai

Main category: cs.CV

TL;DR: MIRROR is a multi-modal self-supervised learning method that integrates histopathology and transcriptomics data for oncology, focusing on both modality alignment and retention of modality-specific structures.

DetailsMotivation: Histopathology and transcriptomics provide orthogonal yet complementary insights in oncology, but their inherent heterogeneity makes conventional multi-modal alignment methods insufficient as they don't adequately preserve modality-specific fidelity.

Method: MIRROR uses dedicated encoders for each modality, a modality alignment module for integration, a modality retention module to preserve unique attributes, and a style clustering module to reduce redundancy and align pathological signatures in clustering space.

Result: Extensive evaluations on TCGA cohorts for cancer subtyping and survival analysis demonstrate MIRROR’s superior performance in constructing comprehensive oncological feature representations.

Conclusion: MIRROR effectively integrates histopathology and transcriptomics while maintaining modality-specific fidelity, benefiting cancer diagnosis through comprehensive feature representations.

Abstract: Histopathology and transcriptomics are fundamental modalities in oncology, encapsulating the morphological and molecular aspects of the disease. Multi-modal self-supervised learning has demonstrated remarkable potential in learning pathological representations by integrating diverse data sources. Conventional multi-modal integration methods primarily emphasize modality alignment, while paying insufficient attention to retaining the modality-specific structures. However, unlike conventional scenarios where multi-modal inputs share highly overlapping features, histopathology and transcriptomics exhibit pronounced heterogeneity, offering orthogonal yet complementary insights. Histopathology provides morphological and spatial context, elucidating tissue architecture and cellular topology, whereas transcriptomics delineates molecular signatures through gene expression patterns. This inherent disparity introduces a major challenge in aligning them while maintaining modality-specific fidelity. To address these challenges, we present MIRROR, a novel multi-modal representation learning method designed to foster both modality alignment and retention. MIRROR employs dedicated encoders to extract comprehensive features for each modality, which is further complemented by a modality alignment module to achieve seamless integration between phenotype patterns and molecular profiles. Furthermore, a modality retention module safeguards unique attributes from each modality, while a style clustering module mitigates redundancy and enhances disease-relevant information by modeling and aligning consistent pathological signatures within a clustering space. Extensive evaluations on TCGA cohorts for cancer subtyping and survival analysis highlight MIRROR’s superior performance, demonstrating its effectiveness in constructing comprehensive oncological feature representations and benefiting the cancer diagnosis.

[212] C-arm Guidance: A Self-supervised Approach To Automated Positioning During Stroke Thrombectomy

Ahmad Arrabi, Jay hwasung Jung, J Le, A Nguyen, J Reed, E Stahl, Nathan Franssen, Scott Raymond, Safwan Wshah

Main category: cs.CV

TL;DR: Deep learning framework for automating thrombectomy procedures using self-supervised landmark classification with regression-based pretext tasks.

DetailsMotivation: Thrombectomy is effective for ischemic stroke but resource-intensive; automation can enhance efficiency and safety.

Method: Self-supervised framework that classifies skeletal landmarks using regression-based pretext tasks.

Result: Model outperforms existing methods in regression and classification; positional pretext task significantly improves downstream classification.

Conclusion: Framework shows promise for autonomous C-arm control to optimize trajectories from pelvis to head during thrombectomy procedures.

Abstract: Thrombectomy is one of the most effective treatments for ischemic stroke, but it is resource and personnel-intensive. We propose employing deep learning to automate critical aspects of thrombectomy, thereby enhancing efficiency and safety. In this work, we introduce a self-supervised framework that classifies various skeletal landmarks using a regression-based pretext task. Our experiments demonstrate that our model outperforms existing methods in both regression and classification tasks. Notably, our results indicate that the positional pretext task significantly enhances downstream classification performance. Future work will focus on extending this framework toward fully autonomous C-arm control, aiming to optimize trajectories from the pelvis to the head during stroke thrombectomy procedures. All code used is available at https://github.com/AhmadArrabi/C_arm_guidance

[213] DuetMatch: Harmonizing Semi-Supervised Brain MRI Segmentation via Decoupled Branch Optimization

Thanh-Huy Nguyen, Hoang-Thien Nguyen, Vi Vu, Ba-Thinh Lam, Phat Huynh, Tianyang Wang, Xingjian Li, Ulas Bagci, Min Xu

Main category: cs.CV

TL;DR: DuetMatch is a dual-branch semi-supervised framework for medical image segmentation that uses asynchronous optimization, decoupled dropout perturbation, pair-wise CutMix cross-guidance, and consistency matching to improve performance with limited annotated data.

DetailsMotivation: Limited availability of annotated medical imaging data makes semi-supervised learning appealing, but joint optimization in teacher-student frameworks can hinder convergence and stability, especially in challenging scenarios.

Method: Dual-branch framework with asynchronous optimization (each branch optimizes either encoder or decoder while keeping the other frozen), Decoupled Dropout Perturbation for regularization, Pair-wise CutMix Cross-Guidance for model diversity, and Consistency Matching to mitigate confirmation bias from noisy pseudo-labels.

Result: Extensive experiments on benchmark brain MRI segmentation datasets (ISLES2022 and BraTS) show that DuetMatch consistently outperforms state-of-the-art methods.

Conclusion: DuetMatch demonstrates effectiveness and robustness across diverse semi-supervised segmentation scenarios in medical imaging.

Abstract: The limited availability of annotated data in medical imaging makes semi-supervised learning increasingly appealing for its ability to learn from imperfect supervision. Recently, teacher-student frameworks have gained popularity for their training benefits and robust performance. However, jointly optimizing the entire network can hinder convergence and stability, especially in challenging scenarios. To address this for medical image segmentation, we propose DuetMatch, a novel dual-branch semi-supervised framework with asynchronous optimization, where each branch optimizes either the encoder or decoder while keeping the other frozen. To improve consistency under noisy conditions, we introduce Decoupled Dropout Perturbation, enforcing regularization across branches. We also design Pair-wise CutMix Cross-Guidance to enhance model diversity by exchanging pseudo-labels through augmented input pairs. To mitigate confirmation bias from noisy pseudo-labels, we propose Consistency Matching, refining labels using stable predictions from frozen teacher models. Extensive experiments on benchmark brain MRI segmentation datasets, including ISLES2022 and BraTS, show that DuetMatch consistently outperforms state-of-the-art methods, demonstrating its effectiveness and robustness across diverse semi-supervised segmentation scenarios.

[214] ActAlign: Zero-Shot Fine-Grained Video Classification via Language-Guided Sequence Alignment

Amir Aghdam, Vincent Tao Hu, Björn Ommer

Main category: cs.CV

TL;DR: ActAlign is a zero-shot video classification method that uses LLM-generated sub-action sequences aligned with video frames via Dynamic Time Warping, achieving strong performance on fine-grained action recognition without training.

DetailsMotivation: To enable zero-shot video classification for extremely fine-grained actions where no video examples or temporal annotations are available for unseen classes, leveraging the generalization of image-language models while addressing their lack of temporal modeling.

Method: Formulates video classification as sequence alignment: LLM generates ordered sub-action sequences for each class, then aligns them with video frames using Dynamic Time Warping in a shared embedding space. Training-free and model-agnostic.

Result: Achieves 30.5% accuracy on ActionAtlas benchmark (human performance: 61.6%), outperforms billion-parameter video-language models while using 8x fewer parameters. Demonstrates domain-general applicability.

Conclusion: Structured language priors combined with classical alignment methods can unlock the open-set recognition potential of image-language models for fine-grained video understanding without video-text supervision or fine-tuning.

Abstract: We address the task of zero-shot video classification for extremely fine-grained actions (e.g., Windmill Dunk in basketball), where no video examples or temporal annotations are available for unseen classes. While image-language models (e.g., CLIP, SigLIP) show strong open-set recognition, they lack temporal modeling needed for video understanding. We propose ActAlign, a truly zero-shot, training-free method that formulates video classification as a sequence alignment problem, preserving the generalization strength of pretrained image-language models. For each class, a large language model (LLM) generates an ordered sequence of sub-actions, which we align with video frames using Dynamic Time Warping (DTW) in a shared embedding space. Without any video-text supervision or fine-tuning, ActAlign achieves 30.5% accuracy on ActionAtlas–the most diverse benchmark of fine-grained actions across multiple sports–where human performance is only 61.6%. ActAlign outperforms billion-parameter video-language models while using 8x fewer parameters. Our approach is model-agnostic and domain-general, demonstrating that structured language priors combined with classical alignment methods can unlock the open-set recognition potential of image-language models for fine-grained video understanding.

[215] Automated C-Arm Positioning via Conformal Landmark Localization

Ahmad Arrabi, Jay Hwasung Jung, Jax Luo, Nathan Franssen, Scott Raymond, Safwan Wshah

Main category: cs.CV

TL;DR: A pipeline for autonomous C-arm navigation to anatomical landmarks using X-ray images, with uncertainty quantification and conformal prediction for reliable deployment.

DetailsMotivation: Manual C-arm positioning in fluoroscopy-guided interventions increases radiation exposure and procedural delays, necessitating automated solutions.

Method: Uses X-ray images from arbitrary starting positions to predict 3D displacement vectors to target landmarks, incorporates aleatoric and epistemic uncertainty modeling with conformal prediction, and combines probabilistic loss with skeletal pose regularization.

Result: Strong localization accuracy across multiple architectures with well-calibrated prediction bounds on synthetic X-ray data from DeepDRR.

Conclusion: The pipeline shows potential as a component for safe and reliable autonomous C-arm systems in clinical workflows.

Abstract: Accurate and reliable C-arm positioning is essential for fluoroscopy-guided interventions. However, clinical workflows rely on manual alignment that increases radiation exposure and procedural delays. In this work, we present a pipeline that autonomously navigates the C-arm to predefined anatomical landmarks utilizing X-ray images. Given an input X-ray image from an arbitrary starting location on the operating table, the model predicts a 3D displacement vector toward each target landmark along the body. To ensure reliable deployment, we capture both aleatoric and epistemic uncertainties in the model’s predictions and further calibrate them using conformal prediction. The derived prediction regions are interpreted as 3D confidence regions around the predicted landmark locations. The training framework combines a probabilistic loss with skeletal pose regularization to encourage anatomically plausible outputs. We validate our approach on a synthetic X-ray dataset generated from DeepDRR. Results show not only strong localization accuracy across multiple architectures but also well-calibrated prediction bounds. These findings highlight the pipeline’s potential as a component in safe and reliable autonomous C-arm systems. Code is available at https://github.com/AhmadArrabi/C_arm_guidance_APAH

[216] Cost Savings from Automatic Quality Assessment of Generated Images

Xavier Giro-i-Nieto, Nefeli Andreou, Anqi Liang, Manel Baradad, Francesc Moreno-Noguer, Aleix Martinez

Main category: cs.CV

TL;DR: The paper presents a cost-saving formula for automatic image quality assessment (IQA) pre-filtering in deep generative model workflows, demonstrating 51.61% cost reduction in background inpainting using AutoML.

DetailsMotivation: Current deep generative models produce low-yield high-quality images requiring expensive manual IQA, creating a need for automated pre-filtering to reduce costs.

Method: Developed a formula to estimate cost savings based on precision and pass yield of IQA engines, and applied it with AutoML in background inpainting use case.

Result: Achieved significant 51.61% cost saving in background inpainting scenario using simple AutoML solution for automatic IQA pre-filtering.

Conclusion: Automatic IQA pre-filtering can substantially reduce costs in generative AI workflows, with the presented formula providing a framework for estimating potential savings.

Abstract: Deep generative models have shown impressive progress in recent years, making it possible to produce high quality images with a simple text prompt or a reference image. However, state of the art technology does not yet meet the quality standards offered by traditional photographic methods. For this reason, production pipelines that use generated images often include a manual stage of image quality assessment (IQA). This process is slow and expensive, especially because of the low yield of automatically generated images that pass the quality bar. The IQA workload can be reduced by introducing an automatic pre-filtering stage, that will increase the overall quality of the images sent to review and, therefore, reduce the average cost required to obtain a high quality image. We present a formula that estimates the cost savings depending on the precision and pass yield of a generic IQA engine. This formula is applied in a use case of background inpainting, showcasing a significant cost saving of 51.61% obtained with a simple AutoML solution.

[217] Seeing Through the Brain: New Insights from Decoding Visual Stimuli with fMRI

Zheng Huang, Enpei Zhang, Yinghao Cai, Weikang Qiu, Carl Yang, Elynn Chen, Xiang Zhang, Rex Ying, Dawei Zhou, Yujun Yan

Main category: cs.CV

TL;DR: PRISM projects fMRI signals into structured text space for better visual stimulus reconstruction, outperforming existing methods by 8% in perceptual loss.

DetailsMotivation: To understand how the brain encodes visual information by reconstructing images from fMRI signals, and to determine the optimal latent space for this transformation.

Method: Projects fMRI signals into structured text space using PRISM model, includes object-centric diffusion module for image composition and attribute-relationship search module to align with neural activity.

Result: Extensive experiments show PRISM outperforms existing methods with up to 8% reduction in perceptual loss on real-world datasets.

Conclusion: Structured text space is superior for bridging fMRI signals and image reconstruction, capturing the compositional nature of visual stimuli effectively.

Abstract: Understanding how the brain encodes visual information is a central challenge in neuroscience and machine learning. A promising approach is to reconstruct visual stimuli, essentially images, from functional Magnetic Resonance Imaging (fMRI) signals. This involves two stages: transforming fMRI signals into a latent space and then using a pretrained generative model to reconstruct images. The reconstruction quality depends on how similar the latent space is to the structure of neural activity and how well the generative model produces images from that space. Yet, it remains unclear which type of latent space best supports this transformation and how it should be organized to represent visual stimuli effectively. We present two key findings. First, fMRI signals are more similar to the text space of a language model than to either a vision based space or a joint text image space. Second, text representations and the generative model should be adapted to capture the compositional nature of visual stimuli, including objects, their detailed attributes, and relationships. Building on these insights, we propose PRISM, a model that Projects fMRI sIgnals into a Structured text space as an interMediate representation for visual stimuli reconstruction. It includes an object centric diffusion module that generates images by composing individual objects to reduce object detection errors, and an attribute relationship search module that automatically identifies key attributes and relationships that best align with the neural activity. Extensive experiments on real world datasets demonstrate that our framework outperforms existing methods, achieving up to an 8% reduction in perceptual loss. These results highlight the importance of using structured text as the intermediate space to bridge fMRI signals and image reconstruction.

[218] Data-Centric AI for Tropical Agricultural Mapping: Challenges, Strategies and Scalable Solutions

Mateus Pinto da Silva, Sabrina P. L. P. Correa, Hugo N. Oliveira, Ian M. Nunes, Jefersson A. dos Santos

Main category: cs.CV

TL;DR: This paper proposes a Data-Centric AI pipeline for agricultural mapping in tropical areas, focusing on data quality and curation to overcome challenges like limited annotated data, high labeling costs, and regional variability.

DetailsMotivation: Tropical agriculture mapping faces unique challenges including high cloudiness, diverse crop calendars, limited datasets, and poor generalization of traditional model-centric approaches, requiring a shift to data-centric methods.

Method: The paper advocates a Data-Centric AI perspective, reviewing and prioritizing techniques like confident learning, core-set selection, data augmentation, and active learning. It identifies 25 distinct strategies and proposes a practical pipeline using the 9 most mature methods.

Result: The paper highlights the readiness and suitability of 25 data-centric strategies for large-scale agricultural mapping pipelines, with a focus on tropical contexts where traditional approaches are limited.

Conclusion: A data-centric approach using curated practical methods is better suited for tropical agriculture mapping, addressing the dynamic realities and limitations of traditional model-centric approaches in these challenging environments.

Abstract: Mapping agriculture in tropical areas through remote sensing presents unique challenges, including the lack of high-quality annotated data, the elevated costs of labeling, data variability, and regional generalisation. This paper advocates a Data-Centric Artificial Intelligence (DCAI) perspective and pipeline, emphasizing data quality and curation as key drivers for model robustness and scalability. It reviews and prioritizes techniques such as confident learning, core-set selection, data augmentation, and active learning. The paper highlights the readiness and suitability of 25 distinct strategies in large-scale agricultural mapping pipelines. The tropical context is of high interest, since high cloudiness, diverse crop calendars, and limited datasets limit traditional model-centric approaches. This tutorial outlines practical solutions as a data-centric approach for curating and training AI models better suited to the dynamic realities of tropical agriculture. Finally, we propose a practical pipeline using the 9 most mature and straightforward methods that can be applied to a large-scale tropical agricultural mapping project.

[219] StretchySnake: Flexible SSM Training Unlocks Action Recognition Across Spatio-Temporal Scales

Nyle Siddiqui, Rohit Gupta, Sirnam Swetha, Mubarak Shah

Main category: cs.CV

TL;DR: Proposes StretchySnake, a flexible training method for video SSMs that enables spatio-temporal flexibility by sampling videos at varying resolutions and dynamically interpolating model weights, outperforming transformers and SSM baselines by up to 28% across diverse action recognition benchmarks.

DetailsMotivation: Current training methods for video understanding are tailored for transformers and fail to leverage SSMs' unique attributes, leading to spatio-temporal inflexibility where models perform poorly on videos with unseen spatial and temporal resolutions.

Method: Samples videos at varying temporal and spatial resolutions during training and dynamically interpolates model weights to accommodate any spatio-temporal scale. Introduces and compares five variants of flexible training for video SSMs.

Result: StretchySnake outperforms transformer and SSM baselines by up to 28% on short-action (UCF-101, HMDB-51) and long-action (COIN, Breakfast) benchmarks, with strong adaptability to fine-grained actions (SSV2, Diving-48).

Conclusion: The method provides a simple drop-in training recipe that makes video SSMs more robust, resolution-agnostic, and efficient across diverse action recognition scenarios.

Abstract: State space models (SSMs) have emerged as a competitive alternative to transformers in various tasks. Their linear complexity and hidden-state recurrence make them particularly attractive for modeling long sequences, whereas attention becomes quadratically expensive. However, current training methods for video understanding are tailored towards transformers and fail to fully leverage the unique attributes of SSMs. For example, video models are often trained at a fixed resolution and video length to balance the quadratic scaling of attention cost against performance. Consequently, these models suffer from degraded performance when evaluated on videos with spatial and temporal resolutions unseen during training; a property we call spatio-temporal inflexibility. In the context of action recognition, this severely limits a model’s ability to retain performance across both short- and long-form videos. Therefore, we propose a flexible training method that leverages and improves the inherent adaptability of SSMs. Our method samples videos at varying temporal and spatial resolutions during training and dynamically interpolates model weights to accommodate any spatio-temporal scale. This instills our SSM, which we call StretchySnake, with spatio-temporal flexibility and enables it to seamlessly handle videos ranging from short, fine-grained clips to long, complex activities. We introduce and compare five different variants of flexible training, and identify the most effective strategy for video SSMs. On short-action (UCF-101, HMDB-51) and long-action (COIN, Breakfast) benchmarks, StretchySnake outperforms transformer and SSM baselines alike by up to 28%, with strong adaptability to fine-grained actions (SSV2, Diving-48). Therefore, our method provides a simple drop-in training recipe that makes video SSMs more robust, resolution-agnostic, and efficient across diverse action recognition scenarios.

[220] VM-BeautyNet: A Synergistic Ensemble of Vision Transformer and Mamba for Facial Beauty Prediction

Djamel Eddine Boukhari

Main category: cs.CV

TL;DR: VM-BeautyNet is a novel ensemble model combining Vision Transformer and Mamba-based Vision model for facial beauty prediction, achieving state-of-the-art results on SCUT-FBP5500 dataset with improved global feature capture and linear complexity.

DetailsMotivation: Existing CNN-based models struggle to capture global facial features important for beauty perception, while Vision Transformers have quadratic complexity limitations. There's a need for models that can efficiently capture both global structure and long-range dependencies.

Method: Proposed VM-BeautyNet, a heterogeneous ensemble architecture that fuses Vision Transformer (for global facial structure and symmetry) with Mamba-based Vision model (for efficient long-range dependency modeling with linear complexity). Uses complementary feature extraction from both backbones.

Result: Achieved state-of-the-art performance on SCUT-FBP5500 dataset: Pearson Correlation of 0.9212, MAE of 0.2085, RMSE of 0.2698. Grad-CAM visualizations confirmed complementary feature extraction between the two backbones.

Conclusion: VM-BeautyNet presents a powerful new architectural paradigm for computational aesthetics, successfully combining global feature capture with efficient long-range dependency modeling, while providing interpretable insights into the decision-making process.

Abstract: Facial Beauty Prediction (FBP) is a complex and challenging computer vision task, aiming to model the subjective and intricate nature of human aesthetic perception. While deep learning models, particularly Convolutional Neural Networks (CNNs), have made significant strides, they often struggle to capture the global, holistic facial features that are critical to human judgment. Vision Transformers (ViT) address this by effectively modeling long-range spatial relationships, but their quadratic complexity can be a bottleneck. This paper introduces a novel, heterogeneous ensemble architecture, \textbf{VM-BeautyNet}, that synergistically fuses the complementary strengths of a Vision Transformer and a Mamba-based Vision model, a recent advancement in State-Space Models (SSMs). The ViT backbone excels at capturing global facial structure and symmetry, while the Mamba backbone efficiently models long-range dependencies with linear complexity, focusing on sequential features and textures. We evaluate our approach on the benchmark SCUT-FBP5500 dataset. Our proposed VM-BeautyNet achieves state-of-the-art performance, with a \textbf{Pearson Correlation (PC) of 0.9212}, a \textbf{Mean Absolute Error (MAE) of 0.2085}, and a \textbf{Root Mean Square Error (RMSE) of 0.2698}. Furthermore, through Grad-CAM visualizations, we provide interpretability analysis that confirms the complementary feature extraction of the two backbones, offering new insights into the model’s decision-making process and presenting a powerful new architectural paradigm for computational aesthetics.

[221] Designing a Convolutional Neural Network for High-Accuracy Oral Cavity Squamous Cell Carcinoma (OCSCC) Detection

Vishal Manikanden, Aniketh Bandlamudi, Daniel Haehn

Main category: cs.CV

TL;DR: A CNN-based system for early detection of Oral Cavity Squamous Cell Carcinoma (OCSCC) using image analysis and hardware enhancement.

DetailsMotivation: OCSCC often goes undetected due to subtle early stages and hidden development areas, leading to preventable deaths, making early detection crucial.

Method: Trained a CNN on 4293 images of benign/malignant tumors and negative samples, tested on images at 5 resolutions, and developed hardware for detailed image capture.

Result: Higher resolution images led to more accurate predictions on a logarithmic scale, showing diminishing returns with increased pixel counts.

Conclusion: CNN with proper image resolution can effectively detect OCSCC, and hardware enhancement improves detection accuracy.

Abstract: Oral Cavity Squamous Cell Carcinoma (OCSCC) is the most common type of head and neck cancer. Due to the subtle nature of its early stages, deep and hidden areas of development, and slow growth, OCSCC often goes undetected, leading to preventable deaths. However, properly trained Convolutional Neural Networks (CNNs), with their precise image segmentation techniques and ability to apply kernel matrices to modify the RGB values of images for accurate image pattern recognition, would be an effective means for early detection of OCSCC. Pairing this neural network with image capturing and processing hardware would allow increased efficacy in OCSCC detection. The aim of our project is to develop a Convolutional Neural Network trained to recognize OCSCC, as well as to design a physical hardware system to capture and process detailed images, in order to determine the image quality required for accurate predictions. A CNN was trained on 4293 training images consisting of benign and malignant tumors, as well as negative samples, and was evaluated for its precision, recall, and Mean Average Precision (mAP) in its predictions of OCSCC. A testing dataset of randomly assorted images of cancerous, non-cancerous, and negative images was chosen, and each image was altered to represent 5 common resolutions. This test data set was thoroughly analyzed by the CNN and predictions were scored on the basis of accuracy. The designed enhancement hardware was used to capture detailed images, and its impact was scored. An application was developed to facilitate the testing process and bring open access to the CNN. Images of increasing resolution resulted in higher-accuracy predictions on a logarithmic scale, demonstrating the diminishing returns of higher pixel counts.

[222] Embody 3D: A Large-scale Multimodal Motion and Behavior Dataset

Claire McLean, Makenzie Meendering, Tristan Swartz, Orri Gabbay, Alexandra Olsen, Rachel Jacobs, Nicholas Rosen, Philippe de Bree, Tony Garcia, Gadsden Merrill, Jake Sandakly, Julia Buffalini, Neham Jain, Steven Krenn, Moneish Kumar, Dejan Markovic, Evonne Ng, Fabian Prada, Andrew Saba, Siwei Zhang, Vasu Agrawal, Tim Godisart, Alexander Richard, Michael Zollhoefer

Main category: cs.CV

TL;DR: Embody 3D is a large-scale multimodal dataset with 500 hours of 3D motion data from 439 participants, featuring both single-person and multi-person interactions with comprehensive tracking and annotations.

DetailsMotivation: To create a comprehensive multimodal dataset for human motion analysis and behavioral studies, addressing the need for large-scale 3D motion data with diverse interaction scenarios.

Method: Collected data from 439 participants in a multi-camera stage, capturing 54 million frames of tracked 3D motion including hand tracking, body shape, text annotations, and separate audio tracks for each participant.

Result: Successfully created a dataset with 500 hours of 3D motion data covering prompted motions, hand gestures, locomotion, discussions, emotional conversations, collaborative activities, and co-living scenarios.

Conclusion: Embody 3D provides a valuable resource for research in human motion analysis, behavioral studies, and multimodal interaction modeling, with its comprehensive tracking and diverse interaction scenarios.

Abstract: The Codec Avatars Lab at Meta introduces Embody 3D, a multimodal dataset of 500 individual hours of 3D motion data from 439 participants collected in a multi-camera collection stage, amounting to over 54 million frames of tracked 3D motion. The dataset features a wide range of single-person motion data, including prompted motions, hand gestures, and locomotion; as well as multi-person behavioral and conversational data like discussions, conversations in different emotional states, collaborative activities, and co-living scenarios in an apartment-like space. We provide tracked human motion including hand tracking and body shape, text annotations, and a separate audio track for each participant.

[223] Proactive Scene Decomposition and Reconstruction

Baicheng Li, Zike Yan, Dong Wu, Hongbin Zha

Main category: cs.CV

TL;DR: Proactive scene decomposition and reconstruction using human-object interactions to dynamically refine environment modeling in real-time.

DetailsMotivation: Human behaviors provide rich cues about scene dynamics, and leveraging these intentional interactions can address ambiguities in static object-level reconstruction methods.

Method: Online approach that integrates camera/object pose estimation, instance decomposition, and map updating using human-object interactions from egocentric live streams, aided by Gaussian splatting for photorealistic rendering.

Result: Achieves accurate and consistent dynamic scene modeling with photorealistic and efficient rendering, validated in multiple real-world scenarios.

Conclusion: The system provides a flexible, progressive alternative to conventional object-level reconstruction methods by capitalizing on human-object interaction cues.

Abstract: Human behaviors are the major causes of scene dynamics and inherently contain rich cues regarding the dynamics. This paper formalizes a new task of proactive scene decomposition and reconstruction, an online approach that leverages human-object interactions to iteratively disassemble and reconstruct the environment. By observing these intentional interactions, we can dynamically refine the decomposition and reconstruction process, addressing inherent ambiguities in static object-level reconstruction. The proposed system effectively integrates multiple tasks in dynamic environments such as accurate camera and object pose estimation, instance decomposition, and online map updating, capitalizing on cues from human-object interactions in egocentric live streams for a flexible, progressive alternative to conventional object-level reconstruction methods. Aided by the Gaussian splatting technique, accurate and consistent dynamic scene modeling is achieved with photorealistic and efficient rendering. The efficacy is validated in multiple real-world scenarios with promising advantages.

[224] Cerberus: Real-Time Video Anomaly Detection via Cascaded Vision-Language Models

Yue Zheng, Xiufang Shi, Jiming Chen, Yuanchao Shu

Main category: cs.CV

TL;DR: Cerberus is a two-stage cascaded system for real-time video anomaly detection that combines lightweight filtering with VLM reasoning, achieving 57.68 fps and 97.2% accuracy while being 151.79x faster than VLM-based methods.

DetailsMotivation: Current VLM-based VAD methods have high computational costs and unstable visual grounding, making them impractical for real-time deployment. There's a need for efficient yet accurate real-time anomaly detection systems.

Method: Two-stage cascaded system: 1) Learns normal behavioral rules offline, 2) Combines lightweight filtering with fine-grained VLM reasoning during online inference. Uses motion mask prompting and rule-based deviation detection to focus VLM attention on relevant motion regions and identify anomalies as deviations from learned norms.

Result: Achieves 57.68 fps on NVIDIA L40S GPU (151.79x speedup compared to VLM-based methods) with 97.2% accuracy comparable to state-of-the-art VLM-based VAD methods across four datasets.

Conclusion: Cerberus establishes itself as a practical solution for real-time video analytics by balancing efficiency and accuracy through its cascaded architecture and innovative attention mechanisms.

Abstract: Video anomaly detection (VAD) has rapidly advanced by recent development of Vision-Language Models (VLMs). While these models offer superior zero-shot detection capabilities, their immense computational cost and unstable visual grounding performance hinder real-time deployment. To overcome these challenges, we introduce Cerberus, a two-stage cascaded system designed for efficient yet accurate real-time VAD. Cerberus learns normal behavioral rules offline, and combines lightweight filtering with fine-grained VLM reasoning during online inference. The performance gains of Cerberus come from two key innovations: motion mask prompting and rule-based deviation detection. The former directs the VLM’s attention to regions relevant to motion, while the latter identifies anomalies as deviations from learned norms rather than enumerating possible anomalies. Extensive evaluations on four datasets show that Cerberus on average achieves 57.68 fps on an NVIDIA L40S GPU, a 151.79$\times$ speedup, and 97.2% accuracy comparable to the state-of-the-art VLM-based VAD methods, establishing it as a practical solution for real-time video analytics.

[225] OpenLVLM-MIA: A Controlled Benchmark Revealing the Limits of Membership Inference Attacks on Large Vision-Language Models

Ryoto Miyamoto, Xin Fan, Fuyuko Kido, Tsuneo Matsumoto, Hayato Yamana

Main category: cs.CV

TL;DR: OpenLVLM-MIA is a new benchmark that addresses distributional bias issues in membership inference attack (MIA) evaluation for large vision-language models, showing that previous high success rates were due to dataset bias rather than true membership detection.

DetailsMotivation: To address fundamental challenges in evaluating membership inference attacks against LVLMs, where prior work reported high success rates but these results were actually due to distributional bias from dataset construction rather than true membership detection.

Method: Created a controlled benchmark with 6,000 images where member and non-member sample distributions are carefully balanced, with ground-truth membership labels provided across three distinct training stages.

Result: Experiments showed that state-of-the-art MIA methods converged to random chance performance under unbiased conditions, revealing that previous high success rates were artifacts of dataset bias.

Conclusion: OpenLVLM-MIA provides a transparent and unbiased benchmark that clarifies current limitations of MIA research on LVLMs and offers a solid foundation for developing stronger privacy-preserving techniques.

Abstract: OpenLVLM-MIA is a new benchmark that highlights fundamental challenges in evaluating membership inference attacks (MIA) against large vision-language models (LVLMs). While prior work has reported high attack success rates, our analysis suggests that these results often arise from detecting distributional bias introduced during dataset construction rather than from identifying true membership status. To address this issue, we introduce a controlled benchmark of 6{,}000 images where the distributions of member and non-member samples are carefully balanced, and ground-truth membership labels are provided across three distinct training stages. Experiments using OpenLVLM-MIA demonstrated that the performance of state-of-the-art MIA methods converged to random chance under unbiased conditions. By offering a transparent and unbiased benchmark, OpenLVLM-MIA clarifies the current limitations of MIA research on LVLMs and provides a solid foundation for developing stronger privacy-preserving techniques.

[226] Stroke2Sketch: Harnessing Stroke Attributes for Training-Free Sketch Generation

Rui Yang, Huining Li, Yiyi Long, Xiaojun Wu, Shengfeng He

Main category: cs.CV

TL;DR: Stroke2Sketch is a training-free framework for sketch generation that transfers stroke attributes from reference styles using cross-image stroke attention while preserving semantic structure and content fidelity.

DetailsMotivation: To enable precise transfer of stroke attributes (line thickness, deformation, texture sparsity) from reference styles to content images while maintaining semantic structure and content fidelity in sketch generation.

Method: Proposes cross-image stroke attention mechanism embedded in self-attention layers to establish fine-grained semantic correspondences for stroke attribute transfer, plus adaptive contrast enhancement and semantic-focused attention for content preservation.

Result: Effectively synthesizes stylistically faithful sketches resembling handcrafted results, outperforming existing methods in expressive stroke control and semantic coherence.

Conclusion: Stroke2Sketch provides an effective training-free solution for style-guided sketch generation with accurate stroke attribute transfer and strong content preservation capabilities.

Abstract: Generating sketches guided by reference styles requires precise transfer of stroke attributes, such as line thickness, deformation, and texture sparsity, while preserving semantic structure and content fidelity. To this end, we propose Stroke2Sketch, a novel training-free framework that introduces cross-image stroke attention, a mechanism embedded within self-attention layers to establish fine-grained semantic correspondences and enable accurate stroke attribute transfer. This allows our method to adaptively integrate reference stroke characteristics into content images while maintaining structural integrity. Additionally, we develop adaptive contrast enhancement and semantic-focused attention to reinforce content preservation and foreground emphasis. Stroke2Sketch effectively synthesizes stylistically faithful sketches that closely resemble handcrafted results, outperforming existing methods in expressive stroke control and semantic coherence. Codes are available at https://github.com/rane7/Stroke2Sketch.

[227] Scaling Laws for Deepfake Detection

Wenhao Wang, Longqi Cai, Taihong Xiao, Yuxiao Wang, Ming-Hsuan Yang

Main category: cs.CV

TL;DR: This paper studies scaling laws in deepfake detection, showing that detection error follows power-law decay as training data (real domains and fake methods) increases, similar to LLMs.

DetailsMotivation: To systematically analyze how deepfake detection performance scales with data quantity and diversity, since no existing dataset was large enough for this research.

Method: Created ScaleDF dataset with 5.8M real images from 51 domains and 8.8M fake images from 102 methods, then analyzed scaling relationships and power-law patterns.

Result: Found predictable power-law scaling where detection error decreases as number of real domains or deepfake methods increases, enabling performance forecasting.

Conclusion: Scaling laws apply to deepfake detection, suggesting data-centric approaches can counter evolving deepfake technology, though scaling has limitations.

Abstract: This paper presents a systematic study of scaling laws for the deepfake detection task. Specifically, we analyze the model performance against the number of real image domains, deepfake generation methods, and training images. Since no existing dataset meets the scale requirements for this research, we construct ScaleDF, the largest dataset to date in this field, which contains over 5.8 million real images from 51 different datasets (domains) and more than 8.8 million fake images generated by 102 deepfake methods. Using ScaleDF, we observe power-law scaling similar to that shown in large language models (LLMs). Specifically, the average detection error follows a predictable power-law decay as either the number of real domains or the number of deepfake methods increases. This key observation not only allows us to forecast the number of additional real domains or deepfake methods required to reach a target performance, but also inspires us to counter the evolving deepfake technology in a data-centric manner. Beyond this, we examine the role of pre-training and data augmentations in deepfake detection under scaling, as well as the limitations of scaling itself.

[228] Scale-DiT: Ultra-High-Resolution Image Generation with Hierarchical Local Attention

Yuyao Zhang, Yu-Wing Tai

Main category: cs.CV

TL;DR: Scale-DiT enables ultra-high-resolution (4K) text-to-image generation using hierarchical local attention with global guidance, achieving efficient inference without requiring native 4K training data.

DetailsMotivation: Current diffusion models are limited to sub-1K resolutions due to quadratic attention complexity and lack of native 4K training data, creating a need for scalable high-resolution generation methods.

Method: Divides high-resolution latents into local windows for near-linear attention complexity, uses low-resolution latent with positional anchors for global semantics, and employs LoRA adaptation to bridge global-local pathways with Hilbert curve token ordering for efficiency.

Result: Achieves 2× faster inference with lower memory usage compared to dense attention, scales to 4K resolution without additional training data, and delivers superior global coherence and local detail on quantitative metrics (FID, IS, CLIP Score).

Conclusion: Hierarchical local attention with guided low-resolution anchors is an effective approach for advancing ultra-high-resolution image generation, matching or outperforming methods requiring native 4K training.

Abstract: Ultra-high-resolution text-to-image generation demands both fine-grained texture synthesis and globally coherent structure, yet current diffusion models remain constrained to sub-$1K \times 1K$ resolutions due to the prohibitive quadratic complexity of attention and the scarcity of native $4K$ training data. We present \textbf{Scale-DiT}, a new diffusion framework that introduces hierarchical local attention with low-resolution global guidance, enabling efficient, scalable, and semantically coherent image synthesis at ultra-high resolutions. Specifically, high-resolution latents are divided into fixed-size local windows to reduce attention complexity from quadratic to near-linear, while a low-resolution latent equipped with scaled positional anchors injects global semantics. A lightweight LoRA adaptation bridges global and local pathways during denoising, ensuring consistency across structure and detail. To maximize inference efficiency, we repermute token sequence in Hilbert curve order and implement a fused-kernel for skipping masked operations, resulting in a GPU-friendly design. Extensive experiments demonstrate that Scale-DiT achieves more than $2\times$ faster inference and lower memory usage compared to dense attention baselines, while reliably scaling to $4K \times 4K$ resolution without requiring additional high-resolution training data. On both quantitative benchmarks (FID, IS, CLIP Score) and qualitative comparisons, Scale-DiT delivers superior global coherence and sharper local detail, matching or outperforming state-of-the-art methods that rely on native 4K training. Taken together, these results highlight hierarchical local attention with guided low-resolution anchors as a promising and effective approach for advancing ultra-high-resolution image generation.

[229] Limitations of Data-Driven Spectral Reconstruction – An Optics-Aware Analysis

Qiang Fu, Matheus Souza, Eunsue Choi, Suhyun Shin, Seung-Hwan Baek, Wolfgang Heidrich

Main category: cs.CV

TL;DR: This paper systematically analyzes limitations of RGB-to-spectral reconstruction methods, revealing overfitting issues, fundamental metameric constraints, and dataset insufficiencies, while exploring optical encoding as a potential improvement.

DetailsMotivation: Current RGB-to-spectral reconstruction methods show high numerical scores but poor real-world performance, creating a need to understand the fundamental limitations and identify paths for improvement.

Method: Systematic analysis including: 1) Overfitting evaluation with reduced training data and cross-dataset validation, 2) Testing with metamer data using metameric black theory, 3) Exploring optical encoding via optical aberrations and deliberate optical design.

Result: RGB-to-spectral methods suffer from overfitting to existing datasets, fundamental inability to handle metameric conditions, and dataset insufficiencies. Optical encoding provides some improvement but remains limited by dataset issues.

Conclusion: Future progress in snapshot spectral imaging depends on generating improved datasets that can enable effective optical encoding strategies, as the RGB-to-spectral inverse problem remains fundamentally ill-posed.

Abstract: Hyperspectral imaging empowers machine vision systems with the distinct capability of identifying materials through recording their spectral signatures. Recent efforts in data-driven spectral reconstruction aim at extracting spectral information from RGB images captured by cost-effective RGB cameras, instead of dedicated hardware. Published work reports exceedingly high numerical scores for this reconstruction task, yet real-world performance lags substantially behind. We systematically analyze the performance of such methods. First, we evaluate the overfitting limitations with respect to current datasets by training the networks with less data, validating the trained models with unseen yet slightly modified data and cross-dataset validation. Second, we reveal fundamental limitations in the ability of RGB to spectral methods to deal with metameric or near-metameric conditions, which have so far gone largely unnoticed due to the insufficiencies of existing datasets. We validate the trained models with metamer data generated by metameric black theory and re-training the networks with various forms of metamers. This methodology can also be used for data augmentation as a partial mitigation of the dataset issues, although the RGB to spectral inverse problem remains fundamentally ill-posed. Finally, we analyze the potential for modifying the problem setting to achieve better performance by exploiting optical encoding provided by either optical aberrations or deliberate optical design. Our experiments show such approaches provide improved results under certain circumstances, but their overall performance is limited by the same dataset issues. We conclude that future progress on snapshot spectral imaging will heavily depend on the generation of improved datasets which can then be used to design effective optical encoding strategies. Code: https://github.com/vccimaging/OpticsAwareHSI-Analysis.

[230] DiffusionX: Efficient Edge-Cloud Collaborative Image Generation with Multi-Round Prompt Evolution

Yi Wei, Shunpu Tang, Liang Zhao, Qiangian Yang

Main category: cs.CV

TL;DR: DiffusionX is a cloud-edge collaborative framework that reduces diffusion model generation time by 15.8% while maintaining image quality through lightweight on-device previews and cloud-based final refinements.

DetailsMotivation: Address the computational intensity of diffusion models and the need for iterative prompt refinement, which increases latency and burdens cloud resources.

Method: A cloud-edge collaborative framework with lightweight on-device diffusion model for rapid previews and high-capacity cloud model for final refinements, plus a noise level predictor for dynamic computation load balancing.

Result: Reduces average generation time by 15.8% compared to Stable Diffusion v1.5 with comparable quality, and only 0.9% slower than Tiny-SD with significantly improved image quality.

Conclusion: DiffusionX demonstrates efficient and scalable prompt-based generation with minimal overhead through cloud-edge collaboration.

Abstract: Recent advances in diffusion models have driven remarkable progress in image generation. However, the generation process remains computationally intensive, and users often need to iteratively refine prompts to achieve the desired results, further increasing latency and placing a heavy burden on cloud resources. To address this challenge, we propose DiffusionX, a cloud-edge collaborative framework for efficient multi-round, prompt-based generation. In this system, a lightweight on-device diffusion model interacts with users by rapidly producing preview images, while a high-capacity cloud model performs final refinements after the prompt is finalized. We further introduce a noise level predictor that dynamically balances the computation load, optimizing the trade-off between latency and cloud workload. Experiments show that DiffusionX reduces average generation time by 15.8% compared with Stable Diffusion v1.5, while maintaining comparable image quality. Moreover, it is only 0.9% slower than Tiny-SD with significantly improved image quality, thereby demonstrating efficiency and scalability with minimal overhead.

[231] Improvement of Spiking Neural Network with Bit Planes and Color Models

Nhan T. Luu, Duong T. Luu, Nam N. Pham, Thang C. Truong

Main category: cs.CV

TL;DR: A novel bit plane coding method for spiking neural networks (SNNs) that improves image classification accuracy without increasing model size, with investigation of color model impacts.

DetailsMotivation: SNNs offer low energy consumption and small memory footprint but face performance optimization challenges. This research aims to enhance SNN performance for images through innovative coding methods.

Method: Proposed a new coding approach using bit plane representation to process images in SNNs, and investigated the effects of different color models on the coding process.

Result: Experimental validation showed effectiveness in achieving performance gains across multiple datasets. This is the first research to consider bit planes and color models in SNN context.

Conclusion: The bit plane coding strategy unlocks new potentials in SNN performance, potentially enabling more efficient and effective SNN models for future applications.

Abstract: Spiking neural network (SNN) has emerged as a promising paradigm in computational neuroscience and artificial intelligence, offering advantages such as low energy consumption and small memory footprint. However, their practical adoption is constrained by several challenges, prominently among them being performance optimization. In this study, we present a novel approach to enhance the performance of SNN for images through a new coding method that exploits bit plane representation. Our proposed technique is designed to improve the accuracy of SNN without increasing model size. Also, we investigate the impacts of color models of the proposed coding process. Through extensive experimental validation, we demonstrate the effectiveness of our coding strategy in achieving performance gain across multiple datasets. To the best of our knowledge, this is the first research that considers bit planes and color models in the context of SNN. By leveraging the unique characteristics of bit planes, we hope to unlock new potentials in SNNs performance, potentially paving the way for more efficient and effective SNNs models in future researches and applications.

[232] TokenAR: Multiple Subject Generation via Autoregressive Token-level enhancement

Haiyue Sun, Qingdong He, Jinlong Peng, Peng Tang, Jiangning Zhang, Junwei Zhu, Xiaobin Hu, Shuicheng Yan

Main category: cs.CV

TL;DR: TokenAR is a token-level enhancement framework for autoregressive models that addresses identity confusion in multi-reference image generation through three key components: token index embedding, instruct token injection, and identity-token disentanglement strategy.

DetailsMotivation: Autoregressive models struggle with decoupling different reference identities in multiple reference generation, leading to identity confusion problems.

Method: Three-part token-level enhancement: 1) Token Index Embedding clusters tokens for better reference representation, 2) Instruct Token Injection adds extra visual features as complementary priors, 3) Identity-Token Disentanglement (ITD) explicitly guides tokens to independently represent each identity’s features.

Result: The framework significantly improves identity consistency while preserving high-quality background reconstruction. Comprehensive experiments show it surpasses current state-of-the-art models in multiple reference image generation.

Conclusion: TokenAR effectively addresses identity confusion in multi-reference generation and introduces the InstructAR Dataset, the first large-scale open-source dataset for this task with 28K training pairs.

Abstract: Autoregressive Model (AR) has shown remarkable success in conditional image generation. However, these approaches for multiple reference generation struggle with decoupling different reference identities. In this work, we propose the TokenAR framework, specifically focused on a simple but effective token-level enhancement mechanism to address reference identity confusion problem. Such token-level enhancement consists of three parts, 1). Token Index Embedding clusters the tokens index for better representing the same reference images; 2). Instruct Token Injection plays as a role of extra visual feature container to inject detailed and complementary priors for reference tokens; 3). The identity-token disentanglement strategy (ITD) explicitly guides the token representations toward independently representing the features of each identity.This token-enhancement framework significantly augments the capabilities of existing AR based methods in conditional image generation, enabling good identity consistency while preserving high quality background reconstruction. Driven by the goal of high-quality and high-diversity in multi-subject generation, we introduce the InstructAR Dataset, the first open-source, large-scale, multi-reference input, open domain image generation dataset that includes 28K training pairs, each example has two reference subjects, a relative prompt and a background with mask annotation, curated for multiple reference image generation training and evaluating. Comprehensive experiments validate that our approach surpasses current state-of-the-art models in multiple reference image generation task. The implementation code and datasets will be made publicly. Codes are available, see https://github.com/lyrig/TokenAR

[233] RL makes MLLMs see better than SFT

Junha Song, Sangdoo Yun, Dongyoon Han, Jaegul Choo, Byeongho Heo

Main category: cs.CV

TL;DR: The paper challenges the assumption that MLLM performance mainly comes from the LLM backbone, showing that training strategies (especially RL vs SFT) fundamentally reshape vision encoder representations, with RL producing stronger and more localized visual representations.

DetailsMotivation: There's a significant gap in understanding how vision encoders in MLLMs are affected by different training paradigms, particularly the shift from SFT to RL, and how this impacts MLLM performance.

Method: Conducted diverse experiments including ImageNet classification, segmentation, and gradient visualization to analyze vision encoder representations after different training strategies (SFT vs RL). Proposed PIVOT method for building strong vision encoders.

Result: RL produces stronger and more precisely localized visual representations than SFT. PIVOT-trained vision encoders outperform larger counterparts with less than 1% computational cost of standard pretraining.

Conclusion: Training strategy fundamentally reshapes MLLM’s visual representations, with RL offering superior performance. PIVOT provides an efficient path for advancing MLLM vision backbones.

Abstract: A dominant assumption in Multimodal Language Model (MLLM) research is that its performance is largely inherited from the LLM backbone, given its immense parameter scale and remarkable capabilities. This has created a void in the understanding of the vision encoder, which determines how MLLMs perceive images. The recent shift in MLLM training paradigms, from Supervised Finetuning (SFT) to Reinforcement Learning (RL), magnifies this oversight-namely, the significant lack of analysis on how such training reshapes the vision encoder as well as the MLLM. To address this, we first investigate the impact of training strategies on MLLMs, where RL shows a clear advantage over SFT in strongly vision-related VQA benchmarks. Motivated by this, we conduct a critical yet under-explored analysis of the vision encoder of MLLMs through diverse and in-depth experiments, ranging from ImageNet classification and segmentation to gradient visualization. Our results demonstrate that MLLM’s post-training strategy (i.e., SFT or RL) not only leads to distinct outcomes on MLLM downstream tasks, but also fundamentally reshapes MLLM’s underlying visual representations. Specifically, the key finding of our study is that RL produces stronger and precisely localized visual representations compared to SFT, boosting the ability of the vision encoder for MLLM. We then reframe our findings into a simple recipe for building strong vision encoders for MLLMs, Preference-Instructed Vision OpTimization (PIVOT). When integrated into MLLMs, a PIVOT-trained vision encoder outperforms even larger and more heavily-trained counterparts, despite requiring less than 1% of the computational cost of standard vision pretraining. This result opens an effective and efficient path for advancing the vision backbones of MLLMs. Project page available at https://june-page.github.io/pivot/

[234] On the Provable Importance of Gradients for Language-Assisted Image Clustering

Bo Peng, Jie Lu, Guangquan Zhang, Zhen Fang

Main category: cs.CV

TL;DR: Proposes GradNorm, a gradient-based framework for filtering positive nouns in Language-assisted Image Clustering, with theoretical guarantees and state-of-the-art performance.

DetailsMotivation: Existing filtering strategies for Language-assisted Image Clustering lack theoretical foundation despite using CLIP features. Need for rigorous methods to identify semantically relevant nouns from unlabeled data.

Method: GradNorm measures noun positiveness using gradient magnitudes from cross-entropy between predicted target distribution and softmax output. Provides theoretical error bounds and subsumes existing methods as special cases.

Result: Extensive experiments show GradNorm achieves state-of-the-art clustering performance across various benchmarks.

Conclusion: GradNorm offers a theoretically grounded and empirically superior approach for noun filtering in Language-assisted Image Clustering, outperforming existing intuitive methods.

Abstract: This paper investigates the recently emerged problem of Language-assisted Image Clustering (LaIC), where textual semantics are leveraged to improve the discriminability of visual representations to facilitate image clustering. Due to the unavailability of true class names, one of core challenges of LaIC lies in how to filter positive nouns, i.e., those semantically close to the images of interest, from unlabeled wild corpus data. Existing filtering strategies are predominantly based on the off-the-shelf feature space learned by CLIP; however, despite being intuitive, these strategies lack a rigorous theoretical foundation. To fill this gap, we propose a novel gradient-based framework, termed as GradNorm, which is theoretically guaranteed and shows strong empirical performance. In particular, we measure the positiveness of each noun based on the magnitude of gradients back-propagated from the cross-entropy between the predicted target distribution and the softmax output. Theoretically, we provide a rigorous error bound to quantify the separability of positive nouns by GradNorm and prove that GradNorm naturally subsumes existing filtering strategies as extremely special cases of itself. Empirically, extensive experiments show that GradNorm achieves the state-of-the-art clustering performance on various benchmarks.

[235] MIRAD - A comprehensive real-world robust anomaly detection dataset for Mass Individualization

Pulin Li, Guocheng Wu, Li Yin, Yuxin Zheng, Wei Zhang, Yanjie Zhou

Main category: cs.CV

TL;DR: The paper introduces MIRAD, the first benchmark dataset for anomaly detection in social manufacturing, addressing challenges of mass individualization, fragmented production, and environmental variations across distributed sites.

DetailsMotivation: Social manufacturing enables mass individualization but faces significant quality control challenges due to customized products, small-batch orders, and varying imaging environments across distributed sites.

Method: Created the MIRAD dataset capturing three critical dimensions: diverse individualized products with large intra-class variation, data from six geographically dispersed manufacturing nodes, and substantial imaging heterogeneity. Evaluated state-of-the-art anomaly detection methods including one-class, multi-class, and zero-shot approaches.

Result: All evaluated models showed significant performance drops compared to conventional benchmarks, highlighting the unresolved complexities of defect detection in real-world individualized production.

Conclusion: MIRAD bridges industrial requirements and academic research, providing a realistic foundation for developing robust quality control solutions essential for Industry 5.0.

Abstract: Social manufacturing leverages community collaboration and scattered resources to realize mass individualization in modern industry. However, this paradigm shift also introduces substantial challenges in quality control, particularly in defect detection. The main difficulties stem from three aspects. First, products often have highly customized configurations. Second, production typically involves fragmented, small-batch orders. Third, imaging environments vary considerably across distributed sites. To overcome the scarcity of real-world datasets and tailored algorithms, we introduce the Mass Individualization Robust Anomaly Detection (MIRAD) dataset. As the first benchmark explicitly designed for anomaly detection in social manufacturing, MIRAD captures three critical dimensions of this domain: (1) diverse individualized products with large intra-class variation, (2) data collected from six geographically dispersed manufacturing nodes, and (3) substantial imaging heterogeneity, including variations in lighting, background, and motion conditions. We then conduct extensive evaluations of state-of-the-art (SOTA) anomaly detection methods on MIRAD, covering one-class, multi-class, and zero-shot approaches. Results show a significant performance drop across all models compared with conventional benchmarks, highlighting the unresolved complexities of defect detection in real-world individualized production. By bridging industrial requirements and academic research, MIRAD provides a realistic foundation for developing robust quality control solutions essential for Industry 5.0. The dataset is publicly available at https://github.com/wu33learn/MIRAD.

[236] Cataract-LMM: Large-Scale, Multi-Source, Multi-Task Benchmark for Deep Learning in Surgical Video Analysis

Mohammad Javad Ahmadi, Iman Gandomi, Parisa Abdi, Seyed-Farzad Mohammadi, Amirhossein Taslimi, Mehdi Khodaparast, Hassan Hashemi, Mahdi Tavakoli, Hamid D. Taghirad

Main category: cs.CV

TL;DR: A large-scale dataset of 3,000 cataract surgery videos with comprehensive annotations for surgical AI development, including temporal phases, instance segmentation, instrument-tissue interactions, and skill scores.

DetailsMotivation: Current cataract surgery datasets lack diversity and annotation depth needed to train generalizable deep-learning models for computer-assisted surgery systems.

Method: Collected 3,000 phacoemulsification cataract surgery videos from two surgical centers with surgeons of varying experience levels, enriched with four annotation layers: temporal surgical phases, instance segmentation, instrument-tissue interaction tracking, and quantitative skill scores based on ICO-OSCAR rubrics.

Result: Benchmarking experiments demonstrated the dataset’s technical quality for key surgical AI tasks including workflow recognition, scene segmentation, and automated skill assessment. Established domain adaptation baseline for phase recognition across surgical centers.

Conclusion: The presented dataset addresses the gap in diverse, deeply annotated cataract surgery resources and supports development of generalizable surgical AI models through comprehensive annotations and benchmarking.

Abstract: The development of computer-assisted surgery systems depends on large-scale, annotated datasets. Current resources for cataract surgery often lack the diversity and annotation depth needed to train generalizable deep-learning models. To address this gap, we present a dataset of 3,000 phacoemulsification cataract surgery videos from two surgical centers, performed by surgeons with a range of experience levels. This resource is enriched with four annotation layers: temporal surgical phases, instance segmentation of instruments and anatomical structures, instrument-tissue interaction tracking, and quantitative skill scores based on the established competency rubrics like the ICO-OSCAR. The technical quality of the dataset is supported by a series of benchmarking experiments for key surgical AI tasks, including workflow recognition, scene segmentation, and automated skill assessment. Furthermore, we establish a domain adaptation baseline for the phase recognition task by training a model on a subset of surgical centers and evaluating its performance on a held-out center. The dataset and annotations are available in Google Form (https://docs.google.com/forms/d/e/1FAIpQLSfmyMAPSTGrIy2sTnz0-TMw08ZagTimRulbAQcWdaPwDy187A/viewform?usp=dialog).

[237] iWatchRoadv2: Pothole Detection, Geospatial Mapping, and Intelligent Road Governance

Rishi Raj Sahoo, Surbhi Saswati Mohanty, Subhankar Mishra

Main category: cs.CV

TL;DR: iWatchRoadv2 is an automated platform for real-time pothole detection, GPS geotagging, and road health visualization using YOLO models and OpenStreetMap, with governance features for contractor accountability.

DetailsMotivation: Road potholes pose significant safety hazards and maintenance challenges on India's diverse and under-maintained road networks, requiring automated solutions for efficient monitoring and repair.

Method: Used self-annotated dataset of 7,000 dashcam frames to fine-tune YOLO model, synchronized OCR timestamps with GPS logs for geolocation, and built backend database with road segment attribution and contractor information.

Result: Developed a fully automated end-to-end platform that detects potholes in real-time, geotags them, and provides dynamic road health visualization with intelligent governance features for automated alerts and accountability.

Conclusion: iWatchRoadv2 enables data-driven smart city management, transparent governance, and sustainable improvements in road infrastructure maintenance by automating the complete pothole monitoring lifecycle from detection to repair verification.

Abstract: Road potholes pose significant safety hazards and maintenance challenges, particularly on India’s diverse and under-maintained road networks. This paper presents iWatchRoadv2, a fully automated end-to-end platform for real-time pothole detection, GPS-based geotagging, and dynamic road health visualization using OpenStreetMap (OSM). We curated a self-annotated dataset of over 7,000 dashcam frames capturing diverse Indian road conditions, weather patterns, and lighting scenarios, which we used to fine-tune the Ultralytics YOLO model for accurate pothole detection. The system synchronizes OCR-extracted video timestamps with external GPS logs to precisely geolocate each detected pothole, enriching detections with comprehensive metadata, including road segment attribution and contractor information managed through an optimized backend database. iWatchRoadv2 introduces intelligent governance features that enable authorities to link road segments with contract metadata through a secure login interface. The system automatically sends alerts to contractors and officials when road health deteriorates, supporting automated accountability and warranty enforcement. The intuitive web interface delivers actionable analytics to stakeholders and the public, facilitating evidence-driven repair planning, budget allocation, and quality assessment. Our cost-effective and scalable solution streamlines frame processing and storage while supporting seamless public engagement for urban and rural deployments. By automating the complete pothole monitoring lifecycle, from detection to repair verification, iWatchRoadv2 enables data-driven smart city management, transparent governance, and sustainable improvements in road infrastructure maintenance. The platform and live demonstration are accessible at https://smlab.niser.ac.in/project/iwatchroad.

[238] Demeter: A Parametric Model of Crop Plant Morphology from the Real World

Tianhang Cheng, Albert J. Zhai, Evan Z. Chen, Rui Zhou, Yawen Deng, Zitong Li, Kejie Zhao, Janice Shiu, Qianyu Zhao, Yide Xu, Xinlei Wang, Yuan Shen, Sheng Wang, Lisa Ainsworth, Kaiyu Guan, Shenlong Wang

Main category: cs.CV

TL;DR: Demeter is a parametric 3D shape model for plants that encodes topology, shape, articulation, and deformation into a learned representation, addressing limitations in existing plant modeling approaches.

DetailsMotivation: Existing parametric models work well for humans and animals but lack expressiveness for plants, which have varying topology across species and multiple sources of shape variation.

Method: Developed a data-driven parametric model that handles varying shape topology and models three sources of shape variation: articulation, subcomponent shape variation, and non-rigid deformation. Used a large-scale soybean farm dataset for training and testing.

Result: Demeter effectively synthesizes shapes, reconstructs structures, and simulates biophysical processes, demonstrating practical utility in agricultural applications.

Conclusion: Demeter advances plant modeling by providing an expressive parametric framework that captures the complex morphological variations in plants, with potential applications in 3D reconstruction, generation, understanding, and simulation for agricultural research.

Abstract: Learning 3D parametric shape models of objects has gained popularity in vision and graphics and has showed broad utility in 3D reconstruction, generation, understanding, and simulation. While powerful models exist for humans and animals, equally expressive approaches for modeling plants are lacking. In this work, we present Demeter, a data-driven parametric model that encodes key factors of a plant morphology, including topology, shape, articulation, and deformation into a compact learned representation. Unlike previous parametric models, Demeter handles varying shape topology across various species and models three sources of shape variation: articulation, subcomponent shape variation, and non-rigid deformation. To advance crop plant modeling, we collected a large-scale, ground-truthed dataset from a soybean farm as a testbed. Experiments show that Demeter effectively synthesizes shapes, reconstructs structures, and simulates biophysical processes. Code and data is available at https://tianhang-cheng.github.io/Demeter/.

[239] SPLite Hand: Sparsity-Aware Lightweight 3D Hand Pose Estimation

Yeh Keng Hao, Hsu Tzu Wei, Sun Min

Main category: cs.CV

TL;DR: A lightweight framework for hand pose estimation on AR/VR edge devices that combines sparse convolution, a novel SPLite decoder, and quantization to achieve significant efficiency gains while maintaining accuracy.

DetailsMotivation: Address the challenge of deploying deep learning models on edge devices for AR/VR applications, which require real-time inference, low power consumption, and minimal latency while balancing efficiency and performance.

Method: Uses encoder-decoder architecture with sparse convolution on ResNet-18 backbone to exploit sparsity in hand pose images, introduces SPLite decoder for faster decoding, and applies quantization-aware training for memory optimization.

Result: Achieved 42% end-to-end efficiency improvement, 3.1x frame rate boost on Raspberry Pi 5, 2.98x overall speed-up on Raspberry Pi 5 CPU, with minimal accuracy loss (PA-MPJPE increased only from 9.0mm to 9.1mm on FreiHAND).

Conclusion: The proposed framework demonstrates comparable accuracy to state-of-the-art methods while significantly enhancing computational efficiency, making it suitable for real-time hand pose estimation on resource-constrained edge devices.

Abstract: With the increasing ubiquity of AR/VR devices, the deployment of deep learning models on edge devices has become a critical challenge. These devices require real-time inference, low power consumption, and minimal latency. Many framework designers face the conundrum of balancing efficiency and performance. We design a light framework that adopts an encoder-decoder architecture and introduces several key contributions aimed at improving both efficiency and accuracy. We apply sparse convolution on a ResNet-18 backbone to exploit the inherent sparsity in hand pose images, achieving a 42% end-to-end efficiency improvement. Moreover, we propose our SPLite decoder. This new architecture significantly boosts the decoding process’s frame rate by 3.1x on the Raspberry Pi 5, while maintaining accuracy on par. To further optimize performance, we apply quantization-aware training, reducing memory usage while preserving accuracy (PA-MPJPE increases only marginally from 9.0 mm to 9.1 mm on FreiHAND). Overall, our system achieves a 2.98x speed-up on a Raspberry Pi 5 CPU (BCM2712 quad-core Arm A76 processor). Our method is also evaluated on compound benchmark datasets, demonstrating comparable accuracy to state-of-the-art approaches while significantly enhancing computational efficiency.

[240] REALM: An MLLM-Agent Framework for Open World 3D Reasoning Segmentation and Editing on Gaussian Splatting

Changyue Shi, Minghao Chen, Yiping Mao, Chuxiao Yang, Xinyuan Hu, Jiajun Ding, Zhou Yu

Main category: cs.CV

TL;DR: REALM is an MLLM-agent framework that bridges 2D vision-language reasoning with 3D spatial understanding for open-world reasoning-based segmentation, using 3D Gaussian Splatting representations and a novel Global-to-Local Spatial Grounding strategy.

DetailsMotivation: Existing 3D segmentation methods struggle with ambiguous, reasoning-based instructions, while 2D vision-language models lack 3D spatial understanding, creating a gap between complex human instructions and precise 3D object grounding.

Method: Segmentation on 3D Gaussian Splatting representations with Global-to-Local Spatial Grounding: multiple global views for coarse localization, then close-up novel views for fine-grained local segmentation, aggregating MLLM responses for robust target identification.

Result: REALM achieves remarkable performance in interpreting both explicit and implicit instructions across LERF, 3D-OVS, and REALM3D benchmarks, and supports various 3D interaction tasks including object removal, replacement, and style transfer.

Conclusion: REALM demonstrates practical utility and versatility as an agent framework that enables open-world reasoning-based segmentation without extensive 3D-specific post-training, effectively bridging 2D reasoning capabilities with 3D spatial understanding.

Abstract: Bridging the gap between complex human instructions and precise 3D object grounding remains a significant challenge in vision and robotics. Existing 3D segmentation methods often struggle to interpret ambiguous, reasoning-based instructions, while 2D vision-language models that excel at such reasoning lack intrinsic 3D spatial understanding. In this paper, we introduce REALM, an innovative MLLM-agent framework that enables open-world reasoning-based segmentation without requiring extensive 3D-specific post-training. We perform segmentation directly on 3D Gaussian Splatting representations, capitalizing on their ability to render photorealistic novel views that are highly suitable for MLLM comprehension. As directly feeding one or more rendered views to the MLLM can lead to high sensitivity to viewpoint selection, we propose a novel Global-to-Local Spatial Grounding strategy. Specifically, multiple global views are first fed into the MLLM agent in parallel for coarse-level localization, aggregating responses to robustly identify the target object. Then, several close-up novel views of the object are synthesized to perform fine-grained local segmentation, yielding accurate and consistent 3D masks. Extensive experiments show that REALM achieves remarkable performance in interpreting both explicit and implicit instructions across LERF, 3D-OVS, and our newly introduced REALM3D benchmarks. Furthermore, our agent framework seamlessly supports a range of 3D interaction tasks, including object removal, replacement, and style transfer, demonstrating its practical utility and versatility. Project page: https://ChangyueShi.github.io/REALM.

[241] SSL4RL: Revisiting Self-supervised Learning as Intrinsic Reward for Visual-Language Reasoning

Xiaojun Guo, Runyu Zhou, Yifei Wang, Qi Zhang, Chenheng Zhang, Stefanie Jegelka, Xiaohan Wang, Jiajun Chai, Guojun Yin, Wei Lin, Yisen Wang

Main category: cs.CV

TL;DR: SSL4RL is a framework that uses self-supervised learning tasks as verifiable rewards for reinforcement learning fine-tuning of vision-language models, eliminating the need for human preference data.

DetailsMotivation: Vision-language models often fail to adequately utilize visual evidence, relying instead on linguistic priors or textual shortcuts. Current RL approaches for VLMs lack scalable and reliable reward mechanisms.

Method: Reformulates self-supervised learning objectives (like predicting image rotation or reconstructing masked patches) into dense, automatic reward signals for RL-based fine-tuning.

Result: Substantially improves performance on both vision-centric and vision-language reasoning benchmarks. Also demonstrates generality by achieving significant gains in graph learning.

Conclusion: SSL4RL establishes a versatile and effective paradigm for aligning multimodal models using verifiable, self-supervised objectives, with identified key factors influencing effectiveness including task difficulty, model scale, and semantic alignment.

Abstract: Vision-language models (VLMs) have shown remarkable abilities by integrating large language models with visual inputs. However, they often fail to utilize visual evidence adequately, either depending on linguistic priors in vision-centric tasks or resorting to textual shortcuts during reasoning. Although reinforcement learning (RL) can align models with desired behaviors, its application to VLMs has been hindered by the lack of scalable and reliable reward mechanisms. To overcome this challenge, we propose SSL4RL, a novel framework that leverages self-supervised learning (SSL) tasks as a source of verifiable rewards for RL-based fine-tuning. Our approach reformulates SSL objectives-such as predicting image rotation or reconstructing masked patches-into dense, automatic reward signals, eliminating the need for human preference data or unreliable AI evaluators. Experiments show that SSL4RL substantially improves performance on both vision-centric and vision-language reasoning benchmarks. Furthermore, through systematic ablations, we identify key factors-such as task difficulty, model scale, and semantic alignment with the target domain-that influence the effectiveness of SSL4RL tasks, offering new design principles for future work. We also demonstrate the framework’s generality by applying it to graph learning, where it yields significant gains. SSL4RL establishes a versatile and effective paradigm for aligning multimodal models using verifiable, self-supervised objectives.

[242] LightGlueStick: a Fast and Robust Glue for Joint Point-Line Matching

Aidyn Ubingazhibov, Rémi Pautrat, Iago Suárez, Shaohui Liu, Marc Pollefeys, Viktor Larsson

Main category: cs.CV

TL;DR: LightGlueStick is a lightweight matcher for points and line segments that achieves state-of-the-art performance through Attentional Line Message Passing (ALMP), enabling efficient communication between nodes while maintaining computational efficiency.

DetailsMotivation: Traditional point and line matching are treated as independent tasks, and while GlueStick proposed joint matching, its heavy architecture prevented real-time applications or deployment to edge devices.

Method: Proposes LightGlueStick with Attentional Line Message Passing (ALMP) that explicitly exposes line connectivity to the network, allowing efficient communication between nodes in a lightweight architecture.

Result: LightGlueStick establishes a new state-of-the-art across different benchmarks while maintaining computational efficiency suitable for real-time applications.

Conclusion: The proposed lightweight matcher successfully combines point and line matching with efficient architecture, making it suitable for real-time applications and edge device deployment.

Abstract: Lines and points are complementary local features, whose combination has proven effective for applications such as SLAM and Structure-from-Motion. The backbone of these pipelines are the local feature matchers, establishing correspondences across images. Traditionally, point and line matching have been treated as independent tasks. Recently, GlueStick proposed a GNN-based network that simultaneously operates on points and lines to establish matches. While running a single joint matching reduced the overall computational complexity, the heavy architecture prevented real-time applications or deployment to edge devices. Inspired by recent progress in point matching, we propose LightGlueStick, a lightweight matcher for points and line segments. The key novel component in our architecture is the Attentional Line Message Passing (ALMP), which explicitly exposes the connectivity of the lines to the network, allowing for efficient communication between nodes. In thorough experiments we show that LightGlueStick establishes a new state-of-the-art across different benchmarks. The code is available at https://github.com/aubingazhib/LightGlueStick.

[243] EDVD-LLaMA: Explainable Deepfake Video Detection via Multimodal Large Language Model Reasoning

Haoran Sun, Chen Cai, Huiping Zhuang, Kong Aik Lee, Lap-Pui Chau, Yi Wang

Main category: cs.CV

TL;DR: This paper proposes EDVD-LLaMA, an explainable deepfake video detection framework using multimodal large language models that provides traceable reasoning processes alongside detection results.

DetailsMotivation: Traditional deepfake detection methods lack transparency and generalization capabilities, creating an urgent need for detectors that can identify forged content while providing verifiable reasoning explanations.

Method: Uses Spatio-Temporal Subtle Information Tokenization (ST-SIT) to extract cross-frame deepfake features, and Fine-grained Multimodal Chain-of-Thought (Fg-MCoT) with facial feature constraints for pixel-level spatio-temporal localization.

Result: EDVD-LLaMA achieves outstanding performance and robustness in detection accuracy, explainability, and cross-forgery/cross-dataset scenarios, outperforming previous methods.

Conclusion: The framework provides a more explainable and superior solution for deepfake video detection compared to traditional methods, with publicly available source code and dataset.

Abstract: The rapid development of deepfake video technology has not only facilitated artistic creation but also made it easier to spread misinformation. Traditional deepfake video detection (DVD) methods face issues such as a lack of transparency in their principles and insufficient generalization capabilities to cope with evolving forgery techniques. This highlights an urgent need for detectors that can identify forged content and provide verifiable reasoning explanations. This paper proposes the explainable deepfake video detection (EDVD) task and designs the EDVD-LLaMA multimodal, a large language model (MLLM) reasoning framework, which provides traceable reasoning processes alongside accurate detection results and trustworthy explanations. Our approach first incorporates a Spatio-Temporal Subtle Information Tokenization (ST-SIT) to extract and fuse global and local cross-frame deepfake features, providing rich spatio-temporal semantic information input for MLLM reasoning. Second, we construct a Fine-grained Multimodal Chain-of-Thought (Fg-MCoT) mechanism, which introduces facial feature data as hard constraints during the reasoning process to achieve pixel-level spatio-temporal video localization, suppress hallucinated outputs, and enhance the reliability of the chain of thought. In addition, we build an Explainable Reasoning FF++ benchmark dataset (ER-FF++set), leveraging structured data to annotate videos and ensure quality control, thereby supporting dual supervision for reasoning and detection. Extensive experiments demonstrate that EDVD-LLaMA achieves outstanding performance and robustness in terms of detection accuracy, explainability, and its ability to handle cross-forgery methods and cross-dataset scenarios. Compared to previous DVD methods, it provides a more explainable and superior solution. The source code and dataset will be publicly available.

[244] Enhancing Rotated Object Detection via Anisotropic Gaussian Bounding Box and Bhattacharyya Distance

Chien Thai, Mai Xuan Trang, Huong Ninh, Hoang Hiep Ly, Anh Son Le

Main category: cs.CV

TL;DR: Improved loss function using Gaussian bounding box representation and Bhattacharyya distance for better rotated object detection accuracy.

DetailsMotivation: Traditional object detection frameworks underperform for rotated objects due to limitations in capturing orientation variations, especially in aerial imagery and autonomous driving applications.

Method: Proposed rotation-invariant loss function leveraging Gaussian bounding box representation and Bhattacharyya distance, with anisotropic Gaussian representation to address isotropic variance issues in square-like objects.

Result: Significant improvements in mean Average Precision metrics compared to existing methods when integrated into state-of-the-art deep learning detectors.

Conclusion: The approach shows potential to establish new benchmarks in rotated object detection with wide applications requiring precise object localization regardless of orientation.

Abstract: Detecting rotated objects accurately and efficiently is a significant challenge in computer vision, particularly in applications such as aerial imagery, remote sensing, and autonomous driving. Although traditional object detection frameworks are effective for axis-aligned objects, they often underperform in scenarios involving rotated objects due to their limitations in capturing orientation variations. This paper introduces an improved loss function aimed at enhancing detection accuracy and robustness by leveraging the Gaussian bounding box representation and Bhattacharyya distance. In addition, we advocate for the use of an anisotropic Gaussian representation to address the issues associated with isotropic variance in square-like objects. Our proposed method addresses these challenges by incorporating a rotation-invariant loss function that effectively captures the geometric properties of rotated objects. We integrate this proposed loss function into state-of-the-art deep learning-based rotated object detection detectors, and extensive experiments demonstrated significant improvements in mean Average Precision metrics compared to existing methods. The results highlight the potential of our approach to establish new benchmark in rotated object detection, with implications for a wide range of applications requiring precise and reliable object localization irrespective of orientation.

[245] VIPAMIN: Visual Prompt Initialization via Embedding Selection and Subspace Expansion

Jaekyun Park, Hye Won Chung

Main category: cs.CV

TL;DR: VIPAMIN is a visual prompt initialization method that enhances adaptation of self-supervised models by aligning prompts with informative regions and injecting new representational directions, achieving state-of-the-art performance with minimal computational overhead.

DetailsMotivation: Existing visual prompt tuning methods often fail to specialize prompts or enrich representation space, especially with self-supervised backbones, which becomes critical in challenging tasks and data-scarce settings where effective adaptation is needed.

Method: VIPAMIN enhances adaptation by (1) aligning prompts with semantically informative regions in embedding space, and (2) injecting novel representational directions beyond the pretrained subspace, requiring only a single forward pass and lightweight operations.

Result: VIPAMIN consistently improves performance across diverse tasks and dataset sizes, setting a new state of the art in visual prompt tuning.

Conclusion: VIPAMIN provides an effective and efficient visual prompt initialization strategy that significantly enhances adaptation of self-supervised models while maintaining computational efficiency.

Abstract: In the era of large-scale foundation models, fully fine-tuning pretrained networks for each downstream task is often prohibitively resource-intensive. Prompt tuning offers a lightweight alternative by introducing tunable prompts while keeping the backbone frozen. However, existing visual prompt tuning methods often fail to specialize the prompts or enrich the representation space–especially when applied to self-supervised backbones. We show that these limitations become especially pronounced in challenging tasks and data-scarce settings, where effective adaptation is most critical. In this work, we introduce VIPAMIN, a visual prompt initialization strategy that enhances adaptation of self-supervised models by (1) aligning prompts with semantically informative regions in the embedding space, and (2) injecting novel representational directions beyond the pretrained subspace. Despite its simplicity–requiring only a single forward pass and lightweight operations–VIPAMIN consistently improves performance across diverse tasks and dataset sizes, setting a new state of the art in visual prompt tuning. Our code is available at https://github.com/iamjaekyun/vipamin.

[246] Instance-Aware Pseudo-Labeling and Class-Focused Contrastive Learning for Weakly Supervised Domain Adaptive Segmentation of Electron Microscopy

Shan Xiong, Jiabao Chen, Ye Wang, Jialin Peng

Main category: cs.CV

TL;DR: This paper proposes a weakly supervised domain adaptation method for mitochondria segmentation in EM images using sparse point labels, featuring multitask learning with cross-teaching and instance-aware pseudo-label selection.

DetailsMotivation: To address the high annotation costs and domain shift issues in mitochondria segmentation from EM images, while achieving better performance than unsupervised domain adaptation methods with minimal annotation effort.

Method: Multitask learning framework combining segmentation and center detection with cross-teaching mechanism, class-focused cross-domain contrastive learning, and segmentation self-training with instance-aware pseudo-label selection strategy.

Result: Outperforms existing UDA and WDA methods, significantly narrowing the performance gap with supervised upper bound, and achieves substantial improvements over other UDA techniques.

Conclusion: The proposed weakly supervised domain adaptation approach effectively leverages sparse point annotations to achieve high-performance mitochondria segmentation with minimal annotation costs.

Abstract: Annotation-efficient segmentation of the numerous mitochondria instances from various electron microscopy (EM) images is highly valuable for biological and neuroscience research. Although unsupervised domain adaptation (UDA) methods can help mitigate domain shifts and reduce the high costs of annotating each domain, they typically have relatively low performance in practical applications. Thus, we investigate weakly supervised domain adaptation (WDA) that utilizes additional sparse point labels on the target domain, which require minimal annotation effort and minimal expert knowledge. To take full use of the incomplete and imprecise point annotations, we introduce a multitask learning framework that jointly conducts segmentation and center detection with a novel cross-teaching mechanism and class-focused cross-domain contrastive learning. While leveraging unlabeled image regions is essential, we introduce segmentation self-training with a novel instance-aware pseudo-label (IPL) selection strategy. Unlike existing methods that typically rely on pixel-wise pseudo-label filtering, the IPL semantically selects reliable and diverse pseudo-labels with the help of the detection task. Comprehensive validations and comparisons on challenging datasets demonstrate that our method outperforms existing UDA and WDA methods, significantly narrowing the performance gap with the supervised upper bound. Furthermore, under the UDA setting, our method also achieves substantial improvements over other UDA techniques.

[247] NavQ: Learning a Q-Model for Foresighted Vision-and-Language Navigation

Peiran Xu, Xicheng Gong, Yadong MU

Main category: cs.CV

TL;DR: Proposes a foresighted VLN agent using Q-learning to predict future outcomes of actions, combining task-agnostic Q-features with navigation instructions for improved decision-making.

DetailsMotivation: Existing VLN methods focus on historical information and overlook future implications of actions, leading to suboptimal navigation decisions.

Method: Trains Q-model with unlabeled trajectory data to generate Q-features for candidate actions, integrates these with navigation instructions via cross-modal encoder, and uses A*-style search combining future and historical scores.

Result: Extensive experiments on goal-oriented VLN datasets validate the method’s effectiveness in improving navigation performance.

Conclusion: The proposed foresighted approach with Q-learning and future-aware decision-making significantly enhances VLN agent performance by considering long-term outcomes.

Abstract: In this work we concentrate on the task of goal-oriented Vision-and-Language Navigation (VLN). Existing methods often make decisions based on historical information, overlooking the future implications and long-term outcomes of the actions. In contrast, we aim to develop a foresighted agent. Specifically, we draw upon Q-learning to train a Q-model using large-scale unlabeled trajectory data, in order to learn the general knowledge regarding the layout and object relations within indoor scenes. This model can generate a Q-feature, analogous to the Q-value in traditional Q-network, for each candidate action, which describes the potential future information that may be observed after taking the specific action. Subsequently, a cross-modal future encoder integrates the task-agnostic Q-feature with navigation instructions to produce a set of action scores reflecting future prospects. These scores, when combined with the original scores based on history, facilitate an A*-style searching strategy to effectively explore the regions that are more likely to lead to the destination. Extensive experiments conducted on widely used goal-oriented VLN datasets validate the effectiveness of the proposed method.

[248] HGC-Avatar: Hierarchical Gaussian Compression for Streamable Dynamic 3D Avatars

Haocheng Tang, Ruoke Yan, Xinhui Yin, Qi Zhang, Xinfeng Zhang, Siwei Ma, Wen Gao, Chuanmin Jia

Main category: cs.CV

TL;DR: HGC-Avatar is a hierarchical Gaussian compression framework for efficient transmission and high-quality rendering of dynamic avatars, using structural and motion layers with facial attention to improve compression efficiency and visual quality.

DetailsMotivation: Current 3D Gaussian Splatting compression methods for digital humans lack human priors, leading to suboptimal bitrate efficiency and reconstruction quality, hindering their use in streamable 3D avatar systems.

Method: Disentangles Gaussian representation into structural layer (StyleUNet-based generator mapping poses to Gaussians) and motion layer (SMPL-X model for compact pose variations), with facial attention mechanism and layer-wise compression.

Result: Provides streamable solution for rapid 3D avatar rendering, significantly outperforming prior methods in both visual quality and compression efficiency.

Conclusion: HGC-Avatar enables efficient transmission and high-quality rendering of dynamic avatars with improved compression efficiency and facial realism.

Abstract: Recent advances in 3D Gaussian Splatting (3DGS) have enabled fast, photorealistic rendering of dynamic 3D scenes, showing strong potential in immersive communication. However, in digital human encoding and transmission, the compression methods based on general 3DGS representations are limited by the lack of human priors, resulting in suboptimal bitrate efficiency and reconstruction quality at the decoder side, which hinders their application in streamable 3D avatar systems. We propose HGC-Avatar, a novel Hierarchical Gaussian Compression framework designed for efficient transmission and high-quality rendering of dynamic avatars. Our method disentangles the Gaussian representation into a structural layer, which maps poses to Gaussians via a StyleUNet-based generator, and a motion layer, which leverages the SMPL-X model to represent temporal pose variations compactly and semantically. This hierarchical design supports layer-wise compression, progressive decoding, and controllable rendering from diverse pose inputs such as video sequences or text. Since people are most concerned with facial realism, we incorporate a facial attention mechanism during StyleUNet training to preserve identity and expression details under low-bitrate constraints. Experimental results demonstrate that HGC-Avatar provides a streamable solution for rapid 3D avatar rendering, while significantly outperforming prior methods in both visual quality and compression efficiency.

[249] PRISMM-Bench: A Benchmark of Peer-Review Grounded Multimodal Inconsistencies

Lukas Selch, Yufang Hou, M. Jehanzeb Mirza, Sivan Doveh, James Glass, Rogerio Feris, Wei Lin

Main category: cs.CV

TL;DR: PRISMM-Bench is the first benchmark using real reviewer-flagged inconsistencies in scientific papers to evaluate Large Multimodal Models’ ability to detect and resolve cross-modal inconsistencies in text, figures, tables, and equations.

DetailsMotivation: Existing benchmarks overlook real-world multimodal inconsistencies in scientific papers, either isolating single modalities or using synthetic errors that fail to capture domain-specific complexity, undermining clarity, reproducibility, and trust.

Method: Created PRISMM-Bench through multi-stage pipeline: review mining, LLM-assisted filtering, and human verification to curate 262 inconsistencies from 242 papers. Designed three tasks (inconsistency identification, remedy, pair matching) and introduced structured JSON-based answer representations to minimize linguistic biases.

Result: Benchmarked 21 leading LMMs including large open-weight and proprietary models. Results show strikingly low performance (26.1-54.2%), highlighting the challenge of multimodal scientific reasoning.

Conclusion: Current LMMs struggle significantly with detecting and resolving real-world multimodal inconsistencies in scientific papers, motivating the need for progress towards more trustworthy scientific assistants.

Abstract: Large Multimodal Models (LMMs) are increasingly applied to scientific research, yet it remains unclear whether they can reliably understand and reason over the multimodal complexity of papers. A central challenge lies in detecting and resolving inconsistencies across text, figures, tables, and equations, issues that are often subtle, domain-specific, and ultimately undermine clarity, reproducibility, and trust. Existing benchmarks overlook this issue, either isolating single modalities or relying on synthetic errors that fail to capture real-world complexity. We introduce PRISMM-Bench (Peer-Review-sourced Inconsistency Set for Multimodal Models), the first benchmark grounded in real reviewer-flagged inconsistencies in scientific papers. Through a multi-stage pipeline of review mining, LLM-assisted filtering and human verification, we curate 262 inconsistencies from 242 papers. Based on this set, we design three tasks, namely inconsistency identification, remedy and pair matching, which assess a model’s capacity to detect, correct, and reason over inconsistencies across different modalities. Furthermore, to address the notorious problem of choice-only shortcuts in multiple-choice evaluation, where models exploit answer patterns without truly understanding the question, we further introduce structured JSON-based answer representations that minimize linguistic biases by reducing reliance on superficial stylistic cues. We benchmark 21 leading LMMs, including large open-weight models (GLM-4.5V 106B, InternVL3 78B) and proprietary models (Gemini 2.5 Pro, GPT-5 with high reasoning). Results reveal strikingly low performance (26.1-54.2%), underscoring the challenge of multimodal scientific reasoning and motivating progress towards trustworthy scientific assistants.

[250] OOS-DSD: Improving Out-of-stock Detection in Retail Images using Auxiliary Tasks

Franko Šikić, Sven Lončarić

Main category: cs.CV

TL;DR: OOS-DSD is a novel deep learning method that enhances out-of-stock detection using auxiliary learning with YOLOv8, adding branches for product segmentation and depth estimation to improve performance.

DetailsMotivation: Out-of-stock detection is crucial for retail verification, but existing methods need improvement in accurately detecting product unavailability on shelves.

Method: Extends YOLOv8 with additional convolutional branches for OOS detection, product segmentation, and depth estimation. Uses pseudo-labeled depth data from Depth Anything V2 with a novel depth normalization procedure.

Result: Achieved 1.8% higher mAP than state-of-the-art OOS detection methods. Ablation studies showed auxiliary learning increased mAP by 3.7% and depth normalization by 4.2%.

Conclusion: The proposed OOS-DSD method effectively improves OOS detection through auxiliary learning and depth normalization, demonstrating significant performance gains over existing approaches.

Abstract: Out-of-stock (OOS) detection is a very important retail verification process that aims to infer the unavailability of products in their designated areas on the shelf. In this paper, we introduce OOS-DSD, a novel deep learning-based method that advances OOS detection through auxiliary learning. In particular, we extend a well-established YOLOv8 object detection architecture with additional convolutional branches to simultaneously detect OOS, segment products, and estimate scene depth. While OOS detection and product segmentation branches are trained using ground truth data, the depth estimation branch is trained using pseudo-labeled annotations produced by the state-of-the-art (SOTA) depth estimation model Depth Anything V2. Furthermore, since the aforementioned pseudo-labeled depth estimates display relative depth, we propose an appropriate depth normalization procedure that stabilizes the training process. The experimental results show that the proposed method surpassed the performance of the SOTA OOS detection methods by 1.8% of the mean average precision (mAP). In addition, ablation studies confirm the effectiveness of auxiliary learning and the proposed depth normalization procedure, with the former increasing mAP by 3.7% and the latter by 4.2%.

[251] Image Categorization and Search via a GAT Autoencoder and Representative Models

Duygu Sap, Martin Lotz, Connor Mattinson

Main category: cs.CV

TL;DR: A representative-centric image categorization and retrieval method using graph attention network (GAT)-based autoencoder to construct context-aware latent representations and category representatives for efficient image comparison.

DetailsMotivation: To develop an effective image categorization and retrieval approach that leverages representative models for images and categories, enabling more accurate and context-aware comparisons through graph-based relationships.

Method: Utilizes a graph structure where nodes represent images/representatives and edges capture similarity relationships. Employs GAT-based autoencoder to highlight important features and construct context-aware latent representations, then obtains category representatives from embeddings for categorization and retrieval.

Result: The method demonstrates effectiveness through experiments comparing GAT autoencoders with standard feature-based techniques, showing improved performance in representative-centric image categorization and retrieval.

Conclusion: The proposed representative-centric approach using GAT-based autoencoders provides an effective framework for image categorization and retrieval by leveraging graph-based relationships and context-aware representations.

Abstract: We propose a method for image categorization and retrieval that leverages graphs and a graph attention network (GAT)-based autoencoder. Our approach is representative-centric, that is, we execute the categorization and retrieval process via the representative models we construct for the images and image categories. We utilize a graph where nodes represent images (or their representatives) and edges capture similarity relationships. GAT highlights important features and relationships between images, enabling the autoencoder to construct context-aware latent representations that capture the key features of each image relative to its neighbors. We obtain category representatives from these embeddings and categorize a query image by comparing its representative to the category representatives. We then retrieve the most similar image to the query image within its identified category. We demonstrate the effectiveness of our representative-centric approach through experiments with both the GAT autoencoders and standard feature-based techniques.

[252] Enhancing Compositional Reasoning in CLIP via Reconstruction and Alignment of Text Descriptions

Jihoon Kwon, Kyle Min, Jy-yong Sohn

Main category: cs.CV

TL;DR: READ is a fine-tuning method that enhances compositional reasoning in vision-language models by adding token-level reconstruction and sentence-level alignment objectives to contrastive learning, achieving state-of-the-art performance on compositional reasoning benchmarks.

DetailsMotivation: Standard contrastive training causes text encoders to focus on individual words rather than their relations, limiting compositional reasoning abilities in vision-language models.

Method: READ adds two auxiliary objectives to contrastive learning: (1) token-level reconstruction using a frozen decoder to reconstruct alternative captions, and (2) sentence-level alignment to explicitly align paraphrased sentences in embedding space.

Result: READ-CLIP achieves state-of-the-art performance across five major compositional reasoning benchmarks, outperforming the strongest baseline by up to 4.1%. The method also improves existing CLIP variants like NegCLIP and FSC-CLIP.

Conclusion: The reconstruction and alignment objectives provide complementary benefits - reconstruction captures word relationships within captions, while alignment ensures consistent representations for paraphrases with different wording.

Abstract: Despite recent advances, vision-language models trained with standard contrastive objectives still struggle with compositional reasoning – the ability to understand structured relationships between visual and linguistic elements. This shortcoming is largely due to the tendency of the text encoder to focus on individual words rather than their relations, a limitation reinforced by contrastive training that primarily aligns words with visual objects. In this paper, we introduce REconstruction and Alignment of text Descriptions (READ), a fine-tuning method designed to enhance compositional reasoning by adding two auxiliary objectives to the contrastive learning: (1) a token-level reconstruction objective, where a frozen pre-trained decoder reconstructs alternative captions based on the embedding of the original caption; and (2) a sentence-level alignment objective, which explicitly aligns paraphrased sentences in the embedding space. We show that READ-CLIP, a model derived by applying the READ method to the pre-trained CLIP model, achieves the state-of-the-art performance across five major compositional reasoning benchmarks, outperforming the strongest conventional fine-tuning baseline by up to 4.1%. Furthermore, applying the READ to existing CLIP variants (including NegCLIP and FSC-CLIP) also improves performance on these benchmarks. Quantitative and qualitative analyses reveal that our proposed objectives – reconstruction and alignment – offer complementary benefits: the former encourages the encoder to capture relationships between words within a caption, while the latter ensures consistent representations for paraphrases expressed with different wording.

[253] Watch Where You Move: Region-aware Dynamic Aggregation and Excitation for Gait Recognition

Binyuan Huang, Yongdong Luo, Xianda Guo, Xiawu Zheng, Zheng Zhu, Jiahui Pan, Chengju Zhou

Main category: cs.CV

TL;DR: GaitRDAE is a novel framework for gait recognition that dynamically adapts to motion regions with varying temporal scales and applies region-specific attention, outperforming existing methods on benchmark datasets.

DetailsMotivation: Existing gait recognition methods use predefined regions with fixed temporal scales, which struggle to model dynamically changing motion regions and adapt to their specific patterns, especially when covariates affect visual appearance.

Method: Proposes Region-aware Dynamic Aggregation and Excitation framework (GaitRDAE) with two core modules: RDA for dynamically searching optimal temporal receptive fields per region, and RDE for emphasizing motion regions with stable behavior patterns while suppressing static regions affected by covariates.

Result: GaitRDAE achieves state-of-the-art performance on several benchmark datasets.

Conclusion: The framework successfully addresses limitations of fixed temporal modeling by dynamically adapting to motion regions and their specific temporal patterns, leading to improved gait recognition accuracy.

Abstract: Deep learning-based gait recognition has achieved great success in various applications. The key to accurate gait recognition lies in considering the unique and diverse behavior patterns in different motion regions, especially when covariates affect visual appearance. However, existing methods typically use predefined regions for temporal modeling, with fixed or equivalent temporal scales assigned to different types of regions, which makes it difficult to model motion regions that change dynamically over time and adapt to their specific patterns. To tackle this problem, we introduce a Region-aware Dynamic Aggregation and Excitation framework (GaitRDAE) that automatically searches for motion regions, assigns adaptive temporal scales and applies corresponding attention. Specifically, the framework includes two core modules: the Region-aware Dynamic Aggregation (RDA) module, which dynamically searches the optimal temporal receptive field for each region, and the Region-aware Dynamic Excitation (RDE) module, which emphasizes the learning of motion regions containing more stable behavior patterns while suppressing attention to static regions that are more susceptible to covariates. Experimental results show that GaitRDAE achieves state-of-the-art performance on several benchmark datasets.

[254] Fit for Purpose? Deepfake Detection in the Real World

Guangyu Lin, Li Lin, Christina P. Walker, Daniel S. Schiff, Shu Hu

Main category: cs.CV

TL;DR: This paper introduces the first systematic benchmark using real-world political deepfakes from social media to evaluate deepfake detectors, finding that current models struggle with generalization and are vulnerable to simple manipulations.

DetailsMotivation: The proliferation of AI-generated content, especially political deepfakes, poses risks to truth and institutional trust. Current detection models are trained on synthetic datasets, limiting their effectiveness against real-world political deepfakes circulating on social media.

Method: Created a benchmark based on the Political Deepfakes Incident Database - a curated collection of real-world political deepfakes shared on social media since 2018. Systematically evaluated state-of-the-art deepfake detectors from academia, government, and industry.

Result: Academic and government detectors performed poorly. Paid tools achieved higher performance than free-access models, but all detectors struggled to generalize to authentic political deepfakes and were vulnerable to simple manipulations, especially in video.

Conclusion: There is an urgent need for politically contextualized deepfake detection frameworks to better protect the public from real-world political deepfake threats.

Abstract: The rapid proliferation of AI-generated content, driven by advances in generative adversarial networks, diffusion models, and multimodal large language models, has made the creation and dissemination of synthetic media effortless, heightening the risks of misinformation, particularly political deepfakes that distort truth and undermine trust in political institutions. In turn, governments, research institutions, and industry have strongly promoted deepfake detection initiatives as solutions. Yet, most existing models are trained and validated on synthetic, laboratory-controlled datasets, limiting their generalizability to the kinds of real-world political deepfakes circulating on social platforms that affect the public. In this work, we introduce the first systematic benchmark based on the Political Deepfakes Incident Database, a curated collection of real-world political deepfakes shared on social media since 2018. Our study includes a systematic evaluation of state-of-the-art deepfake detectors across academia, government, and industry. We find that the detectors from academia and government perform relatively poorly. While paid detection tools achieve relatively higher performance than free-access models, all evaluated detectors struggle to generalize effectively to authentic political deepfakes, and are vulnerable to simple manipulations, especially in the video domain. Results urge the need for politically contextualized deepfake detection frameworks to better safeguard the public in real-world settings.

[255] SHIELD: Suppressing Hallucinations In LVLM Encoders via Bias and Vulnerability Defense

Yiyang Huang, Liang Shi, Yitian Zhang, Yi Xu, Yun Fu

Main category: cs.CV

TL;DR: SHIELD is a training-free framework that addresses object hallucination in Large Vision-Language Models by targeting visual encoder issues through three strategies: re-weighting visual tokens, introducing noise-derived tokens, and applying adversarial attacks with contrastive decoding.

DetailsMotivation: Object hallucination in LVLMs remains a significant challenge where models produce plausible but inaccurate object descriptions. Previous work focused on LLM components, but this paper identifies visual encoders as the primary source of hallucinations.

Method: Proposes SHIELD framework with three training-free strategies: 1) re-weighting visual tokens to reduce statistical bias, 2) introducing noise-derived tokens to counter inherent bias, and 3) applying adversarial attacks with contrastive decoding to address vulnerability.

Result: SHIELD effectively mitigates object hallucinations across diverse benchmarks and LVLM families. It also achieves strong performance on general LVLM benchmarks, demonstrating broad applicability.

Conclusion: SHIELD successfully addresses object hallucination in LVLMs by targeting visual encoder issues rather than LLM components, providing an effective training-free solution with wide applicability across different model families.

Abstract: Large Vision-Language Models (LVLMs) excel in diverse cross-modal tasks. However, object hallucination, where models produce plausible but inaccurate object descriptions, remains a significant challenge. In contrast to previous work focusing on LLM components, this paper is the first to trace LVLM hallucinations to visual encoders and identifies three key issues: statistical bias, inherent bias, and vulnerability. To address these challenges, we propose SHIELD, a training-free framework that mitigates hallucinations through three strategies: re-weighting visual tokens to reduce statistical bias, introducing noise-derived tokens to counter inherent bias, and applying adversarial attacks with contrastive decoding to address vulnerability. Experiments demonstrate that SHIELD effectively mitigates object hallucinations across diverse benchmarks and LVLM families. Moreover, SHIELD achieves strong performance on the general LVLM benchmark, highlighting its broad applicability. Code will be released.

[256] VisionSelector: End-to-End Learnable Visual Token Compression for Efficient Multimodal LLMs

Jiaying Zhu, Yurui Zhu, Xin Lu, Wenrui Yan, Dong Li, Kunlin Liu, Xueyang Fu, Zheng-Jun Zha

Main category: cs.CV

TL;DR: VisionSelector is a lightweight plug-and-play framework for adaptive token compression in MLLMs that uses a differentiable Top-K mechanism and curriculum annealing to efficiently select critical visual tokens while maintaining performance.

DetailsMotivation: MLLMs face computational bottlenecks from massive visual tokens in high-resolution/multi-image inputs. Existing compression methods risk information loss and suffer from biases like attention sinks, leading to performance drops under aggressive compression.

Method: Reformulates token compression as end-to-end learnable decision process with VisionSelector - a scorer module decoupled from MLLM backbone, incorporating differentiable Top-K mechanism and curriculum annealing strategy to bridge training-inference gap.

Result: Achieves 100% accuracy on MME with 30% retention budget, outperforms prior methods by 12.14% at 10% retention budget, doubles prefill speed, with only 12.85M trainable parameters. Demonstrates generalization across various compression rates.

Conclusion: VisionSelector provides efficient and adaptive token selection for MLLMs, preserving critical information while significantly reducing computational overhead across various compression budgets.

Abstract: Multimodal Large Language Models (MLLMs) encounter significant computational and memory bottlenecks from the massive number of visual tokens generated by high-resolution images or multi-image inputs. Previous token compression techniques are often constrained by heuristic rules that risk discarding critical information. They may suffer from biases, such as attention sinks, that lead to sharp performance drops under aggressive compression ratios. To address these limitations, we reformulate token compression as a lightweight plug-and-play framework that reformulates token compression into an end-to-end learnable decision process. To be specific, we propose VisionSelector, a scorer module decoupled from the MLLM backbone that incorporates a differentiable Top-K mechanism and a curriculum annealing strategy to bridge the training-inference gap, enabling efficient and adaptive token selection various arbitrary compression rates. Remarkably lightweight with only 12.85M trainable parameters, VisionSelector demonstrates generalization across various compression rates and adaptively identifying critical tokens. This leads to superior performance across all compression budgets, evidenced by preserving 100% accuracy on MME with 30% retention budget, outperforming prior methods by 12.14% at 10% retention budget, and doubling prefill speed. Our code is available at https://github.com/JulietChoo/VisionSelector .

[257] A Deep Learning Framework for Real-Time Image Processing in Medical Diagnostics: Enhancing Accuracy and Speed in Clinical Applications

Melika Filvantorkaman, Maral Filvan Torkaman

Main category: cs.CV

TL;DR: A deep learning framework for real-time medical image analysis that integrates U-Net, EfficientNet, and Transformer models with optimization techniques to achieve high accuracy and fast inference for X-ray, CT, and MRI diagnostics.

DetailsMotivation: To address the limitations of traditional image processing in medical diagnostics, including time-consuming interpretation, clinician variability, and lack of precision/robustness for real-time clinical use.

Method: Integrates U-Net, EfficientNet, and Transformer-based neural networks with real-time optimization strategies (model pruning, quantization, GPU acceleration) and enables flexible deployment across edge devices, servers, and cloud infrastructure.

Result: Achieved state-of-the-art performance: >92% classification accuracy, >91% segmentation Dice scores, and <80ms inference times on public benchmark datasets, with enhanced transparency through Grad-CAM and segmentation overlays.

Conclusion: The framework can substantially accelerate diagnostic workflows, reduce clinician workload, and support trustworthy AI integration in time-critical healthcare environments.

Abstract: Medical imaging plays a vital role in modern diagnostics; however, interpreting high-resolution radiological data remains time-consuming and susceptible to variability among clinicians. Traditional image processing techniques often lack the precision, robustness, and speed required for real-time clinical use. To overcome these limitations, this paper introduces a deep learning framework for real-time medical image analysis designed to enhance diagnostic accuracy and computational efficiency across multiple imaging modalities, including X-ray, CT, and MRI. The proposed system integrates advanced neural network architectures such as U-Net, EfficientNet, and Transformer-based models with real-time optimization strategies including model pruning, quantization, and GPU acceleration. The framework enables flexible deployment on edge devices, local servers, and cloud infrastructures, ensuring seamless interoperability with clinical systems such as PACS and EHR. Experimental evaluations on public benchmark datasets demonstrate state-of-the-art performance, achieving classification accuracies above 92%, segmentation Dice scores exceeding 91%, and inference times below 80 milliseconds. Furthermore, visual explanation tools such as Grad-CAM and segmentation overlays enhance transparency and clinical interpretability. These results indicate that the proposed framework can substantially accelerate diagnostic workflows, reduce clinician workload, and support trustworthy AI integration in time-critical healthcare environments.

[258] Self-Supervised Learning to Fly using Efficient Semantic Segmentation and Metric Depth Estimation for Low-Cost Autonomous UAVs

Sebastian Mocanu, Emil Slusanschi, Marius Leordeanu

Main category: cs.CV

TL;DR: Vision-only autonomous flight system for small UAVs using semantic segmentation and monocular depth estimation for obstacle avoidance and safe landing in GPS-denied indoor environments.

DetailsMotivation: Enable autonomous drone navigation without GPS or expensive sensors like LiDAR by using vision-only approaches suitable for resource-constrained platforms.

Method: Combines semantic segmentation with monocular depth estimation using adaptive scale factor algorithm for metric distance conversion. Uses knowledge distillation with SVM teacher to train lightweight U-Net student network for real-time segmentation.

Result: Achieved 14.4 cm mean distance error, 100% success rate in 30 real-world and 100 digital-twin flight tests. End-to-end learning achieved 87.5% autonomous mission success rate.

Conclusion: The system advances practical vision-based drone navigation in structured environments, solving metric depth estimation and computational efficiency challenges for deployment on resource-constrained platforms.

Abstract: This paper presents a vision-only autonomous flight system for small UAVs operating in controlled indoor environments. The system combines semantic segmentation with monocular depth estimation to enable obstacle avoidance, scene exploration, and autonomous safe landing operations without requiring GPS or expensive sensors such as LiDAR. A key innovation is an adaptive scale factor algorithm that converts non-metric monocular depth predictions into accurate metric distance measurements by leveraging semantic ground plane detection and camera intrinsic parameters, achieving a mean distance error of 14.4 cm. The approach uses a knowledge distillation framework where a color-based Support Vector Machine (SVM) teacher generates training data for a lightweight U-Net student network (1.6M parameters) capable of real-time semantic segmentation. For more complex environments, the SVM teacher can be replaced with a state-of-the-art segmentation model. Testing was conducted in a controlled 5x4 meter laboratory environment with eight cardboard obstacles simulating urban structures. Extensive validation across 30 flight tests in a real-world environment and 100 flight tests in a digital-twin environment demonstrates that the combined segmentation and depth approach increases the distance traveled during surveillance and reduces mission time while maintaining 100% success rates. The system is further optimized through end-to-end learning, where a compact student neural network learns complete flight policies from demonstration data generated by our best-performing method, achieving an 87.5% autonomous mission success rate. This work advances practical vision-based drone navigation in structured environments, demonstrating solutions for metric depth estimation and computational efficiency challenges that enable deployment on resource-constrained platforms.

[259] MultiVerse: A Multi-Turn Conversation Benchmark for Evaluating Large Vision and Language Models

Young-Jun Lee, Byung-Kwan Lee, Jianshu Zhang, Yechan Hwang, Byungsoo Ko, Han-Gyu Kim, Dongyu Yao, Xuankun Rong, Eojin Joo, Seung-Ho Han, Bowon Ko, Ho-Jin Choi

Main category: cs.CV

TL;DR: MultiVerse is a new multi-turn conversation benchmark with 647 dialogues from 12 VLM evaluation benchmarks, using GPT-4o as automated evaluator to assess 18 VLMs across 37 aspects.

DetailsMotivation: Real-world applications require complex multi-turn dialogues, but existing datasets only partially capture conversational scenarios. Current VLMs excel on single-turn benchmarks but struggle with multi-turn interactions.

Method: Created MultiVerse benchmark with 647 dialogues (4 turns average) from 12 diverse VLM benchmarks, covering 484 tasks across factual knowledge, perception, reasoning, math, and coding. Uses checklist-based evaluation with GPT-4o measuring 37 key aspects.

Result: Even strongest models like GPT-4o achieve only 50% success rate in complex multi-turn conversations. Providing full dialogue context significantly improves performance for smaller/weaker models, highlighting importance of in-context learning.

Conclusion: MultiVerse serves as a comprehensive landscape for evaluating multi-turn interaction abilities in VLMs, revealing significant challenges in complex conversational scenarios that current models struggle with.

Abstract: Vision-and-Language Models (VLMs) have shown impressive capabilities on single-turn benchmarks, yet real-world applications often demand more intricate multi-turn dialogues. Existing multi-turn datasets (e.g, MMDU, ConvBench) only partially capture the breadth and depth of conversational scenarios encountered by users. In this work, we introduce MultiVerse, a novel multi-turn conversation benchmark featuring 647 dialogues - each averaging four turns - derived from a diverse set of 12 popular VLM evaluation benchmarks. With 484 tasks and 484 interaction goals, MultiVerse covers a wide range of topics, from factual knowledge and perception to advanced reasoning tasks such as mathematics and coding. To facilitate robust assessment, we propose a checklist-based evaluation method that leverages GPT-4o as the automated evaluator, measuring performance across 37 key aspects, including perceptual accuracy, linguistic clarity, and factual correctness. We evaluate 18 VLMs on MultiVerse, revealing that even the strongest models (e.g., GPT-4o) achieve only a 50% success rate in complex multi-turn conversations, highlighting the dataset’s challenging nature. Notably, we find that providing full dialogue context significantly enhances performance for smaller or weaker models, emphasizing the importance of in-context learning. We believe MultiVerse is a landscape of evaluating multi-turn interaction abilities for VLMs.

[260] Structured Interfaces for Automated Reasoning with 3D Scene Graphs

Aaron Ray, Jacob Arkin, Harel Biggie, Chuchu Fan, Luca Carlone, Nicholas Roy

Main category: cs.CV

TL;DR: Using Retrieval Augmented Generation with Cypher query language to connect LLMs with 3D scene graphs for natural language grounding, enabling scalable processing of large scene graphs.

DetailsMotivation: Existing methods that encode entire 3D scene graphs as text in LLM context windows don't scale well to large or rich graphs, creating a need for more efficient grounding approaches.

Method: Proposed using Retrieval Augmented Generation with a graph database encoding of 3DSGs and providing Cypher query language as a tool for LLMs to retrieve relevant scene data for language grounding tasks.

Result: The Cypher interface approach scales significantly better to large, rich graphs on both local and cloud-based models, leading to large performance improvements in grounded language tasks while substantially reducing token count.

Conclusion: Using Cypher as an interface to 3D scene graphs provides an effective and scalable solution for connecting LLMs with complex world representations for natural language grounding.

Abstract: In order to provide a robot with the ability to understand and react to a user’s natural language inputs, the natural language must be connected to the robot’s underlying representations of the world. Recently, large language models (LLMs) and 3D scene graphs (3DSGs) have become a popular choice for grounding natural language and representing the world. In this work, we address the challenge of using LLMs with 3DSGs to ground natural language. Existing methods encode the scene graph as serialized text within the LLM’s context window, but this encoding does not scale to large or rich 3DSGs. Instead, we propose to use a form of Retrieval Augmented Generation to select a subset of the 3DSG relevant to the task. We encode a 3DSG in a graph database and provide a query language interface (Cypher) as a tool to the LLM with which it can retrieve relevant data for language grounding. We evaluate our approach on instruction following and scene question-answering tasks and compare against baseline context window and code generation methods. Our results show that using Cypher as an interface to 3D scene graphs scales significantly better to large, rich graphs on both local and cloud-based models. This leads to large performance improvements in grounded language tasks while also substantially reducing the token count of the scene graph content. A video supplement is available at https://www.youtube.com/watch?v=zY_YI9giZSA.

[261] Universal and Transferable Attacks on Pathology Foundation Models

Yuntian Wang, Xilin Yang, Che-Yung Shen, Nir Pillar, Aydogan Ozcan

Main category: cs.CV

TL;DR: UTAP is a universal and transferable adversarial attack method that uses fixed weak noise patterns to systematically disrupt pathology foundation models’ feature representations, causing performance drops across various downstream tasks and unseen data distributions.

DetailsMotivation: To reveal critical vulnerabilities in pathology foundation models and establish a high-standard benchmark for model robustness evaluation, highlighting the need for improved defense mechanisms for safe AI deployment in pathology.

Method: Optimized using deep learning, UTAP generates fixed weak noise patterns that are added to pathology images to disrupt feature representations. The method demonstrates universality (works across diverse field-of-views) and transferability (affects various external black-box models).

Result: UTAP causes significant performance drops across various state-of-the-art pathology foundation models on multiple datasets with visually imperceptible modifications. It successfully degrades performance of external black-box models never seen during development.

Conclusion: UTAP constitutes a broad threat to emerging pathology foundation models and their applications, establishing a critical benchmark for robustness evaluation and highlighting the need for advancing defense mechanisms and adversarial training for reliable AI deployment in pathology.

Abstract: We introduce Universal and Transferable Adversarial Perturbations (UTAP) for pathology foundation models that reveal critical vulnerabilities in their capabilities. Optimized using deep learning, UTAP comprises a fixed and weak noise pattern that, when added to a pathology image, systematically disrupts the feature representation capabilities of multiple pathology foundation models. Therefore, UTAP induces performance drops in downstream tasks that utilize foundation models, including misclassification across a wide range of unseen data distributions. In addition to compromising the model performance, we demonstrate two key features of UTAP: (1) universality: its perturbation can be applied across diverse field-of-views independent of the dataset that UTAP was developed on, and (2) transferability: its perturbation can successfully degrade the performance of various external, black-box pathology foundation models - never seen before. These two features indicate that UTAP is not a dedicated attack associated with a specific foundation model or image dataset, but rather constitutes a broad threat to various emerging pathology foundation models and their applications. We systematically evaluated UTAP across various state-of-the-art pathology foundation models on multiple datasets, causing a significant drop in their performance with visually imperceptible modifications to the input images using a fixed noise pattern. The development of these potent attacks establishes a critical, high-standard benchmark for model robustness evaluation, highlighting a need for advancing defense mechanisms and potentially providing the necessary assets for adversarial training to ensure the safe and reliable deployment of AI in pathology.

[262] HYDRA: HYbrid knowledge Distillation and spectral Reconstruction Algorithm for high channel hyperspectral camera applications

Christopher Thirgood, Oscar Mendez, Erin Ling, Jon Storey, Simon Hadfield

Main category: cs.CV

TL;DR: HYDRA introduces a hybrid knowledge distillation approach for spectral reconstruction, achieving SOTA performance with 18% accuracy improvement and faster inference than existing methods.

DetailsMotivation: Address limitations of previous Multi-Scale Attention methods that only work well with sparse spectra, while modern hyperspectral sensors have hundreds of channels.

Method: Uses Teacher-Student architecture with knowledge distillation - Teacher encodes latent hyperspectral data, Student learns mappings from natural images to Teacher’s encoded domain with novel training method.

Result: Achieves high-quality spectral reconstruction with SOTA performance across all metrics, including 18% accuracy boost and faster inference times at various channel depths.

Conclusion: HYDRA successfully addresses key limitations of prior spectral reconstruction models and provides superior performance for modern hyperspectral imaging applications.

Abstract: Hyperspectral images (HSI) promise to support a range of new applications in computer vision. Recent research has explored the feasibility of generalizable Spectral Reconstruction (SR), the problem of recovering a HSI from a natural three-channel color image in unseen scenarios. However, previous Multi-Scale Attention (MSA) works have only demonstrated sufficient generalizable results for very sparse spectra, while modern HSI sensors contain hundreds of channels. This paper introduces a novel approach to spectral reconstruction via our HYbrid knowledge Distillation and spectral Reconstruction Architecture (HYDRA). Using a Teacher model that encapsulates latent hyperspectral image data and a Student model that learns mappings from natural images to the Teacher’s encoded domain, alongside a novel training method, we achieve high-quality spectral reconstruction. This addresses key limitations of prior SR models, providing SOTA performance across all metrics, including an 18% boost in accuracy, and faster inference times than current SOTA models at various channel depths.

[263] Pursuing Minimal Sufficiency in Spatial Reasoning

Yejie Guo, Yunzhong Hou, Wufei Ma, Meng Tang, Ming-Hsuan Yang

Main category: cs.CV

TL;DR: MSSR is a dual-agent framework that improves spatial reasoning in VLMs by first extracting sufficient 3D information using expert models, then iteratively refining it to achieve minimality through redundant detail pruning and missing information requests.

DetailsMotivation: Address two fundamental bottlenecks in spatial reasoning for VLMs: inadequate 3D understanding from 2D-centric pre-training, and reasoning failures caused by redundant 3D information.

Method: Dual-agent framework with Perception Agent that programmatically queries 3D scenes using perception toolbox (including novel SOG module for direction grounding), and Reasoning Agent that iteratively refines information to achieve minimal sufficiency through closed-loop pruning and requests.

Result: Significantly improves accuracy and achieves state-of-the-art performance across two challenging benchmarks, while producing interpretable reasoning paths.

Conclusion: The explicit pursuit of both sufficiency and minimality in 3D information processing effectively addresses spatial reasoning challenges in VLMs, offering a promising approach for generating high-quality training data.

Abstract: Spatial reasoning, the ability to ground language in 3D understanding, remains a persistent challenge for Vision-Language Models (VLMs). We identify two fundamental bottlenecks: inadequate 3D understanding capabilities stemming from 2D-centric pre-training, and reasoning failures induced by redundant 3D information. To address these, we first construct a Minimal Sufficient Set (MSS) of information before answering a given question: a compact selection of 3D perception results from \textit{expert models}. We introduce MSSR (Minimal Sufficient Spatial Reasoner), a dual-agent framework that implements this principle. A Perception Agent programmatically queries 3D scenes using a versatile perception toolbox to extract sufficient information, including a novel SOG (Situated Orientation Grounding) module that robustly extracts language-grounded directions. A Reasoning Agent then iteratively refines this information to pursue minimality, pruning redundant details and requesting missing ones in a closed loop until the MSS is curated. Extensive experiments demonstrate that our method, by explicitly pursuing both sufficiency and minimality, significantly improves accuracy and achieves state-of-the-art performance across two challenging benchmarks. Furthermore, our framework produces interpretable reasoning paths, offering a promising source of high-quality training data for future models. Source code is available at https://github.com/gyj155/mssr.

[264] SDPA++: A General Framework for Self-Supervised Denoising with Patch Aggregation

Huy Minh Nhat Nguyen, Triet Hoang Minh Dao, Chau Vinh Hoang Truong, Cuong Tuan Nguyen

Main category: cs.CV

TL;DR: SDPA++ is a self-supervised denoising framework for OCT images that uses only noisy images to generate pseudo-ground-truth through self-fusion and trains ensemble models via patch aggregation, achieving improved image quality without clean reference data.

DetailsMotivation: OCT imaging is crucial for retinal disease diagnosis but suffers from intrinsic speckle noise. Acquiring paired clean/noisy datasets for supervised denoising is challenging in clinical settings due to practical constraints.

Method: Proposes SDPA++ framework that leverages only noisy OCT images to generate pseudo-ground-truth through self-fusion and self-supervised denoising. Uses these refined images to train an ensemble of denoising models with patch-based strategy for enhanced clarity.

Result: Validated on IEEE SPS VIP Cup dataset containing only real-world noisy OCT images. Shows performance improvements in Contrast-to-Noise Ratio (CNR), Mean Square Ratio (MSR), Texture Preservation (TP), and Edge Preservation (EP) metrics.

Conclusion: The method demonstrates potential for improving OCT image quality and diagnostic outcomes in clinical practice without requiring clean reference images, addressing the challenge of limited paired datasets.

Abstract: Optical Coherence Tomography (OCT) is a widely used non-invasive imaging technique that provides detailed three-dimensional views of the retina, which are essential for the early and accurate diagnosis of ocular diseases. Consequently, OCT image analysis and processing have emerged as key research areas in biomedical imaging. However, acquiring paired datasets of clean and real-world noisy OCT images for supervised denoising models remains a formidable challenge due to intrinsic speckle noise and practical constraints in clinical imaging environments. To address these issues, we propose SDPA++: A General Framework for Self-Supervised Denoising with Patch Aggregation. Our novel approach leverages only noisy OCT images by first generating pseudo-ground-truth images through self-fusion and self-supervised denoising. These refined images then serve as targets to train an ensemble of denoising models using a patch-based strategy that effectively enhances image clarity. Performance improvements are validated via metrics such as Contrast-to-Noise Ratio (CNR), Mean Square Ratio (MSR), Texture Preservation (TP), and Edge Preservation (EP) on the real-world dataset from the IEEE SPS Video and Image Processing Cup. Notably, the VIP Cup dataset contains only real-world noisy OCT images without clean references, highlighting our method’s potential for improving image quality and diagnostic outcomes in clinical practice.

[265] Connecting Domains and Contrasting Samples: A Ladder for Domain Generalization

Tianxin Wei, Yifan Chen, Xinrui He, Wenxuan Bao, Jingrui He

Main category: cs.CV

TL;DR: Proposes Domain-Connecting Contrastive Learning (DCCL) to address domain generalization challenges by improving intra-class connectivity across domains through aggressive data augmentation, cross-domain positive samples, model anchoring, and generative transformation loss.

DetailsMotivation: Distribution shifts between training and testing samples impair model generalization. Direct application of contrastive learning (CL) deteriorates performance in domain generalization (DG) due to lack of intra-class connectivity across domains.

Method: DCCL enhances intra-class connectivity through: 1) aggressive data augmentation and cross-domain positive samples on data side, 2) model anchoring to exploit pre-trained representations and generative transformation loss on model side.

Result: Extensive experiments on five standard DG benchmarks show DCCL outperforms state-of-the-art baselines without requiring domain supervision.

Conclusion: DCCL successfully addresses the intra-class connectivity deficiency in DG settings and provides generalizable representations that improve domain generalization performance.

Abstract: Distribution shifts between training and testing samples frequently occur in practice and impede model generalization performance. This crucial challenge thereby motivates studies on domain generalization (DG), which aim to predict the label on unseen target domain data by solely using data from source domains. It is intuitive to conceive the class-separated representations learned in contrastive learning (CL) are able to improve DG, while the reality is quite the opposite: users observe directly applying CL deteriorates the performance. We analyze the phenomenon with the insights from CL theory and discover lack of intra-class connectivity in the DG setting causes the deficiency. We thus propose a new paradigm, domain-connecting contrastive learning (DCCL), to enhance the conceptual connectivity across domains and obtain generalizable representations for DG. On the data side, more aggressive data augmentation and cross-domain positive samples are introduced to improve intra-class connectivity. On the model side, to better embed the unseen test domains, we propose model anchoring to exploit the intra-class connectivity in pre-trained representations and complement the anchoring with generative transformation loss. Extensive experiments on five standard DG benchmarks are performed. The results verify that DCCL outperforms state-of-the-art baselines even without domain supervision. The detailed model implementation and the code are provided through https://github.com/weitianxin/DCCL

[266] HumanCM: One Step Human Motion Prediction

Liu Haojie, Gao Suixiang

Main category: cs.CV

TL;DR: HumanCM is a one-step human motion prediction framework using consistency models that achieves comparable accuracy to diffusion models with significantly faster inference.

DetailsMotivation: To overcome the inefficiency of multi-step denoising in diffusion-based motion prediction methods by developing a single-step generation approach.

Method: Uses consistency models to learn self-consistent mapping between noisy and clean motion states, with Transformer-based spatiotemporal architecture and temporal embeddings for long-range dependencies.

Result: Achieves comparable or superior accuracy to state-of-the-art diffusion models on Human3.6M and HumanEva-I datasets while reducing inference steps by up to two orders of magnitude.

Conclusion: HumanCM provides an efficient alternative to diffusion models for human motion prediction, maintaining high accuracy with dramatically faster inference through one-step generation.

Abstract: We present HumanCM, a one-step human motion prediction framework built upon consistency models. Instead of relying on multi-step denoising as in diffusion-based methods, HumanCM performs efficient single-step generation by learning a self-consistent mapping between noisy and clean motion states. The framework adopts a Transformer-based spatiotemporal architecture with temporal embeddings to model long-range dependencies and preserve motion coherence. Experiments on Human3.6M and HumanEva-I demonstrate that HumanCM achieves comparable or superior accuracy to state-of-the-art diffusion models while reducing inference steps by up to two orders of magnitude.

[267] Eliciting Grounded Chain-of-Thought Reasoning in 3D Scenes

Xiongkun Linghu, Jiangyong Huang, Ziyu Zhu, Baoxiong Jia, Siyuan Huang

Main category: cs.CV

TL;DR: A novel framework for 3D scene understanding using grounded Chain-of-Thought reasoning, featuring SCENECOT method and a large-scale dataset of 185K instances.

DetailsMotivation: Existing 3D LLMs struggle with grounded question-answering due to lack of human-like scene-object grounded reasoning mechanisms.

Method: Introduces SCENECOT - a grounded Chain-of-Thought reasoning method that decouples complex tasks into simpler problems using multimodal expert modules to build visual clues.

Result: Achieves strong performance across various 3D scene reasoning benchmarks with high grounding-QA coherence, demonstrating the first successful application of CoT reasoning to 3D scene understanding.

Conclusion: The framework enables step-by-step human-like reasoning in 3D scenes and shows potential for extension to broader 3D scene understanding scenarios.

Abstract: Existing research on 3D Large Language Models (LLMs) still struggles to achieve grounded question-answering, primarily due to the under-exploration of the mech- anism of human-like scene-object grounded reasoning. This paper bridges the gap by presenting a novel framework. We first introduce a grounded Chain-of- Thought reasoning method in 3D scenes (SCENECOT), decoupling a complex reasoning task into simpler and manageable problems, and building corresponding visual clues based on multimodal expert modules. To enable such a method, we develop SCENECOT-185K, the first large-scale grounded CoT reasoning dataset, consisting of 185K high-quality instances. Extensive experiments across various complex 3D scene reasoning benchmarks demonstrate that our new framework achieves strong performance with high grounding-QA coherence. To the best of our knowledge, this is the first successful application of CoT reasoning to 3D scene understanding, enabling step-by-step human-like reasoning and showing potential for extension to broader 3D scene understanding scenarios.

[268] Vision-Centric 4D Occupancy Forecasting and Planning via Implicit Residual World Models

Jianbiao Mei, Yu Yang, Xuemeng Yang, Licheng Wen, Jiajun Lv, Botian Shi, Yong Liu

Main category: cs.CV

TL;DR: IR-WM is an Implicit Residual World Model that focuses on modeling dynamic changes in autonomous driving scenes rather than full scene reconstruction, using BEV representations and residual prediction to improve efficiency and accuracy.

DetailsMotivation: Current vision-centric world models in autonomous driving inefficiently reconstruct entire future scenes, including static backgrounds, wasting computational capacity on redundant modeling.

Method: Uses BEV representation of current state, leverages previous timestep BEV features as temporal prior, predicts only residual changes conditioned on ego-vehicle actions and scene context, and applies alignment module to correct semantic/dynamic misalignments.

Result: Achieves top performance on nuScenes benchmark for both 4D occupancy forecasting and trajectory planning, with implicit future state generation substantially improving planning accuracy.

Conclusion: The proposed IR-WM approach effectively focuses on modeling world evolution rather than full reconstruction, demonstrating superior performance in autonomous driving tasks through efficient residual prediction and temporal modeling.

Abstract: End-to-end autonomous driving systems increasingly rely on vision-centric world models to understand and predict their environment. However, a common ineffectiveness in these models is the full reconstruction of future scenes, which expends significant capacity on redundantly modeling static backgrounds. To address this, we propose IR-WM, an Implicit Residual World Model that focuses on modeling the current state and evolution of the world. IR-WM first establishes a robust bird’s-eye-view representation of the current state from the visual observation. It then leverages the BEV features from the previous timestep as a strong temporal prior and predicts only the “residual”, i.e., the changes conditioned on the ego-vehicle’s actions and scene context. To alleviate error accumulation over time, we further apply an alignment module to calibrate semantic and dynamic misalignments. Moreover, we investigate different forecasting-planning coupling schemes and demonstrate that the implicit future state generated by world models substantially improves planning accuracy. On the nuScenes benchmark, IR-WM achieves top performance in both 4D occupancy forecasting and trajectory planning.

[269] UKANFormer: Noise-Robust Semantic Segmentation for Coral Reef Mapping via a Kolmogorov-Arnold Network-Transformer Hybrid

Tianyang Dou, Ming Li, Jiangying Qin, Xuan Liao, Jiageng Zhong, Armin Gruen, Mengyi Deng

Main category: cs.CV

TL;DR: UKANFormer is a semantic segmentation model that achieves high-precision coral reef mapping using noisy supervision from Allen Coral Atlas, outperforming baselines and producing more accurate predictions than the training labels.

DetailsMotivation: Global coral reef mapping products like Allen Coral Atlas have limited spatial precision and semantic consistency, especially for fine-grained boundary delineation, requiring improved methods for accurate large-scale conservation mapping.

Method: UKANFormer builds on UKAN architecture with a Global-Local Transformer (GL-Trans) block in the decoder to extract both global semantic structures and local boundary details, enabling high-precision mapping under noisy supervision.

Result: Achieved coral-class IoU of 67.00% and pixel accuracy of 83.98%, outperforming conventional baselines under the same noisy labels setting, and produced predictions more accurate than the training labels themselves.

Conclusion: Architectural design can mitigate label noise and support scalable mapping under imperfect supervision, challenging the notion that data quality directly limits model performance, providing foundation for ecological monitoring with scarce reliable labels.

Abstract: Coral reefs are vital yet fragile ecosystems that require accurate large-scale mapping for effective conservation. Although global products such as the Allen Coral Atlas provide unprecedented coverage of global coral reef distri-bution, their predictions are frequently limited in spatial precision and semantic consistency, especially in regions requiring fine-grained boundary delineation. To address these challenges, we propose UKANFormer, a novel se-mantic segmentation model designed to achieve high-precision mapping under noisy supervision derived from Allen Coral Atlas. Building upon the UKAN architecture, UKANFormer incorporates a Global-Local Transformer (GL-Trans) block in the decoder, enabling the extraction of both global semantic structures and local boundary details. In experiments, UKANFormer achieved a coral-class IoU of 67.00% and pixel accuracy of 83.98%, outperforming conventional baselines under the same noisy labels setting. Remarkably, the model produces predictions that are visually and structurally more accurate than the noisy labels used for training. These results challenge the notion that data quality directly limits model performance, showing that architectural design can mitigate label noise and sup-port scalable mapping under imperfect supervision. UKANFormer provides a foundation for ecological monitoring where reliable labels are scarce.

[270] A Comprehensive Survey on World Models for Embodied AI

Xinqing Li, Xin He, Le Zhang, Yun Liu

Main category: cs.CV

TL;DR: This survey presents a unified framework for world models in embodied AI, proposing a three-axis taxonomy covering functionality, temporal modeling, and spatial representation. It systematizes data resources and metrics, offers quantitative comparisons of state-of-the-art models, and identifies key open challenges.

DetailsMotivation: Embodied AI requires agents that can perceive, act, and anticipate how actions reshape future world states. World models serve as internal simulators that capture environment dynamics to support perception, prediction, and decision making.

Method: The paper formalizes the problem setting and learning objectives, and proposes a three-axis taxonomy: (1) Functionality: Decision-Coupled vs. General-Purpose; (2) Temporal Modeling: Sequential Simulation and Inference vs. Global Difference Prediction; (3) Spatial Representation: Global Latent Vector, Token Feature Sequence, Spatial Latent Grid, and Decomposed Rendering Representation.

Result: The survey systematizes data resources and metrics across robotics, autonomous driving, and general video settings, covering pixel prediction quality, state-level understanding, and task performance. It provides a quantitative comparison of state-of-the-art models.

Conclusion: Key open challenges include: scarcity of unified datasets, need for evaluation metrics that assess physical consistency over pixel fidelity, trade-off between model performance and computational efficiency for real-time control, and achieving long-horizon temporal consistency while mitigating error accumulation.

Abstract: Embodied AI requires agents that perceive, act, and anticipate how actions reshape future world states. World models serve as internal simulators that capture environment dynamics, enabling forward and counterfactual rollouts to support perception, prediction, and decision making. This survey presents a unified framework for world models in embodied AI. Specifically, we formalize the problem setting and learning objectives, and propose a three-axis taxonomy encompassing: (1) Functionality, Decision-Coupled vs. General-Purpose; (2) Temporal Modeling, Sequential Simulation and Inference vs. Global Difference Prediction; (3) Spatial Representation, Global Latent Vector, Token Feature Sequence, Spatial Latent Grid, and Decomposed Rendering Representation. We systematize data resources and metrics across robotics, autonomous driving, and general video settings, covering pixel prediction quality, state-level understanding, and task performance. Furthermore, we offer a quantitative comparison of state-of-the-art models and distill key open challenges, including the scarcity of unified datasets and the need for evaluation metrics that assess physical consistency over pixel fidelity, the trade-off between model performance and the computational efficiency required for real-time control, and the core modeling difficulty of achieving long-horizon temporal consistency while mitigating error accumulation. Finally, we maintain a curated bibliography at https://github.com/Li-Zn-H/AwesomeWorldModels.

[271] Visual Autoregressive Models Beat Diffusion Models on Inference Time Scaling

Erik Riise, Mehmet Onurcan Kaya, Dim P. Papadopoulos

Main category: cs.CV

TL;DR: Beam search significantly improves text-to-image generation in autoregressive models, enabling a 2B parameter model to outperform a 12B parameter diffusion model, highlighting the importance of discrete token spaces for effective inference-time optimization.

DetailsMotivation: While inference-time scaling through search has revolutionized LLMs, similar gains in image generation have been difficult to achieve, with recent attempts in diffusion models showing limited benefits compared to simple random sampling.

Method: Applied beam search to discrete, sequential visual autoregressive models, leveraging the discrete token space for early pruning and computational reuse in text-to-image generation.

Result: Beam search substantially improved text-to-image generation, with a 2B parameter autoregressive model outperforming a 12B parameter diffusion model across benchmarks. Systematic ablations confirmed the advantage comes from discrete token space properties.

Conclusion: Model architecture, not just scale, is critical for inference-time optimization in visual generation, with discrete autoregressive models enabling effective search strategies that continuous diffusion models cannot match.

Abstract: While inference-time scaling through search has revolutionized Large Language Models, translating these gains to image generation has proven difficult. Recent attempts to apply search strategies to continuous diffusion models show limited benefits, with simple random sampling often performing best. We demonstrate that the discrete, sequential nature of visual autoregressive models enables effective search for image generation. We show that beam search substantially improves text-to-image generation, enabling a 2B parameter autoregressive model to outperform a 12B parameter diffusion model across benchmarks. Systematic ablations show that this advantage comes from the discrete token space, which allows early pruning and computational reuse, and our verifier analysis highlights trade-offs between speed and reasoning capability. These findings suggest that model architecture, not just scale, is critical for inference-time optimization in visual generation.

[272] Prominence-Aware Artifact Detection and Dataset for Image Super-Resolution

Ivan Molodetskikh, Kirill Malyshev, Mark Mirgaleev, Nikita Zagainov, Evgeney Bogatyrev, Dmitriy Vatolin

Main category: cs.CV

TL;DR: This paper introduces a dataset of SR artifacts with prominence scores and trains a regressor to detect prominent artifacts, arguing that artifacts should be evaluated by their visual prominence rather than as binary defects.

DetailsMotivation: As SR models improve, they increasingly produce artifacts that vary in perceptual impact - some are barely noticeable while others strongly degrade image quality. Current methods treat artifacts as uniform binary defects rather than considering their varying prominence to human observers.

Method: Created a dataset of 1302 artifact examples from 11 SR methods with crowdsourced prominence scores. Trained a lightweight regressor to produce spatial prominence heatmaps for artifact detection.

Result: The trained regressor outperforms existing methods at detecting prominent artifacts and produces spatial prominence heatmaps that effectively identify visually disturbing artifacts.

Conclusion: The proposed prominence-aware approach provides better evaluation and mitigation of SR artifacts. The dataset and code are released to facilitate further research in this direction.

Abstract: Generative image super-resolution (SR) is rapidly advancing in visual quality and detail restoration. As the capacity of SR models expands, however, so does their tendency to produce artifacts: incorrect, visually disturbing details that reduce perceived quality. Crucially, their perceptual impact varies: some artifacts are barely noticeable while others strongly degrade the image. We argue that artifacts should be characterized by their prominence to human observers rather than treated as uniform binary defects. Motivated by this, we present a novel dataset of 1302 artifact examples from 11 contemporary image-SR methods, where each artifact is paired with a crowdsourced prominence score. Building on this dataset, we train a lightweight regressor that produces spatial prominence heatmaps and outperforms existing methods at detecting prominent artifacts. We release the dataset and code to facilitate prominence-aware evaluation and mitigation of SR artifacts.

[273] WaMaIR: Image Restoration via Multiscale Wavelet Convolutions and Mamba-based Channel Modeling with Texture Enhancement

Shengyu Zhu, Fan, Fuxuan Zhang

Main category: cs.CV

TL;DR: WaMaIR is a CNN-based image restoration framework that uses wavelet transforms and Mamba-based modules to improve texture detail reconstruction while maintaining computational efficiency.

DetailsMotivation: Previous CNN-based image restoration methods struggle with restoring fine texture details due to limited receptive fields and lack of channel feature modeling.

Method: Proposes three key components: Global Multiscale Wavelet Transform Convolutions (GMWTConvs) to expand receptive field, Mamba-Based Channel-Aware Module (MCAM) to capture long-range dependencies, and Multiscale Texture Enhancement Loss (MTELoss) to preserve texture structures.

Result: Extensive experiments show WaMaIR outperforms state-of-the-art methods in image restoration quality while maintaining efficient computational performance.

Conclusion: WaMaIR effectively addresses texture detail restoration challenges in image restoration through its novel architecture combining wavelet transforms, channel-aware modeling, and specialized loss functions.

Abstract: Image restoration is a fundamental and challenging task in computer vision, where CNN-based frameworks demonstrate significant computational efficiency. However, previous CNN-based methods often face challenges in adequately restoring fine texture details, which are limited by the small receptive field of CNN structures and the lack of channel feature modeling. In this paper, we propose WaMaIR, which is a novel framework with a large receptive field for image perception and improves the reconstruction of texture details in restored images. Specifically, we introduce the Global Multiscale Wavelet Transform Convolutions (GMWTConvs) for expandding the receptive field to extract image features, preserving and enriching texture features in model inputs. Meanwhile, we propose the Mamba-Based Channel-Aware Module (MCAM), explicitly designed to capture long-range dependencies within feature channels, which enhancing the model sensitivity to color, edges, and texture information. Additionally, we propose Multiscale Texture Enhancement Loss (MTELoss) for image restoration to guide the model in preserving detailed texture structures effectively. Extensive experiments confirm that WaMaIR outperforms state-of-the-art methods, achieving better image restoration and efficient computational performance of the model.

[274] Region in Context: Text-condition Image editing with Human-like semantic reasoning

Thuy Phuong Vu, Dinh-Cuong Hoang, Minhhuy Le, Phan Xuan Tan

Main category: cs.CV

TL;DR: Region in Context is a text-conditioned image editing framework that performs multilevel semantic alignment between vision and language to enable precise and harmonized image edits by understanding regions within global context.

DetailsMotivation: Current approaches treat image regions in isolation, relying only on local cues without considering how each part contributes to the overall composition, leading to inconsistent edits, unnatural transitions, and loss of coherence.

Method: Introduces a dual-level guidance mechanism: regions are represented with full-image context and aligned with detailed region-level descriptions, while the entire image is simultaneously matched to comprehensive scene-level descriptions generated by a large vision-language model.

Result: Experiments show the method produces more coherent and instruction-aligned results compared to existing approaches.

Conclusion: The proposed framework enables precise and harmonized image editing by encouraging regions to understand their role within the global image context through multilevel semantic alignment between vision and language.

Abstract: Recent research has made significant progress in localizing and editing image regions based on text. However, most approaches treat these regions in isolation, relying solely on local cues without accounting for how each part contributes to the overall visual and semantic composition. This often results in inconsistent edits, unnatural transitions, or loss of coherence across the image. In this work, we propose Region in Context, a novel framework for text-conditioned image editing that performs multilevel semantic alignment between vision and language, inspired by the human ability to reason about edits in relation to the whole scene. Our method encourages each region to understand its role within the global image context, enabling precise and harmonized changes. At its core, the framework introduces a dual-level guidance mechanism: regions are represented with full-image context and aligned with detailed region-level descriptions, while the entire image is simultaneously matched to a comprehensive scene-level description generated by a large vision-language model. These descriptions serve as explicit verbal references of the intended content, guiding both local modifications and global structure. Experiments show that it produces more coherent and instruction-aligned results. Code is available at: https://github.com/thuyvuphuong/Region-in-Context.git

[275] EMRRG: Efficient Fine-Tuning Pre-trained X-ray Mamba Networks for Radiology Report Generation

Mingzheng Zhang, Jinfeng Gao, Dan Xu, Jiangrui Yu, Yuhan Qiao, Lan Chen, Jin Tang, Xiao Wang

Main category: cs.CV

TL;DR: EMRRG is a novel X-ray medical report generation framework that fine-tunes pre-trained Mamba networks using parameter-efficient methods, achieving strong performance on benchmark datasets.

DetailsMotivation: Existing MRG models rely heavily on LLMs with limited exploration of pre-trained vision foundation models, advanced fine-tuning techniques, and non-Transformer architectures like Mamba networks for medical report generation.

Method: X-ray images are divided into patches, tokenized, and processed by an SSM-based vision backbone for feature extraction using Partial LoRA. An LLM with hybrid decoder generates medical reports through end-to-end training.

Result: Extensive experiments on three widely used benchmark datasets fully validated the effectiveness of the proposed strategies for X-ray medical report generation.

Conclusion: The proposed EMRRG framework demonstrates the potential of fine-tuning pre-trained Mamba networks with parameter-efficient methods for medical report generation, achieving strong results on benchmark datasets.

Abstract: X-ray image-based medical report generation (MRG) is a pivotal area in artificial intelligence that can significantly reduce diagnostic burdens for clinicians and patient wait times. Existing MRG models predominantly rely on Large Language Models (LLMs) to improve report generation, with limited exploration of pre-trained vision foundation models or advanced fine-tuning techniques. Mainstream frameworks either avoid fine-tuning or utilize simplistic methods like LoRA, often neglecting the potential of enhancing cross-attention mechanisms. Additionally, while Transformer-based models dominate vision-language tasks, non-Transformer architectures, such as the Mamba network, remain underexplored for medical report generation, presenting a promising avenue for future research. In this paper, we propose EMRRG, a novel X-ray report generation framework that fine-tunes pre-trained Mamba networks using parameter-efficient methods. Specifically, X-ray images are divided into patches, tokenized, and processed by an SSM-based vision backbone for feature extraction, with Partial LoRA yielding optimal performance. An LLM with a hybrid decoder generates the medical report, enabling end-to-end training and achieving strong results on benchmark datasets. Extensive experiments on three widely used benchmark datasets fully validated the effectiveness of our proposed strategies for the X-ray MRG. The source code of this paper will be released on https://github.com/Event-AHU/Medical_Image_Analysis.

[276] GS2POSE: Marry Gaussian Splatting to 6D Object Pose Estimation

Junbo Li, Weimin Yuan, Yinuo Wang, Yue Zeng, Shihao Shu, Cai Meng, Xiangzhi Bai

Main category: cs.CV

TL;DR: GS2POSE is a novel 6D object pose estimation method that uses Bundle Adjustment principles with 3D Gaussian Splatting to handle textureless objects and varying illumination.

DetailsMotivation: Current 6D pose estimation methods struggle with textureless objects and varying illumination conditions when establishing 2D-3D correspondences.

Method: Formulates pose regression using Bundle Adjustment principles, extends 3DGS with Lie algebra for pose-differentiable rendering, and iteratively optimizes pose by comparing input and rendered images while updating color parameters.

Result: Achieves accuracy improvements of 1.4%, 2.8% and 2.5% on T-LESS, LineMod-Occlusion and LineMod datasets respectively compared to previous models.

Conclusion: GS2POSE effectively addresses challenges in 6D pose estimation for textureless objects under varying illumination through differentiable rendering and iterative optimization.

Abstract: Accurate 6D pose estimation of 3D objects is a fundamental task in computer vision, and current research typically predicts the 6D pose by establishing correspondences between 2D image features and 3D model features. However, these methods often face difficulties with textureless objects and varying illumination conditions. To overcome these limitations, we propose GS2POSE, a novel approach for 6D object pose estimation. GS2POSE formulates a pose regression algorithm inspired by the principles of Bundle Adjustment (BA). By leveraging Lie algebra, we extend the capabilities of 3DGS to develop a pose-differentiable rendering pipeline, which iteratively optimizes the pose by comparing the input image to the rendered image. Additionally, GS2POSE updates color parameters within the 3DGS model, enhancing its adaptability to changes in illumination. Compared to previous models, GS2POSE demonstrates accuracy improvements of 1.4%, 2.8% and 2.5% on the T-LESS, LineMod-Occlusion and LineMod datasets, respectively.

[277] Xiaoice: Training-Free Video Understanding via Self-Supervised Spatio-Temporal Clustering of Semantic Features

Shihao Ji, Zihui Song

Main category: cs.CV

TL;DR: Training-free video understanding framework using pre-trained VLMs and clustering algorithms for zero-shot structural analysis without end-to-end training.

DetailsMotivation: Current video understanding models require extensive annotated datasets and task-specific training, which is costly and lacks scalability. The goal is to translate zero-shot reasoning capabilities from static images to videos without training.

Method: Reframe video understanding as self-supervised spatio-temporal clustering. Transform video into semantic feature trajectory using frozen VLM encoder, apply Kernel Temporal Segmentation (KTS) for event segmentation, then use density-based clustering to identify recurring scenes and themes.

Result: Automatically produces structured, multi-modal video summaries by selecting representative keyframes from discovered clusters and generating textual descriptions using VLM capabilities.

Conclusion: Provides an effective, interpretable, and model-agnostic pathway for zero-shot automated structural analysis of video content without requiring training.

Abstract: The remarkable zero-shot reasoning capabilities of large-scale Visual Language Models (VLMs) on static images have yet to be fully translated to the video domain. Conventional video understanding models often rely on extensive, task-specific training on annotated datasets, a process that is both costly and limited in scalability. This paper introduces a novel, training-free framework for video understanding that circumvents end-to-end training by synergistically combining the rich semantic priors of pre-trained VLMs with classic machine learning algorithms for pattern discovery. Our core idea is to reframe video understanding as a self-supervised spatio-temporal clustering problem within a high-dimensional semantic feature space. The proposed pipeline first transforms a video stream into a semantic feature trajectory using the frozen visual encoder of a pre-trained VLM. Subsequently, we employ Kernel Temporal Segmentation (KTS), a robust machine learning technique, to partition the continuous feature stream into discrete, semantically coherent event segments. These segments are then subjected to unsupervised density-based clustering to identify recurring macroscopic scenes and themes throughout the video. By selecting representative keyframes from each discovered cluster and leveraging the VLM’s generative capabilities for textual description, our framework automatically produces a structured, multi-modal summary of the video content. This approach provides an effective, interpretable, and model-agnostic pathway for zero-shot, automated structural analysis of video content.

[278] Segmentation as A Plug-and-Play Capability for Frozen Multimodal LLMs

Jiazhen Liu, Long Chen

Main category: cs.CV

TL;DR: LENS is a plug-and-play method that enables Multimodal Large Language Models (MLLMs) to perform pixel-level segmentation without finetuning, preserving the model’s generalization capabilities while achieving competitive segmentation performance.

DetailsMotivation: Current methods for adding segmentation to MLLMs require finetuning, which alters the model's output space and compromises its intrinsic generalization, undermining the goal of building unified multimodal models.

Method: LENS attaches a lightweight, trainable head to a frozen MLLM, refining spatial cues from attention maps to extract keypoints and generate point-wise features compatible with mask decoders.

Result: Extensive experiments show LENS achieves segmentation performance competitive with or superior to retraining-based methods while fully preserving the MLLM’s generalization capabilities.

Conclusion: LENS establishes an efficient paradigm for extending MLLMs with segmentation capabilities without compromising their generalization, paving the way for truly multi-talented unified models.

Abstract: Integrating diverse visual capabilities into a unified model is a significant trend in Multimodal Large Language Models (MLLMs). Among these, the inclusion of segmentation poses a distinct set of challenges. To equip MLLMs with pixel-level segmentation abilities, prevailing methods require finetuning the model to produce specific outputs compatible with a mask decoder. This process typically alters the model’s output space and compromises its intrinsic generalization, which undermines the goal of building a unified model. We introduce LENS (Leveraging kEypoiNts for MLLMs’ Segmentation), a novel plug-and-play solution. LENS attaches a lightweight, trainable head to a completely frozen MLLM. By refining the spatial cues embedded in attention maps, LENS extracts keypoints and describes them into point-wise features directly compatible with the mask decoder. Extensive experiments validate our approach: LENS achieves segmentation performance competitive with or superior to that of retraining-based methods. Crucially, it does so while fully preserving the MLLM’s generalization capabilities, which are significantly degraded by finetuning approaches. As such, the attachable design of LENS establishes an efficient and powerful paradigm for extending MLLMs, paving the way for truly multi-talented, unified models.

[279] Unsupervised Monocular Road Segmentation for Autonomous Driving via Scene Geometry

Sara Hatami Rostami, Behrooz Nasihatkon

Main category: cs.CV

TL;DR: Unsupervised binary road segmentation using geometric priors and temporal consistency, achieving 0.82 IoU on Cityscapes without manual labels.

DetailsMotivation: Eliminate reliance on costly manually labeled datasets for road segmentation in autonomous driving by leveraging unsupervised methods.

Method: Uses geometric priors to generate weak labels (pixels above horizon as non-road, quadrilateral in front as road), then enforces temporal consistency through feature tracking and mutual information maximization.

Result: Achieves 0.82 Intersection-over-Union (IoU) on Cityscapes dataset with high accuracy and temporal stability.

Conclusion: Combining geometric constraints and temporal consistency enables scalable unsupervised road segmentation for autonomous driving applications.

Abstract: This paper presents a fully unsupervised approach for binary road segmentation (road vs. non-road), eliminating the reliance on costly manually labeled datasets. The method leverages scene geometry and temporal cues to distinguish road from non-road regions. Weak labels are first generated from geometric priors, marking pixels above the horizon as non-road and a predefined quadrilateral in front of the vehicle as road. In a refinement stage, temporal consistency is enforced by tracking local feature points across frames and penalizing inconsistent label assignments using mutual information maximization. This enhances both precision and temporal stability. On the Cityscapes dataset, the model achieves an Intersection-over-Union (IoU) of 0.82, demonstrating high accuracy with a simple design. These findings demonstrate the potential of combining geometric constraints and temporal consistency for scalable unsupervised road segmentation in autonomous driving.

[280] Personalized Image Filter: Mastering Your Photographic Style

Chengxuan Zhu, Shuchen Weng, Jiacong Fang, Peixuan Zhang, Si Li, Chao Xu, Boxin Shi

Main category: cs.CV

TL;DR: PIF (Personalized Image Filter) is a method that learns photographic styles from reference images using text-to-image diffusion models and textual inversion, enabling effective style transfer while preserving content.

DetailsMotivation: Previous methods fail to learn meaningful photographic concepts from reference images or cannot preserve content image integrity when transferring photographic styles.

Method: Based on pretrained text-to-image diffusion model, PIF learns average appearance of photographic concepts and adjusts them via text prompts. Uses textual inversion technique to optimize prompts for photographic concepts from reference images.

Result: PIF shows outstanding performance in extracting and transferring various kinds of photographic style.

Conclusion: PIF effectively addresses limitations of previous methods by leveraging diffusion model generative prior and textual inversion for photographic style learning and transfer.

Abstract: Photographic style, as a composition of certain photographic concepts, is the charm behind renowned photographers. But learning and transferring photographic style need a profound understanding of how the photo is edited from the unknown original appearance. Previous works either fail to learn meaningful photographic concepts from reference images, or cannot preserve the content of the content image. To tackle these issues, we proposed a Personalized Image Filter (PIF). Based on a pretrained text-to-image diffusion model, the generative prior enables PIF to learn the average appearance of photographic concepts, as well as how to adjust them according to text prompts. PIF then learns the photographic style of reference images with the textual inversion technique, by optimizing the prompts for the photographic concepts. PIF shows outstanding performance in extracting and transferring various kinds of photographic style. Project page: https://pif.pages.dev/

[281] An RGB-D Image Dataset for Lychee Detection and Maturity Classification for Robotic Harvesting

Zhenpeng Zhang, Yi Wang, Shanglei Chai, Yingying Liu, Zekai Xie, Wenhao Huang, Pengyu Li, Zipei Luo, Dajiang Lu, Yibin Tian

Main category: cs.CV

TL;DR: Created a comprehensive lychee dataset with 11,414 images (RGB and depth) featuring multiple varieties and ripeness stages, annotated for detection and maturity classification, to support vision-based harvesting robot development.

DetailsMotivation: Address the lack of consistently annotated open-source lychee datasets for developing vision-based harvesting robots, which can improve productivity and reduce labor dependency in lychee cultivation.

Method: Collected color (RGB) images under diverse weather conditions and times across multiple lychee varieties, applied data augmentation, included depth images, and implemented rigorous annotation process with multiple independent labelers and verification.

Result: Produced a dataset with 11,414 images (878 raw RGB, 8,780 augmented RGB, 1,756 depth) annotated with 9,658 label pairs for detection and maturity classification across three ripeness stages.

Conclusion: The dataset enables development of vision-based harvesting robots for lychee, is publicly available for academic use, and was validated using three deep learning models.

Abstract: Lychee is a high-value subtropical fruit. The adoption of vision-based harvesting robots can significantly improve productivity while reduce reliance on labor. High-quality data are essential for developing such harvesting robots. However, there are currently no consistently and comprehensively annotated open-source lychee datasets featuring fruits in natural growing environments. To address this, we constructed a dataset to facilitate lychee detection and maturity classification. Color (RGB) images were acquired under diverse weather conditions, and at different times of the day, across multiple lychee varieties, such as Nuomici, Feizixiao, Heiye, and Huaizhi. The dataset encompasses three different ripeness stages and contains 11,414 images, consisting of 878 raw RGB images, 8,780 augmented RGB images, and 1,756 depth images. The images are annotated with 9,658 pairs of lables for lychee detection and maturity classification. To improve annotation consistency, three individuals independently labeled the data, and their results were then aggregated and verified by a fourth reviewer. Detailed statistical analyses were done to examine the dataset. Finally, we performed experiments using three representative deep learning models to evaluate the dataset. It is publicly available for academic

[282] ReefNet: A Large scale, Taxonomically Enriched Dataset and Benchmark for Hard Coral Classification

Yahia Battach, Abdulwahab Felemban, Faizan Farooq Khan, Yousef A. Radwan, Xiang Li, Fabio Marchese, Sara Beery, Burton H. Jones, Francesca Benzoni, Mohamed Elhoseiny

Main category: cs.CV

TL;DR: ReefNet is a large public coral reef image dataset with fine-grained genus-level annotations mapped to WoRMS, providing challenging benchmarks for domain generalization and fine-grained coral classification.

DetailsMotivation: Coral reefs are declining rapidly due to climate change, creating urgent need for scalable automated monitoring. Existing datasets are limited by size, geography, or coarse labels and are not ML-ready.

Method: Aggregated imagery from 76 curated CoralNet sources and Al Wajh site, totaling ~925K genus-level hard coral annotations with expert-verified labels mapped to WoRMS. Proposed two evaluation settings: within-source and cross-source benchmarks.

Result: Supervised within-source performance is promising but drops sharply across domains. Zero-shot models perform poorly across the board, especially for rare and visually similar genera.

Conclusion: ReefNet provides a challenging benchmark to catalyze advances in domain generalization and fine-grained coral classification, supporting robust global coral reef monitoring and conservation.

Abstract: Coral reefs are rapidly declining due to anthropogenic pressures such as climate change, underscoring the urgent need for scalable, automated monitoring. We introduce ReefNet, a large public coral reef image dataset with point-label annotations mapped to the World Register of Marine Species (WoRMS). ReefNet aggregates imagery from 76 curated CoralNet sources and an additional site from Al Wajh in the Red Sea, totaling approximately 925000 genus-level hard coral annotations with expert-verified labels. Unlike prior datasets, which are often limited by size, geography, or coarse labels and are not ML-ready, ReefNet offers fine-grained, taxonomically mapped labels at a global scale to WoRMS. We propose two evaluation settings: (i) a within-source benchmark that partitions each source’s images for localized evaluation, and (ii) a cross-source benchmark that withholds entire sources to test domain generalization. We analyze both supervised and zero-shot classification performance on ReefNet and find that while supervised within-source performance is promising, supervised performance drops sharply across domains, and performance is low across the board for zero-shot models, especially for rare and visually similar genera. This provides a challenging benchmark intended to catalyze advances in domain generalization and fine-grained coral classification. We will release our dataset, benchmarking code, and pretrained models to advance robust, domain-adaptive, global coral reef monitoring and conservation.

[283] Robust Cross-Domain Adaptation in Texture Features Transferring for Wood Chip Moisture Content Prediction

Abdur Rahman, Mohammad Marufuzzaman, Jason Street, Haifeng Wang, Veera G. Gude, Randy Buchanan

Main category: cs.CV

TL;DR: This paper proposes AdaptMoist, a domain adaptation method that uses texture features from wood chip images to predict moisture content across different wood sources, achieving 80% accuracy compared to 57% for non-adapted models.

DetailsMotivation: Current moisture prediction methods for wood chips are either slow/destructive (oven drying) or inaccurate when dealing with wood from various sources due to data distribution shifts. There's a need for robust approaches that handle source variability.

Method: Comprehensive analysis of five texture feature types from wood chip images, combined feature sets, and a domain adaptation method (AdaptMoist) that transfers knowledge between wood chip sources using texture features with model saving based on adjusted mutual information.

Result: Combined texture features achieved 95% accuracy for moisture prediction. AdaptMoist improved cross-domain prediction accuracy by 23%, achieving 80% average accuracy compared to 57% for non-adapted models.

Conclusion: AdaptMoist is an effective robust solution for wood chip moisture content estimation across domains, making it suitable for wood chip-reliant industries by addressing source variability issues.

Abstract: Accurate and quick prediction of wood chip moisture content is critical for optimizing biofuel production and ensuring energy efficiency. The current widely used direct method (oven drying) is limited by its longer processing time and sample destructiveness. On the other hand, existing indirect methods, including near-infrared spectroscopy-based, electrical capacitance-based, and image-based approaches, are quick but not accurate when wood chips come from various sources. Variability in the source material can alter data distributions, undermining the performance of data-driven models. Therefore, there is a need for a robust approach that effectively mitigates the impact of source variability. Previous studies show that manually extracted texture features have the potential to predict wood chip moisture class. Building on this, in this study, we conduct a comprehensive analysis of five distinct texture feature types extracted from wood chip images to predict moisture content. Our findings reveal that a combined feature set incorporating all five texture features achieves an accuracy of 95% and consistently outperforms individual texture features in predicting moisture content. To ensure robust moisture prediction, we propose a domain adaptation method named AdaptMoist that utilizes the texture features to transfer knowledge from one source of wood chip data to another, addressing variability across different domains. We also proposed a criterion for model saving based on adjusted mutual information. The AdaptMoist method improves prediction accuracy across domains by 23%, achieving an average accuracy of 80%, compared to 57% for non-adapted models. These results highlight the effectiveness of AdaptMoist as a robust solution for wood chip moisture content estimation across domains, making it a potential solution for wood chip-reliant industries.

[284] From Mannequin to Human: A Pose-Aware and Identity-Preserving Video Generation Framework for Lifelike Clothing Display

Xiangyu Mu, Dongliang Zhou, Jie Hou, Haijun Zhang, Weili Guan

Main category: cs.CV

TL;DR: M2HVideo is a framework for generating photorealistic human videos from mannequin footage, addressing head-body misalignment and identity drift through pose-aware head encoding, mirror loss, and distribution-aware feature alignment.

DetailsMotivation: Mannequin-based clothing displays are cost-effective but lack realism and expressive detail compared to real-model showcases, limiting their effectiveness for online fashion presentation.

Method: Proposes M2HVideo with dynamic pose-aware head encoder for consistent identity embeddings, mirror loss in pixel space via DDIM-based denoising, and distribution-aware adapter for temporal coherence.

Result: Extensive experiments on UBC fashion, ASOS, and MannequinVideos datasets show superior performance in clothing consistency, identity preservation, and video fidelity compared to state-of-the-art methods.

Conclusion: M2HVideo effectively addresses key challenges in mannequin-to-human video generation, providing a practical solution for realistic fashion presentation while maintaining identity consistency and temporal coherence.

Abstract: Mannequin-based clothing displays offer a cost-effective alternative to real-model showcases for online fashion presentation, but lack realism and expressive detail. To overcome this limitation, we introduce a new task called mannequin-to-human (M2H) video generation, which aims to synthesize identity-controllable, photorealistic human videos from footage of mannequins. We propose M2HVideo, a pose-aware and identity-preserving video generation framework that addresses two key challenges: the misalignment between head and body motion, and identity drift caused by temporal modeling. In particular, M2HVideo incorporates a dynamic pose-aware head encoder that fuses facial semantics with body pose to produce consistent identity embeddings across frames. To address the loss of fine facial details due to latent space compression, we introduce a mirror loss applied in pixel space through a denoising diffusion implicit model (DDIM)-based one-step denoising. Additionally, we design a distribution-aware adapter that aligns statistical distributions of identity and clothing features to enhance temporal coherence. Extensive experiments on the UBC fashion dataset, our self-constructed ASOS dataset, and the newly collected MannequinVideos dataset captured on-site demonstrate that M2HVideo achieves superior performance in terms of clothing consistency, identity preservation, and video fidelity in comparison to state-of-the-art methods.

[285] 2DGS-R: Revisiting the Normal Consistency Regularization in 2D Gaussian Splatting

Haofan Ren, Qingsong Yan, Ming Lu, Rongfeng Lu, Zunjie Zhu

Main category: cs.CV

TL;DR: 2DGS-R improves 3D Gaussian Splatting by using hierarchical training to achieve both high-quality rendering and precise geometric structures, with minimal storage and training time overhead.

DetailsMotivation: Current 3DGS methods struggle to balance high-quality rendering with accurate surface representation, while 2DGS improves geometry but compromises rendering quality, making single-stage optimization infeasible.

Method: Uses hierarchical training: first trains 2D Gaussians with normal consistency regularization, then selects underperforming Gaussians for in-place cloning, and finally fine-tunes with frozen opacity.

Result: Achieves high-quality rendering while preserving fine geometric structures with only 1% more storage and minimal additional training time compared to original 2DGS.

Conclusion: The approach effectively balances efficiency with performance, leading to improvements in both visual fidelity and geometric reconstruction accuracy.

Abstract: Recent advancements in 3D Gaussian Splatting (3DGS) have greatly influenced neural fields, as it enables high-fidelity rendering with impressive visual quality. However, 3DGS has difficulty accurately representing surfaces. In contrast, 2DGS transforms the 3D volume into a collection of 2D planar Gaussian disks. Despite advancements in geometric fidelity, rendering quality remains compromised, highlighting the challenge of achieving both high-quality rendering and precise geometric structures. This indicates that optimizing both geometric and rendering quality in a single training stage is currently unfeasible. To overcome this limitation, we present 2DGS-R, a new method that uses a hierarchical training approach to improve rendering quality while maintaining geometric accuracy. 2DGS-R first trains the original 2D Gaussians with the normal consistency regularization. Then 2DGS-R selects the 2D Gaussians with inadequate rendering quality and applies a novel in-place cloning operation to enhance the 2D Gaussians. Finally, we fine-tune the 2DGS-R model with opacity frozen. Experimental results show that compared to the original 2DGS, our method requires only 1% more storage and minimal additional training time. Despite this negligible overhead, it achieves high-quality rendering results while preserving fine geometric structures. These findings indicate that our approach effectively balances efficiency with performance, leading to improvements in both visual fidelity and geometric reconstruction accuracy.

[286] ArmFormer: Lightweight Transformer Architecture for Real-Time Multi-Class Weapon Segmentation and Classification

Akhila Kambhatla, Taminul Islam, Khaled R Ahmed

Main category: cs.CV

TL;DR: ArmFormer is a lightweight transformer-based semantic segmentation framework that achieves pixel-level weapon detection with high accuracy and computational efficiency, making it suitable for real-time edge deployment in security applications.

DetailsMotivation: The need for automated weapon detection systems with pixel-level precision for real-time threat assessment, as traditional object detection provides only coarse bounding boxes and existing segmentation models are either inaccurate or computationally intensive for edge devices.

Method: Integrates Convolutional Block Attention Module (CBAM) with MixVisionTransformer architecture, combining CBAM-enhanced encoder backbone with attention-integrated hamburger decoder for multi-class weapon segmentation across five categories.

Result: Achieves state-of-the-art performance with 80.64% mIoU and 89.13% mFscore while maintaining real-time inference at 82.26 FPS, with only 4.886G FLOPs and 3.66M parameters.

Conclusion: ArmFormer outperforms heavyweight models requiring up to 48x more computation, establishing it as the optimal solution for deployment on portable security cameras, surveillance drones, and embedded AI accelerators in distributed security infrastructure.

Abstract: The escalating threat of weapon-related violence necessitates automated detection systems capable of pixel-level precision for accurate threat assessment in real-time security applications. Traditional weapon detection approaches rely on object detection frameworks that provide only coarse bounding box localizations, lacking the fine-grained segmentation required for comprehensive threat analysis. Furthermore, existing semantic segmentation models either sacrifice accuracy for computational efficiency or require excessive computational resources incompatible with edge deployment scenarios. This paper presents ArmFormer, a lightweight transformer-based semantic segmentation framework that strategically integrates Convolutional Block Attention Module (CBAM) with MixVisionTransformer architecture to achieve superior accuracy while maintaining computational efficiency suitable for resource-constrained edge devices. Our approach combines CBAM-enhanced encoder backbone with attention-integrated hamburger decoder to enable multi-class weapon segmentation across five categories: handgun, rifle, knife, revolver, and human. Comprehensive experiments demonstrate that ArmFormer achieves state-of-the-art performance with 80.64% mIoU and 89.13% mFscore while maintaining real-time inference at 82.26 FPS. With only 4.886G FLOPs and 3.66M parameters, ArmFormer outperforms heavyweight models requiring up to 48x more computation, establishing it as the optimal solution for deployment on portable security cameras, surveillance drones, and embedded AI accelerators in distributed security infrastructure.

[287] BARL: Bilateral Alignment in Representation and Label Spaces for Semi-Supervised Volumetric Medical Image Segmentation

Shujian Gao, Yuan Wang, Zekuan Yu

Main category: cs.CV

TL;DR: BARL introduces a semi-supervised medical image segmentation framework that enforces alignment in both representation and label spaces to improve performance while reducing annotation costs.

DetailsMotivation: Current SSMIS methods focus only on label-space consistency but overlook representation-space alignment, leading to models that struggle to learn discriminative and spatially coherent representations for complex medical images.

Method: BARL uses two collaborative branches with Dual-Path Regularization (DPR) and Progressively Cognitive Bias Correction (PCBC) for label-space alignment, plus region-level and lesion-instance matching for representation-space alignment.

Result: Extensive experiments on four public benchmarks and a proprietary CBCT dataset show BARL consistently outperforms state-of-the-art SSMIS methods.

Conclusion: BARL effectively addresses the limitations of existing SSMIS methods by enforcing bilateral alignment, demonstrating superior performance across multiple medical imaging datasets.

Abstract: Semi-supervised medical image segmentation (SSMIS) seeks to match fully supervised performance while sharply reducing annotation cost. Mainstream SSMIS methods rely on \emph{label-space consistency}, yet they overlook the equally critical \emph{representation-space alignment}. Without harmonizing latent features, models struggle to learn representations that are both discriminative and spatially coherent. To this end, we introduce \textbf{Bilateral Alignment in Representation and Label spaces (BARL)}, a unified framework that couples two collaborative branches and enforces alignment in both spaces. For label-space alignment, inspired by co-training and multi-scale decoding, we devise \textbf{Dual-Path Regularization (DPR)} and \textbf{Progressively Cognitive Bias Correction (PCBC)} to impose fine-grained cross-branch consistency while mitigating error accumulation from coarse to fine scales. For representation-space alignment, we conduct region-level and lesion-instance matching between branches, explicitly capturing the fragmented, complex pathological patterns common in medical imagery. Extensive experiments on four public benchmarks and a proprietary CBCT dataset demonstrate that BARL consistently surpasses state-of-the-art SSMIS methods. Ablative studies further validate the contribution of each component. Code will be released soon.

[288] Registration is a Powerful Rotation-Invariance Learner for 3D Anomaly Detection

Yuyang Yu, Zhengwei Chen, Xuemiao Xu, Lei Zhang, Haoxin Yang, Yongwei Nie, Shengfeng He

Main category: cs.CV

TL;DR: A registration-induced, rotation-invariant feature extraction framework that integrates point-cloud registration with memory-based anomaly detection to improve 3D anomaly detection performance.

DetailsMotivation: Current memory bank-based methods for 3D anomaly detection suffer from inconsistent feature transformations and limited discriminative capacity, especially in capturing local geometric details and achieving rotation invariance, particularly when registration fails.

Method: Proposes a framework that embeds feature extraction into the registration learning process, jointly optimizing alignment and representation learning to acquire rotation-invariant and locally discriminative features.

Result: Extensive experiments on Anomaly-ShapeNet and Real3D-AD datasets show the method consistently outperforms existing approaches in effectiveness and generalizability.

Conclusion: Point-cloud registration plays an essential role in guiding feature extraction toward rotation-invariant and locally discriminative representations, and integrating registration with anomaly detection significantly improves performance.

Abstract: 3D anomaly detection in point-cloud data is critical for industrial quality control, aiming to identify structural defects with high reliability. However, current memory bank-based methods often suffer from inconsistent feature transformations and limited discriminative capacity, particularly in capturing local geometric details and achieving rotation invariance. These limitations become more pronounced when registration fails, leading to unreliable detection results. We argue that point-cloud registration plays an essential role not only in aligning geometric structures but also in guiding feature extraction toward rotation-invariant and locally discriminative representations. To this end, we propose a registration-induced, rotation-invariant feature extraction framework that integrates the objectives of point-cloud registration and memory-based anomaly detection. Our key insight is that both tasks rely on modeling local geometric structures and leveraging feature similarity across samples. By embedding feature extraction into the registration learning process, our framework jointly optimizes alignment and representation learning. This integration enables the network to acquire features that are both robust to rotations and highly effective for anomaly detection. Extensive experiments on the Anomaly-ShapeNet and Real3D-AD datasets demonstrate that our method consistently outperforms existing approaches in effectiveness and generalizability.

[289] Uncovering Brain-Like Hierarchical Patterns in Vision-Language Models through fMRI-Based Neural Encoding

Yudan Ren, Xinlong Wang, Kexin Wang, Tian Xia, Zihan Ma, Zhaowei Li, Xiangrong Bi, Xiao Li, Xiaowei He

Main category: cs.CV

TL;DR: A neuron-level analysis framework reveals brain-like processing in vision-language models, showing shared representational mechanisms, functional redundancy, polarity patterns, and architecture-dependent neural activation between artificial and biological neurons.

DetailsMotivation: Current ANN studies have limitations: unimodal approaches don't capture brain's multimodal processing, and multimodal research focuses on high-level outputs while neglecting individual neurons' crucial role.

Method: Proposed a novel neuron-level analysis framework combining fine-grained artificial neuron analysis with fMRI-based voxel encoding to examine two architecturally distinct VLMs (CLIP and METER).

Result: Four key findings: (1) ANs predict BN activities across multiple functional networks; (2) Both show functional redundancy; (3) ANs exhibit polarity patterns paralleling BNs; (4) Different architectures drive distinct BN activations - CLIP shows modality-specific specialization while METER yields unified cross-modal activation.

Conclusion: The results provide compelling evidence for brain-like hierarchical processing in VLMs at the neuronal level, demonstrating shared representational mechanisms between artificial and biological neural systems.

Abstract: While brain-inspired artificial intelligence(AI) has demonstrated promising results, current understanding of the parallels between artificial neural networks (ANNs) and human brain processing remains limited: (1) unimodal ANN studies fail to capture the brain’s inherent multimodal processing capabilities, and (2) multimodal ANN research primarily focuses on high-level model outputs, neglecting the crucial role of individual neurons. To address these limitations, we propose a novel neuron-level analysis framework that investigates the multimodal information processing mechanisms in vision-language models (VLMs) through the lens of human brain activity. Our approach uniquely combines fine-grained artificial neuron (AN) analysis with fMRI-based voxel encoding to examine two architecturally distinct VLMs: CLIP and METER. Our analysis reveals four key findings: (1) ANs successfully predict biological neurons (BNs) activities across multiple functional networks (including language, vision, attention, and default mode), demonstrating shared representational mechanisms; (2) Both ANs and BNs demonstrate functional redundancy through overlapping neural representations, mirroring the brain’s fault-tolerant and collaborative information processing mechanisms; (3) ANs exhibit polarity patterns that parallel the BNs, with oppositely activated BNs showing mirrored activation trends across VLM layers, reflecting the complexity and bidirectional nature of neural information processing; (4) The architectures of CLIP and METER drive distinct BNs: CLIP’s independent branches show modality-specific specialization, whereas METER’s cross-modal design yields unified cross-modal activation, highlighting the architecture’s influence on ANN brain-like properties. These results provide compelling evidence for brain-like hierarchical processing in VLMs at the neuronal level.

[290] Class-N-Diff: Classification-Induced Diffusion Model Can Make Fair Skin Cancer Diagnosis

Nusrat Munia, Abdullah Imran

Main category: cs.CV

TL;DR: Class-N-Diff is a classification-induced diffusion model that simultaneously generates and classifies dermoscopic images by integrating a classifier within a diffusion model for better class-conditioned image synthesis.

DetailsMotivation: Traditional class-conditioned generative models struggle to generate accurate medical images for specific categories, limiting their usefulness in applications like skin cancer diagnosis.

Method: Integrates a classifier within a diffusion model to guide image generation based on class conditions, enabling better control over class-conditioned image synthesis.

Result: Generates more realistic and diverse dermoscopic images, and the classifier shows improved performance for downstream diagnostic tasks.

Conclusion: Class-N-Diff is a robust tool for enhancing the quality and utility of diffusion model-based synthetic dermoscopic image generation.

Abstract: Generative models, especially Diffusion Models, have demonstrated remarkable capability in generating high-quality synthetic data, including medical images. However, traditional class-conditioned generative models often struggle to generate images that accurately represent specific medical categories, limiting their usefulness for applications such as skin cancer diagnosis. To address this problem, we propose a classification-induced diffusion model, namely, Class-N-Diff, to simultaneously generate and classify dermoscopic images. Our Class-N-Diff model integrates a classifier within a diffusion model to guide image generation based on its class conditions. Thus, the model has better control over class-conditioned image synthesis, resulting in more realistic and diverse images. Additionally, the classifier demonstrates improved performance, highlighting its effectiveness for downstream diagnostic tasks. This unique integration in our Class-N-Diff makes it a robust tool for enhancing the quality and utility of diffusion model-based synthetic dermoscopic image generation. Our code is available at https://github.com/Munia03/Class-N-Diff.

[291] Uniworld-V2: Reinforce Image Editing with Diffusion Negative-aware Finetuning and MLLM Implicit Feedback

Zongjian Li, Zheyuan Liu, Qihui Zhang, Bin Lin, Shenghai Yuan, Zhiyuan Yan, Yang Ye, Wangbo Yu, Yuwei Niu, Li Yuan

Main category: cs.CV

TL;DR: Edit-R1 is a post-training framework using policy optimization to enhance instruction-based image editing models, overcoming overfitting and improving generalization through DiffusionNFT and MLLM-based rewards.

DetailsMotivation: Models trained via supervised fine-tuning often overfit to annotated patterns, limiting their ability to generalize beyond training distributions in instruction-based image editing.

Method: Uses Diffusion Negative-aware Finetuning (DiffusionNFT) for policy optimization and employs a Multimodal Large Language Model as a unified, training-free reward model with low-variance group filtering to reduce noise.

Result: UniWorld-V2 achieves state-of-the-art results on ImgEdit (4.49) and GEdit-Bench (7.83), with the framework being model-agnostic and delivering substantial performance gains across diverse base models.

Conclusion: The Edit-R1 framework effectively addresses overfitting and generalization issues in instruction-based image editing through policy optimization and MLLM-based rewards, demonstrating wide applicability and superior performance.

Abstract: Instruction-based image editing has achieved remarkable progress; however, models solely trained via supervised fine-tuning often overfit to annotated patterns, hindering their ability to explore and generalize beyond training distributions. To this end, we introduce Edit-R1, a novel post-training framework for instruction-based image editing based on policy optimization. Specifically, we utilize Diffusion Negative-aware Finetuning (DiffusionNFT), a likelihood-free policy optimization method consistent with the flow matching forward process, thereby enabling the use of higher-order samplers and more efficient training. Another key challenge here is the absence of a universal reward model, resulting from the diverse nature of editing instructions and tasks. To bridge this gap, we employ a Multimodal Large Language Model (MLLM) as a unified, training-free reward model, leveraging its output logits to provide fine-grained feedback. Furthermore, we carefully design a low-variance group filtering mechanism to reduce MLLM scoring noise and stabilize optimization. UniWorld-V2, trained with this framework, achieves \textbf{state-of-the-art} results on the ImgEdit and GEdit-Bench benchmarks, scoring 4.49 and 7.83, respectively. Crucially, our framework is model-agnostic, delivering substantial performance gains when applied to diverse base models like Qwen-Image-Edit and FLUX-Kontext, demonstrating its wide applicability. Code and models are publicly available at https://github.com/PKU-YuanGroup/UniWorld-V2.

[292] Contrail-to-Flight Attribution Using Ground Visible Cameras and Flight Surveillance Data

Ramon Dalmau, Gabriel Jarry, Philippe Very

Main category: cs.CV

TL;DR: A modular framework for attributing contrails observed by ground-based cameras to their source flights using aircraft surveillance and meteorological data.

DetailsMotivation: Aviation's non-CO2 effects, particularly contrails, significantly contribute to climate impact. Validating contrail models requires linking observed contrails to source flights, which is challenging with satellites due to limited resolution and contrail drift.

Method: Uses ground-based cameras to capture contrails shortly after formation when they remain thin and distinct. Introduces a modular framework with multiple geometric representations, distance metrics, temporal smoothing, and probability-based assignment strategies.

Result: Establishes a strong baseline for contrail-to-flight attribution using the ground visible camera contrail sequences (GVCCS) dataset.

Conclusion: Provides a modular framework for future research in linking contrails to their source flights, enabling better validation of contrail formation and evolution models.

Abstract: Aviation’s non-CO2 effects, particularly contrails, are a significant contributor to its climate impact. Persistent contrails can evolve into cirrus-like clouds that trap outgoing infrared radiation, with radiative forcing potentially comparable to or exceeding that of aviation’s CO2 emissions. While physical models simulate contrail formation, evolution and dissipation, validating and calibrating these models requires linking observed contrails to the flights that generated them, a process known as contrail-to-flight attribution. Satellite-based attribution is challenging due to limited spatial and temporal resolution, as contrails often drift and deform before detection. In this paper, we evaluate an alternative approach using ground-based cameras, which capture contrails shortly after formation at high spatial and temporal resolution, when they remain thin, linear, and visually distinct. Leveraging the ground visible camera contrail sequences (GVCCS) dataset, we introduce a modular framework for attributing contrails observed using ground-based cameras to theoretical contrails derived from aircraft surveillance and meteorological data. The framework accommodates multiple geometric representations and distance metrics, incorporates temporal smoothing, and enables flexible probability-based assignment strategies. This work establishes a strong baseline and provides a modular framework for future research in linking contrails to their source flight.

[293] Beyond RGB: Leveraging Vision Transformers for Thermal Weapon Segmentation

Akhila Kambhatla, Ahmed R Khaled

Main category: cs.CV

TL;DR: This paper evaluates four transformer-based architectures (SegFormer, DeepLabV3+, SegNeXt, Swin Transformer) for thermal weapon segmentation, achieving state-of-the-art performance with SegFormer-b5 reaching 94.15% mIoU and SegFormer-b0 achieving 98.32 FPS.

DetailsMotivation: Thermal weapon segmentation is crucial for surveillance in low-light conditions where RGB systems fail. While CNNs dominate thermal segmentation, they struggle with long-range dependencies and fine details. Transformers offer global context modeling but remain underexplored for thermal weapon segmentation.

Method: Adapted and evaluated four transformer architectures on a custom thermal dataset of 9,711 images from real surveillance videos, automatically annotated using SAM2. Used standard augmentation strategies within MMSegmentation framework for robust training and fair comparison.

Result: SegFormer-b5 achieved highest mIoU (94.15%) and Pixel Accuracy (97.04%), while SegFormer-b0 provided fastest inference (98.32 FPS) with competitive mIoU (90.84%). SegNeXt-mscans offered balanced performance (85.12 FPS, 92.24% mIoU), and DeepLabV3+ R101-D8 reached 92.76% mIoU at 29.86 FPS.

Conclusion: Transformer architectures demonstrate robust generalization for weapon detection in low-light thermal environments, offering flexible accuracy-speed trade-offs suitable for diverse real-time security applications.

Abstract: Thermal weapon segmentation is crucial for surveillance and security applications, enabling robust detection under lowlight and visually obscured conditions where RGB-based systems fail. While convolutional neural networks (CNNs) dominate thermal segmentation literature, their ability to capture long-range dependencies and fine structural details is limited. Vision Transformers (ViTs), with their global context modeling capabilities, have achieved state-of-the-art results in RGB segmentation tasks, yet their potential in thermal weapon segmentation remains underexplored. This work adapts and evaluates four transformer-based architectures SegFormer, DeepLabV3+, SegNeXt, and Swin Transformer for binary weapon segmentation on a custom thermal dataset comprising 9,711 images collected from real world surveillance videos and automatically annotated using SAM2. We employ standard augmentation strategies within the MMSegmentation framework to ensure robust model training and fair architectural comparison. Experimental results demonstrate significant improvements in segmentation performance: SegFormer-b5 achieves the highest mIoU (94.15%) and Pixel Accuracy (97.04%), while SegFormer-b0 provides the fastest inference speed (98.32 FPS) with competitive mIoU (90.84%). SegNeXt-mscans offers balanced performance with 85.12 FPS and 92.24% mIoU, and DeepLabV3+ R101-D8 reaches 92.76% mIoU at 29.86 FPS. The transformer architectures demonstrate robust generalization capabilities for weapon detection in low-light and occluded thermal environments, with flexible accuracy-speed trade-offs suitable for diverse real-time security applications.

[294] Res-Bench: Benchmarking the Robustness of Multimodal Large Language Models to Dynamic Resolution Input

Chenxu Li, Zhicai Wang, Yuan Sheng, Xingyu Zhu, Yanbin Hao, Xiang Wang

Main category: cs.CV

TL;DR: Res-Bench is a comprehensive benchmark for evaluating resolution robustness in Multimodal Large Language Models (MLLMs), introducing novel metrics to assess performance stability across varying input resolutions.

DetailsMotivation: Current MLLM evaluations focus on semantic performance but overlook resolution robustness - whether performance remains stable across different input resolutions, creating a critical gap in assessment.

Method: Developed Res-Bench with 14,400 samples across 12 resolution levels and 6 capability dimensions, introducing novel robustness metrics (Spearman’s correlation, Absolute/Relative Continuous Error) and conducting large-scale evaluation of leading MLLMs.

Result: The benchmark enables comprehensive analysis of model-centric and task-centric robustness, investigation of preprocessing strategies (padding, super-resolution), and exploration of fine-tuning for stability enhancement.

Conclusion: Res-Bench provides a systematic framework for evaluating resolution robustness in MLLMs, addressing a critical gap in current evaluation paradigms and enabling better understanding of model performance stability across varying input resolutions.

Abstract: Multimodal Large Language Models (MLLMs) increasingly support dynamic image resolutions. However, current evaluation paradigms primarily assess semantic performance, overlooking the critical question of resolution robustness - whether performance remains stable across varying input resolutions. To address this gap, we introduce \textbf{Res-Bench}, a comprehensive benchmark comprising 14,400 samples across 12 resolution levels and six core capability dimensions. We designed a novel evaluation framework that goes beyond traditional accuracy metrics to capture performance stability. This framework introduces multiple robustness metrics: Spearman’s correlation for assessing resolution-performance trends, and Absolute/Relative Continuous Error (ACE/RCE) for measuring performance volatility. Using these metrics, we conducted a large-scale evaluation of leading MLLMs. Our analysis encompasses: (1) model-centric and task-centric robustness examination, (2) investigation of preprocessing strategies including padding and super-resolution, and (3) exploration of fine-tuning for stability enhancement.

[295] Foundation Models in Medical Image Analysis: A Systematic Review and Meta-Analysis

Praveenbalaji Rajendran, Mojtaba Safari, Wenfeng He, Mingzhe Hu, Shansong Wang, Jun Zhou, Xiaofeng Yang

Main category: cs.CV

TL;DR: This review paper provides a comprehensive analysis of foundation models (FMs) in medical image analysis, systematically categorizing research into vision-only and vision-language models, analyzing trends, and discussing challenges and future directions.

DetailsMotivation: The field of foundation models in medical imaging is fragmented and lacks a unified synthesis that systematically maps the evolution of architectures, training paradigms, and clinical applications across modalities.

Method: Systematic categorization of studies into vision-only and vision-language FMs based on architectural foundations, training strategies, and downstream clinical tasks. Quantitative meta-analysis of temporal trends in dataset utilization and application domains.

Result: The review provides a structured analysis of FM research in medical imaging, identifying trends in dataset usage and application domains, and critically discussing persistent challenges and emerging solutions.

Conclusion: Key future research directions are identified to enhance robustness, explainability, and clinical integration of FMs, accelerating their translation into real-world medical practice.

Abstract: Recent advancements in artificial intelligence (AI), particularly foundation models (FMs), have revolutionized medical image analysis, demonstrating strong zero- and few-shot performance across diverse medical imaging tasks, from segmentation to report generation. Unlike traditional task-specific AI models, FMs leverage large corpora of labeled and unlabeled multimodal datasets to learn generalized representations that can be adapted to various downstream clinical applications with minimal fine-tuning. However, despite the rapid proliferation of FM research in medical imaging, the field remains fragmented, lacking a unified synthesis that systematically maps the evolution of architectures, training paradigms, and clinical applications across modalities. To address this gap, this review article provides a comprehensive and structured analysis of FMs in medical image analysis. We systematically categorize studies into vision-only and vision-language FMs based on their architectural foundations, training strategies, and downstream clinical tasks. Additionally, a quantitative meta-analysis of the studies was conducted to characterize temporal trends in dataset utilization and application domains. We also critically discuss persistent challenges, including domain adaptation, efficient fine-tuning, computational constraints, and interpretability along with emerging solutions such as federated learning, knowledge distillation, and advanced prompting. Finally, we identify key future research directions aimed at enhancing the robustness, explainability, and clinical integration of FMs, thereby accelerating their translation into real-world medical practice.

[296] One-step Diffusion Models with Bregman Density Ratio Matching

Yuanzhi Zhu, Eleftherios Tsonis, Lucas Degeorge, Vicky Kalogeiton

Main category: cs.CV

TL;DR: Di-Bregman is a unified framework for diffusion distillation that uses Bregman divergence-based density-ratio matching to accelerate multi-step diffusion models into efficient one-step generators.

DetailsMotivation: Diffusion and flow models achieve high quality but are computationally expensive due to slow multi-step sampling. Existing distillation methods lack a unified theoretical foundation.

Method: Proposes Di-Bregman framework that formulates diffusion distillation as Bregman divergence-based density-ratio matching, providing a convex-analytic view that connects existing objectives.

Result: Experiments on CIFAR-10 and text-to-image generation show improved one-step FID over reverse-KL distillation while maintaining high visual fidelity compared to teacher models.

Conclusion: Bregman density-ratio matching provides a practical and theoretically-grounded approach for efficient one-step diffusion generation.

Abstract: Diffusion and flow models achieve high generative quality but remain computationally expensive due to slow multi-step sampling. Distillation methods accelerate them by training fast student generators, yet most existing objectives lack a unified theoretical foundation. In this work, we propose Di-Bregman, a compact framework that formulates diffusion distillation as Bregman divergence-based density-ratio matching. This convex-analytic view connects several existing objectives through a common lens. Experiments on CIFAR-10 and text-to-image generation demonstrate that Di-Bregman achieves improved one-step FID over reverse-KL distillation and maintains high visual fidelity compared to the teacher model. Our results highlight Bregman density-ratio matching as a practical and theoretically-grounded route toward efficient one-step diffusion generation.

[297] CARE: Contrastive Alignment for ADL Recognition from Event-Triggered Sensor Streams

Junhao Zhao, Zishuai Liu, Ruili Fang, Jin Lu, Linghan Zhang, Fei Dou

Main category: cs.CV

TL;DR: CARE is a framework for ADL recognition that aligns sequence- and image-based representations through contrastive learning to overcome limitations of existing methods.

DetailsMotivation: Existing ADL recognition methods have limitations: sequence-based approaches are sensitive to noise and lack spatial awareness, while image-based approaches compress temporal dynamics and distort sensor layouts. Naive fusion methods fail to properly align these complementary representations.

Method: CARE uses Sequence-Image Contrastive Alignment (SICA) to jointly optimize representation learning and classification. It integrates time-aware sequence encoding with spatially-informed image representations, and employs a joint contrastive-classification objective for end-to-end learning.

Result: CARE achieves state-of-the-art performance on three CASAS datasets: 89.8% on Milan, 88.9% on Cairo, and 73.3% on Kyoto7. It also demonstrates robustness to sensor malfunctions and layout variability.

Conclusion: The CARE framework effectively leverages complementary strengths of sequence- and image-based representations through contrastive alignment, enabling reliable ADL recognition in smart homes with improved performance and robustness.

Abstract: The recognition of Activities of Daily Living (ADLs) from event-triggered ambient sensors is an essential task in Ambient Assisted Living, yet existing methods remain constrained by representation-level limitations. Sequence-based approaches preserve temporal order of sensor activations but are sensitive to noise and lack spatial awareness, while image-based approaches capture global patterns and implicit spatial correlations but compress fine-grained temporal dynamics and distort sensor layouts. Naive fusion (e.g., feature concatenation) fail to enforce alignment between sequence- and image-based representation views, underutilizing their complementary strengths. We propose Contrastive Alignment for ADL Recognition from Event-Triggered Sensor Streams (CARE), an end-to-end framework that jointly optimizes representation learning via Sequence-Image Contrastive Alignment (SICA) and classification via cross-entropy, ensuring both cross-representation alignment and task-specific discriminability. CARE integrates (i) time-aware, noise-resilient sequence encoding with (ii) spatially-informed and frequency-sensitive image representations, and employs (iii) a joint contrastive-classification objective for end-to-end learning of aligned and discriminative embeddings. Evaluated on three CASAS datasets, CARE achieves state-of-the-art performance (89.8% on Milan, 88.9% on Cairo, and 73.3% on Kyoto7) and demonstrates robustness to sensor malfunctions and layout variability, highlighting its potential for reliable ADL recognition in smart homes.

[298] Training-free Online Video Step Grounding

Luca Zanella, Massimiliano Mancini, Yiming Wang, Alessio Tonioni, Elisa Ricci

Main category: cs.CV

TL;DR: This paper introduces BaGLM, a method for Video Step Grounding (VSG) that performs online step detection without training by leveraging Large Multimodal Models (LMMs) and Bayesian filtering principles.

DetailsMotivation: Standard VSG approaches require labeled training data and process full videos offline, which is costly and limits online applications. This work explores performing VSG online without training.

Method: Uses LMMs to predict steps from restricted frame sets, then develops BaGLM which injects knowledge of past frames using Bayesian filtering with step transitions modeled via dependency matrices from LLMs and step progress estimation.

Result: The online strategy without task-specific tuning outperforms offline training-based models. BaGLM shows superior performance over state-of-the-art training-based offline methods on three datasets.

Conclusion: BaGLM enables effective online VSG without training requirements, demonstrating the power of LMMs and Bayesian filtering for step grounding tasks.

Abstract: Given a task and a set of steps composing it, Video Step Grounding (VSG) aims to detect which steps are performed in a video. Standard approaches for this task require a labeled training set (e.g., with step-level annotations or narrations), which may be costly to collect. Moreover, they process the full video offline, limiting their applications for scenarios requiring online decisions. Thus, in this work, we explore how to perform VSG online and without training. We achieve this by exploiting the zero-shot capabilities of recent Large Multimodal Models (LMMs). In particular, we use LMMs to predict the step associated with a restricted set of frames, without access to the whole video. We show that this online strategy without task-specific tuning outperforms offline and training-based models. Motivated by this finding, we develop Bayesian Grounding with Large Multimodal Models (BaGLM), further injecting knowledge of past frames into the LMM-based predictions. BaGLM exploits Bayesian filtering principles, modeling step transitions via (i) a dependency matrix extracted through large language models and (ii) an estimation of step progress. Experiments on three datasets show superior performance of BaGLM over state-of-the-art training-based offline methods.

[299] An empirical study of the effect of video encoders on Temporal Video Grounding

Ignacio M. De la Jara, Cristian Rodriguez-Opazo, Edison Marrese-Taylor, Felipe Bravo-Marquez

Main category: cs.CV

TL;DR: This paper presents an empirical study on how different video feature representations impact temporal video grounding performance, showing that simply changing video encoders (CNN, temporal reasoning, transformers) significantly affects model performance and reveals feature complementarity patterns.

DetailsMotivation: Current temporal video grounding research focuses on a limited selection of video representations, which may lead to architectural overfitting. The authors aim to investigate the impact of different video features on model performance.

Method: Extracted features from three benchmarks (Charades-STA, ActivityNet-Captions, YouCookII) using video encoders based on CNNs, temporal reasoning, and transformers, then evaluated their impact on a classical architecture.

Result: Results show significant performance differences when changing video encoders, revealing clear patterns and errors from using certain features, indicating potential feature complementarity.

Conclusion: Different video feature representations significantly impact temporal video grounding performance, suggesting that exploring diverse feature types could improve model robustness and reveal complementary strengths.

Abstract: Temporal video grounding is a fundamental task in computer vision, aiming to localize a natural language query in a long, untrimmed video. It has a key role in the scientific community, in part due to the large amount of video generated every day. Although we find extensive work in this task, we note that research remains focused on a small selection of video representations, which may lead to architectural overfitting in the long run. To address this issue, we propose an empirical study to investigate the impact of different video features on a classical architecture. We extract features for three well-known benchmarks, Charades-STA, ActivityNet-Captions and YouCookII, using video encoders based on CNNs, temporal reasoning and transformers. Our results show significant differences in the performance of our model by simply changing the video encoder, while also revealing clear patterns and errors derived from the use of certain features, ultimately indicating potential feature complementarity.

[300] Do Satellite Tasks Need Special Pretraining?

Ani Vanyan, Alvard Barseghyan, Hakob Tamazyan, Tigran Galstyan, Vahan Huroyan, Naira Hovakimyan, Hrant Khachatrian

Main category: cs.CV

TL;DR: The paper challenges the need for specialized remote sensing foundation models, showing that general-purpose vision models perform equally well at small scales.

DetailsMotivation: To test whether specialized foundation models for remote sensing provide meaningful advantages over general-purpose vision foundation models.

Method: Created a benchmark to measure generalization to lower resolution images, and trained iBOT on MillionAID dataset with remote sensing-specific modifications.

Result: No consistent improvements were found from specialized pretrained models over general-purpose baselines at ViT-B scale.

Conclusion: Specialized remote sensing foundation models may not be necessary when working at small scales, as general-purpose models perform equally well.

Abstract: Foundation models have advanced machine learning across various modalities, including images. Recently multiple teams trained foundation models specialized for remote sensing applications. This line of research is motivated by the distinct characteristics of remote sensing imagery, specific applications and types of robustness useful for satellite image analysis. In this work we systematically challenge the idea that specific foundation models are more useful than general-purpose vision foundation models, at least in the small scale. First, we design a simple benchmark that measures generalization of remote sensing models towards images with lower resolution for two downstream tasks. Second, we train iBOT, a self-supervised vision encoder, on MillionAID, an ImageNet-scale satellite imagery dataset, with several modifications specific to remote sensing. We show that none of those pretrained models bring consistent improvements upon general-purpose baselines at the ViT-B scale.

[301] Where, Not What: Compelling Video LLMs to Learn Geometric Causality for 3D-Grounding

Yutong Zhong

Main category: cs.CV

TL;DR: W2R2 is a training framework that addresses 2D semantic bias in multimodal 3D grounding by disentangling 2D semantic and 3D spatial features, using dual-objective loss functions for improved localization without changing inference architecture.

DetailsMotivation: Current multimodal 3D grounding models suffer from severe "2D semantic bias" - over-relying on 2D image features for coarse localization while largely ignoring 3D geometric inputs, leading to suboptimal fusion performance.

Method: Proposes What-Where Representation Re-Forming (W2R2) framework with disentangled representation learning: 2D features as semantic beacons for “What” identification and 3D features as spatial anchors for “Where” localization. Uses dual-objective loss with Alignment Loss for multimodal synergy and Pseudo-Label Loss to penalize 2D-dominant outputs.

Result: Experiments on ScanRefer and ScanQA show significant gains in localization accuracy and robustness, particularly in cluttered outdoor scenes.

Conclusion: W2R2 effectively addresses 2D semantic bias through disentangled representation learning and targeted shortcut suppression, enabling precise 3D grounding without architectural modifications.

Abstract: Multimodal 3D grounding has garnered considerable interest in Vision-Language Models (VLMs) \cite{yin2025spatial} for advancing spatial reasoning in complex environments. However, these models suffer from a severe “2D semantic bias” that arises from over-reliance on 2D image features for coarse localization, largely disregarding 3D geometric inputs and resulting in suboptimal fusion performance. In this paper, we propose a novel training framework called What-Where Representation Re-Forming (W2R2) to tackle this issue via disentangled representation learning and targeted shortcut suppression. Our approach fundamentally reshapes the model’s internal space by designating 2D features as semantic beacons for “What” identification and 3D features as spatial anchors for “Where” localization, enabling precise 3D grounding without modifying inference architecture. Key components include a dual-objective loss function with an Alignment Loss that supervises fused predictions using adapted cross-entropy for multimodal synergy, and a Pseudo-Label Loss that penalizes overly effective 2D-dominant pseudo-outputs via a margin-based mechanism. Experiments conducted on ScanRefer and ScanQA demonstrate the effectiveness of W2R2, with significant gains in localization accuracy and robustness, particularly in cluttered outdoor scenes.

[302] Conditional Synthetic Live and Spoof Fingerprint Generation

Syed Konain Abbas, Sandip Purnapatra, M. G. Sarwar Murshed, Conor Miller-Lynch, Lambert Igene, Soumyabrata Dey, Stephanie Schuckers, Faraz Hussain

Main category: cs.CV

TL;DR: This paper presents a novel approach for generating synthetic fingerprint images using conditional StyleGAN2-ADA and StyleGAN3 architectures for live fingerprints and CycleGANs for spoof fingerprints, addressing privacy, cost, and accessibility issues in biometric data collection.

DetailsMotivation: Large fingerprint datasets are time-consuming and expensive to collect with strict privacy requirements. Synthetic fingerprint data can overcome these limitations while enabling development of robust spoof detection systems.

Method: Uses conditional StyleGAN2-ADA and StyleGAN3 to generate high-resolution synthetic live fingerprints conditioned on finger identities, and CycleGANs to translate these into realistic spoof fingerprints simulating various presentation attack materials.

Result: Created two synthetic datasets (DB2 and DB3) with 1,500 fingerprint images each. StyleGAN3 achieved FID as low as 5 and TAR of 99.47% at 0.01% FAR. StyleGAN2-ADA achieved TAR of 98.67%. Quality metrics (NFIQ2, MINDTCT) confirm strong performance with no significant identity leakage.

Conclusion: The synthetic fingerprint generation approach successfully addresses privacy, cost, and accessibility concerns while producing high-quality fingerprints suitable for training and evaluation, with strong privacy-preserving properties confirmed through matching experiments.

Abstract: Large fingerprint datasets, while important for training and evaluation, are time-consuming and expensive to collect and require strict privacy measures. Researchers are exploring the use of synthetic fingerprint data to address these issues. This paper presents a novel approach for generating synthetic fingerprint images (both spoof and live), addressing concerns related to privacy, cost, and accessibility in biometric data collection. Our approach utilizes conditional StyleGAN2-ADA and StyleGAN3 architectures to produce high-resolution synthetic live fingerprints, conditioned on specific finger identities (thumb through little finger). Additionally, we employ CycleGANs to translate these into realistic spoof fingerprints, simulating a variety of presentation attack materials (e.g., EcoFlex, Play-Doh). These synthetic spoof fingerprints are crucial for developing robust spoof detection systems. Through these generative models, we created two synthetic datasets (DB2 and DB3), each containing 1,500 fingerprint images of all ten fingers with multiple impressions per finger, and including corresponding spoofs in eight material types. The results indicate robust performance: our StyleGAN3 model achieves a Fr'echet Inception Distance (FID) as low as 5, and the generated fingerprints achieve a True Accept Rate of 99.47% at a 0.01% False Accept Rate. The StyleGAN2-ADA model achieved a TAR of 98.67% at the same 0.01% FAR. We assess fingerprint quality using standard metrics (NFIQ2, MINDTCT), and notably, matching experiments confirm strong privacy preservation, with no significant evidence of identity leakage, confirming the strong privacy-preserving properties of our synthetic datasets.

[303] Click, Predict, Trust: Clinician-in-the-Loop AI Segmentation for Lung Cancer CT-Based Prognosis within the Knowledge-to-Action Framework

Mohammad R. Salmanpour, Sonya Falahati, Amir Hossein Pouria, Amin Mousavi, Somayeh Sadat Mehrnia, Morteza Alizadeh, Arman Gorji, Zeinab Farsangi, Alireza Safarian, Mehdi Maghsudi, Carlos Uribe, Arman Rahmim, Ren Yuan

Main category: cs.CV

TL;DR: A clinician-in-the-loop deep learning pipeline using VNet with semi-supervised learning achieves accurate, reproducible lung cancer segmentation and prognosis from CT scans, with radiologists preferring AI-generated masks for refinement rather than replacement.

DetailsMotivation: Lung cancer is the leading cause of cancer mortality, and while CT imaging is central to screening and treatment, manual segmentation is variable and time-intensive. Deep learning offers automation but faces barriers to clinical adoption, requiring enhanced reproducibility, prognostic accuracy, and clinical trust.

Method: Used multi-center CT data from 999 patients across 12 public datasets with five DL models (3D Attention U-Net, ResUNet, VNet, ReconNet, SAM-Med3D), benchmarked against expert contours. Assessed segmentation reproducibility using 497 radiomic features and compared supervised vs semi-supervised learning across multiple dimensionality reduction strategies and classifiers. Six physicians qualitatively evaluated masks across clinical domains.

Result: VNet achieved best performance (Dice = 0.83, IoU = 0.71), radiomic stability (mean correlation = 0.76, ICC = 0.65), and predictive accuracy under SSL (accuracy = 0.88, F1 = 0.83). SSL consistently outperformed SL across models. Radiologists favored VNet for peritumoral representation and preferred AI-generated masks for refinement rather than replacement.

Conclusion: Integrating VNet with SSL yields accurate, reproducible, and clinically trusted CT-based lung cancer prognosis, highlighting a feasible path toward physician-centered AI translation with clinician-in-the-loop workflows.

Abstract: Lung cancer remains the leading cause of cancer mortality, with CT imaging central to screening, prognosis, and treatment. Manual segmentation is variable and time-intensive, while deep learning (DL) offers automation but faces barriers to clinical adoption. Guided by the Knowledge-to-Action framework, this study develops a clinician-in-the-loop DL pipeline to enhance reproducibility, prognostic accuracy, and clinical trust. Multi-center CT data from 999 patients across 12 public datasets were analyzed using five DL models (3D Attention U-Net, ResUNet, VNet, ReconNet, SAM-Med3D), benchmarked against expert contours on whole and click-point cropped images. Segmentation reproducibility was assessed using 497 PySERA-extracted radiomic features via Spearman correlation, ICC, Wilcoxon tests, and MANOVA, while prognostic modeling compared supervised (SL) and semi-supervised learning (SSL) across 38 dimensionality reduction strategies and 24 classifiers. Six physicians qualitatively evaluated masks across seven domains, including clinical meaningfulness, boundary quality, prognostic value, trust, and workflow integration. VNet achieved the best performance (Dice = 0.83, IoU = 0.71), radiomic stability (mean correlation = 0.76, ICC = 0.65), and predictive accuracy under SSL (accuracy = 0.88, F1 = 0.83). SSL consistently outperformed SL across models. Radiologists favored VNet for peritumoral representation and smoother boundaries, preferring AI-generated initial masks for refinement rather than replacement. These results demonstrate that integrating VNet with SSL yields accurate, reproducible, and clinically trusted CT-based lung cancer prognosis, highlighting a feasible path toward physician-centered AI translation.

[304] Video Reasoning without Training

Deepak Sridhar, Kartikeya Bhardwaj, Jeya Pradha Jeyaraj, Nuno Vasconcelos, Ankita Nayak, Harris Teague

Main category: cs.CV

TL;DR: V-Reason improves video reasoning in LMMs by using entropy signals to optimize exploration-exploitation behavior during inference, achieving near-RL performance without training while reducing computational costs.

DetailsMotivation: Current video reasoning methods using LMMs rely on expensive RL training and verbose chain-of-thought, causing computational overhead and limited control over the thinking process.

Method: Uses entropy of model outputs as signal to guide micro-exploration/exploitation. Adapts LMM’s value cache during inference via a small trainable controller with entropy-based objective, requiring no dataset supervision or RL.

Result: Significant improvements over base models across video reasoning datasets, narrowing gap with RL-trained models to within 0.6% average accuracy. Reduces output tokens by 58.6% compared to RL models.

Conclusion: V-Reason enables efficient video reasoning by optimizing inference-time behavior using entropy signals, achieving competitive performance without costly training while offering substantial efficiency benefits.

Abstract: Video reasoning using Large Multimodal Models (LMMs) relies on costly reinforcement learning (RL) and verbose chain-of-thought, resulting in substantial computational overhead during both training and inference. Moreover, the mechanisms that control the thinking process in these reasoning models are very limited. In this paper, using entropy of the model’s output as a signal, we discover that the high-quality models go through a series of micro-explorations and micro-exploitations which keep the reasoning process grounded (i.e., avoid excessive randomness while the model is exploring or thinking through an answer). We further observe that once this “thinking” process is over, more accurate models demonstrate a better convergence by reducing the entropy significantly via a final exploitation phase (i.e., a more certain convergence towards a solution trajectory). We then use these novel, theoretically-grounded insights to tune the model’s behavior directly at inference, without using any RL or supervised fine-tuning. Specifically, during inference, our proposed approach called V-Reason (Video-Reason) adapts the value cache of the LMM via a few optimization steps on a small, trainable controller using an entropy-based objective, i.e., no supervision from any dataset or RL is necessary. This tuning improves the model’s micro-exploration and exploitation behavior during inference. Our experiments show that our proposed method achieves significant improvements over the base instruction-tuned models across several video reasoning datasets, narrowing the gap with RL-trained models to within 0.6% average accuracy without any training, while offering massive efficiency benefits: output tokens are reduced by 58.6% compared to the RL model.

[305] How Universal Are SAM2 Features?

Masoud Khairi Atani, Alon Harell, Hyomin Choi, Runyu Yang, Fabien Racape, Ivan V. Bajic

Main category: cs.CV

TL;DR: This paper investigates the trade-off between general-purpose foundation models (Hiera) and specialized models (SAM2) for vision tasks, quantifying the information-theoretic cost of specialization and revealing performance differences across task domains.

DetailsMotivation: To understand the trade-off between general-purpose foundation vision models and specialized counterparts for efficient feature coding design, as this trade-off is not yet fully understood.

Method: Compare feature versatility of Hiera encoder vs SAM2 using lightweight trainable neck to probe adaptability of frozen features, quantify information-theoretic cost of specialization, and conduct cross-neck analysis on SAM2.

Result: SAM2’s specialization is highly effective for spatially-related tasks like depth estimation but underperforms Hiera on conceptually distant tasks (pose estimation, image captioning), showing measurable loss of broader semantic information. Cross-neck analysis reveals each adaptation level creates further representational bottlenecks.

Conclusion: The analysis illuminates trade-offs in feature universality, providing quantitative foundation for designing efficient feature coding and adaptation strategies for diverse downstream applications.

Abstract: The trade-off between general-purpose foundation vision models and their specialized counterparts is critical for efficient feature coding design and is not yet fully understood. We investigate this trade-off by comparing the feature versatility of the general-purpose Hiera encoder against the segmentation-specialized Segment Anything Model 2 (SAM2). Using a lightweight, trainable neck to probe the adaptability of their frozen features, we quantify the information-theoretic cost of specialization. Our results reveal that while SAM2’s specialization is highly effective for spatially-related tasks like depth estimation, it comes at a cost. The specialized SAM2 encoder underperforms its generalist predecessor, Hiera, on conceptually distant tasks such as pose estimation and image captioning, demonstrating a measurable loss of broader semantic information. A novel cross-neck analysis on SAM2 reveals that each level of adaptation creates a further representational bottleneck. Our analysis illuminates these trade-offs in feature universality, providing a quantitative foundation for designing efficient feature coding and adaptation strategies for diverse downstream applications.

[306] ProDAT: Progressive Density-Aware Tail-Drop for Point Cloud Coding

Zhe Luo, Wenjing Jia, Stuart Perry

Main category: cs.CV

TL;DR: ProDAT enables progressive point cloud coding with density-aware tail-drop mechanism, achieving superior compression efficiency with single-model multi-bitrate support.

DetailsMotivation: 3D point clouds require real-time processing but face bandwidth constraints; existing learning-based methods lack progressive decoding capability.

Method: Proposed ProDAT with density-aware tail-drop mechanism that adaptively decodes latent features and coordinates based on significance using density guidance.

Result: Achieved over 28.6% BD-rate improvement for PSNR-D2 on SemanticKITTI and over 18.15% on ShapeNet compared to state-of-the-art methods.

Conclusion: ProDAT successfully bridges the progressive coding gap while maintaining superior compression performance across multiple datasets.

Abstract: Three-dimensional (3D) point clouds are becoming increasingly vital in applications such as autonomous driving, augmented reality, and immersive communication, demanding real-time processing and low latency. However, their large data volumes and bandwidth constraints hinder the deployment of high-quality services in resource-limited environments. Progres- sive coding, which allows for decoding at varying levels of detail, provides an alternative by allowing initial partial decoding with subsequent refinement. Although recent learning-based point cloud geometry coding methods have achieved notable success, their fixed latent representation does not support progressive decoding. To bridge this gap, we propose ProDAT, a novel density-aware tail-drop mechanism for progressive point cloud coding. By leveraging density information as a guidance signal, latent features and coordinates are decoded adaptively based on their significance, therefore achieving progressive decoding at multiple bitrates using one single model. Experimental results on benchmark datasets show that the proposed ProDAT not only enables progressive coding but also achieves superior coding efficiency compared to state-of-the-art learning-based coding techniques, with over 28.6% BD-rate improvement for PSNR- D2 on SemanticKITTI and over 18.15% for ShapeNet

[307] Towards a Generalizable Fusion Architecture for Multimodal Object Detection

Jad Berjawi, Yoann Dupas, Christophe C’erin

Main category: cs.CV

TL;DR: FMCAF is a preprocessing architecture that enhances RGB-IR fusion using frequency filtering and cross-attention, improving multimodal object detection performance across different datasets without dataset-specific tuning.

DetailsMotivation: To improve robustness in challenging conditions by leveraging complementary cues from multiple sensor modalities, particularly RGB and infrared inputs, for better multimodal object detection.

Method: Combines frequency-domain filtering (Freq-Filter) to suppress redundant spectral features with cross-attention-based fusion (MCAF) to improve intermodal feature sharing.

Result: Outperforms traditional concatenation fusion, achieving +13.9% mAP@50 on VEDAI and +1.1% on LLVIP datasets.

Conclusion: FMCAF shows potential as a flexible foundation for robust multimodal fusion in future detection pipelines, demonstrating generalizability across different multimodal challenges.

Abstract: Multimodal object detection improves robustness in chal- lenging conditions by leveraging complementary cues from multiple sensor modalities. We introduce Filtered Multi- Modal Cross Attention Fusion (FMCAF), a preprocess- ing architecture designed to enhance the fusion of RGB and infrared (IR) inputs. FMCAF combines a frequency- domain filtering block (Freq-Filter) to suppress redun- dant spectral features with a cross-attention-based fusion module (MCAF) to improve intermodal feature sharing. Unlike approaches tailored to specific datasets, FMCAF aims for generalizability, improving performance across different multimodal challenges without requiring dataset- specific tuning. On LLVIP (low-light pedestrian detec- tion) and VEDAI (aerial vehicle detection), FMCAF outper- forms traditional fusion (concatenation), achieving +13.9% mAP@50 on VEDAI and +1.1% on LLVIP. These results support the potential of FMCAF as a flexible foundation for robust multimodal fusion in future detection pipelines.

[308] GSPlane: Concise and Accurate Planar Reconstruction via Structured Representation

Ruitong Gan, Junran Peng, Yang Liu, Chuanchen Luo, Qing Li, Zhaoxiang Zhang

Main category: cs.CV

TL;DR: GSPlane enhances Gaussian Splatting by incorporating planar priors to improve geometry reconstruction and mesh quality for planar regions in 3D scenes.

DetailsMotivation: Existing Gaussian Splatting methods struggle with reconstructing smooth and precise planar regions, which are fundamental in man-made environments for scene editing and physical simulations.

Method: Leverages segmentation and normal prediction models to extract planar priors, uses structured representations for planar Gaussian coordinates, introduces Dynamic Gaussian Re-classifier to handle high-gradient regions, and refines mesh layouts using optimized planar priors.

Result: Significantly improves geometric accuracy of extracted meshes without sacrificing rendering quality, produces clean mesh connectivity with reduced vertices/faces, and enables object decoupling and manipulation on supportive planes.

Conclusion: GSPlane effectively addresses planar reconstruction limitations in Gaussian Splatting through structured planar priors, enhancing both geometric accuracy and mesh topology while maintaining rendering performance.

Abstract: Planes are fundamental primitives of 3D sences, especially in man-made environments such as indoor spaces and urban streets. Representing these planes in a structured and parameterized format facilitates scene editing and physical simulations in downstream applications. Recently, Gaussian Splatting (GS) has demonstrated remarkable effectiveness in the Novel View Synthesis task, with extensions showing great potential in accurate surface reconstruction. However, even state-of-the-art GS representations often struggle to reconstruct planar regions with sufficient smoothness and precision. To address this issue, we propose GSPlane, which recovers accurate geometry and produces clean and well-structured mesh connectivity for plane regions in the reconstructed scene. By leveraging off-the-shelf segmentation and normal prediction models, GSPlane extracts robust planar priors to establish structured representations for planar Gaussian coordinates, which help guide the training process by enforcing geometric consistency. To further enhance training robustness, a Dynamic Gaussian Re-classifier is introduced to adaptively reclassify planar Gaussians with persistently high gradients as non-planar, ensuring more reliable optimization. Furthermore, we utilize the optimized planar priors to refine the mesh layouts, significantly improving topological structure while reducing the number of vertices and faces. We also explore applications of the structured planar representation, which enable decoupling and flexible manipulation of objects on supportive planes. Extensive experiments demonstrate that, with no sacrifice in rendering quality, the introduction of planar priors significantly improves the geometric accuracy of the extracted meshes across various baselines.

[309] Boosting Fidelity for Pre-Trained-Diffusion-Based Low-Light Image Enhancement via Condition Refinement

Xiaogang Xu, Jian Wang, Yunfan Lu, Ruihang Chu, Ruixing Wang, Jiafei Wu, Bei Yu, Liang Lin

Main category: cs.CV

TL;DR: Proposes a novel optimization strategy for pre-trained diffusion models to enhance fidelity in low-level vision tasks, particularly in low-light scenarios, while preserving realism and aesthetics.

DetailsMotivation: Pre-Trained Diffusion-Based (PTDB) methods often sacrifice content fidelity for perceptual realism, especially in low-light scenarios where degraded information limits effective control. Two main causes identified: absence of suitable conditional latent modeling and lack of bidirectional interaction between conditional and noisy latents.

Method: Introduces a latent refinement pipeline to recover spatial details lost during VAE encoding, incorporating generative priors. The refined latent condition dynamically interacts with the noisy latent in the diffusion process. The approach is plug-and-play and integrates into existing diffusion networks.

Result: Extensive experiments demonstrate significant fidelity improvements in PTDB methods while maintaining realism and aesthetics.

Conclusion: The proposed optimization strategy effectively addresses fidelity loss in diffusion-based methods by enabling better conditional latent modeling and bidirectional interaction, providing more effective control in low-level vision tasks.

Abstract: Diffusion-based methods, leveraging pre-trained large models like Stable Diffusion via ControlNet, have achieved remarkable performance in several low-level vision tasks. However, Pre-Trained Diffusion-Based (PTDB) methods often sacrifice content fidelity to attain higher perceptual realism. This issue is exacerbated in low-light scenarios, where severely degraded information caused by the darkness limits effective control. We identify two primary causes of fidelity loss: the absence of suitable conditional latent modeling and the lack of bidirectional interaction between the conditional latent and noisy latent in the diffusion process. To address this, we propose a novel optimization strategy for conditioning in pre-trained diffusion models, enhancing fidelity while preserving realism and aesthetics. Our method introduces a mechanism to recover spatial details lost during VAE encoding, i.e., a latent refinement pipeline incorporating generative priors. Additionally, the refined latent condition interacts dynamically with the noisy latent, leading to improved restoration performance. Our approach is plug-and-play, seamlessly integrating into existing diffusion networks to provide more effective control. Extensive experiments demonstrate significant fidelity improvements in PTDB methods.

[310] Towards Imperceptible Watermarking Via Environment Illumination for Consumer Cameras

Hodaka Kawachi, Tomoya Nakamura, Hiroaki Santo, SaiKiran Kumar Tedla, Trevor Dalton Canham, Yasushi Yagi, Michael S. Brown

Main category: cs.CV

TL;DR: A method using LED lighting to create invisible watermarks for cameras by optimizing spectral profiles that are undetectable to humans but visible to consumer cameras, enabling metadata embedding for privacy and verification.

DetailsMotivation: To develop an imperceptible watermarking system that leverages environmental lighting to embed metadata in video content for privacy protection and content verification without being noticeable to human observers.

Method: Optimizes LED spectral profiles using spectral modulation (not intensity modulation) that considers human visual sensitivity, camera sensor characteristics, and LED capabilities to produce white light while embedding detectable watermarks.

Result: Successfully embeds 128 bits within 10-second video clips at standard frame rates (30-60 fps), providing sufficient capacity for essential metadata while maintaining visual imperceptibility.

Conclusion: The approach enables practical invisible watermarking through environmental lighting, supporting privacy protection and content verification applications with modest but adequate data transfer rates.

Abstract: This paper introduces a method for using LED-based environmental lighting to produce visually imperceptible watermarks for consumer cameras. Our approach optimizes an LED light source’s spectral profile to be minimally visible to the human eye while remaining highly detectable by typical consumer cameras. The method jointly considers the human visual system’s sensitivity to visible spectra, modern consumer camera sensors’ spectral sensitivity, and narrowband LEDs’ ability to generate broadband spectra perceived as “white light” (specifically, D65 illumination). To ensure imperceptibility, we employ spectral modulation rather than intensity modulation. Unlike conventional visible light communication, our approach enables watermark extraction at standard low frame rates (30-60 fps). While the information transfer rate is modest-embedding 128 bits within a 10-second video clip-this capacity is sufficient for essential metadata supporting privacy protection and content verification.

[311] GOOD: Training-Free Guided Diffusion Sampling for Out-of-Distribution Detection

Xin Gao, Jiyao Liu, Guanghao Li, Yueming Lyu, Jianxiong Gao, Weichen Yu, Ningsheng Xu, Liang Wang, Caifeng Shan, Ziwei Liu, Chenyang Si

Main category: cs.CV

TL;DR: GOOD is a framework that uses dual-level guidance (image-level and feature-level) with diffusion models to generate diverse out-of-distribution samples, improving OOD detection performance.

DetailsMotivation: Existing methods for generating OOD samples using text-to-image diffusion models suffer from semantic instability and insufficient shift diversity due to text-embedding perturbations, limiting generalization to realistic OOD scenarios.

Method: GOOD guides diffusion sampling trajectories using ID classifiers with dual-level guidance: image-level guidance reduces input likelihood via gradient of log partition, and feature-level guidance promotes sampling in feature-sparse regions using k-NN distance in classifier’s latent space.

Result: Training with GOOD-generated samples notably enhances OOD detection performance, as demonstrated through thorough quantitative and qualitative analyses.

Conclusion: GOOD enables more controllable and diverse OOD sample generation through its dual-guidance design and unified OOD scoring, substantially improving OOD detection capabilities.

Abstract: Recent advancements have explored text-to-image diffusion models for synthesizing out-of-distribution (OOD) samples, substantially enhancing the performance of OOD detection. However, existing approaches typically rely on perturbing text-conditioned embeddings, resulting in semantic instability and insufficient shift diversity, which limit generalization to realistic OOD. To address these challenges, we propose GOOD, a novel and flexible framework that directly guides diffusion sampling trajectories towards OOD regions using off-the-shelf in-distribution (ID) classifiers. GOOD incorporates dual-level guidance: (1) Image-level guidance based on the gradient of log partition to reduce input likelihood, drives samples toward low-density regions in pixel space. (2) Feature-level guidance, derived from k-NN distance in the classifier’s latent space, promotes sampling in feature-sparse regions. Hence, this dual-guidance design enables more controllable and diverse OOD sample generation. Additionally, we introduce a unified OOD score that adaptively combines image and feature discrepancies, enhancing detection robustness. We perform thorough quantitative and qualitative analyses to evaluate the effectiveness of GOOD, demonstrating that training with samples generated by GOOD can notably enhance OOD detection performance.

[312] KineDiff3D: Kinematic-Aware Diffusion for Category-Level Articulated Object Shape Reconstruction and Generation

WenBo Xu, Liu Liu, Li Zhang, Ran Zhang, Hao Wu, Dan Guo, Meng Wang

Main category: cs.CV

TL;DR: KineDiff3D is a unified framework for reconstructing articulated objects and estimating poses from single-view inputs using kinematic-aware diffusion models and iterative optimization.

DetailsMotivation: Articulated objects like laptops and drawers pose challenges for 3D reconstruction due to their multi-part geometries and variable joint configurations, which create structural diversity across different states.

Method: The framework uses a Kinematic-Aware VAE to encode geometry, joint angles, and part segmentation into a structured latent space, then employs two conditional diffusion models for pose/joint parameter regression and kinematic-aware latent code generation, followed by iterative optimization with Chamfer-distance minimization.

Result: Experimental results on synthetic, semi-synthetic, and real-world datasets show the approach effectively reconstructs articulated objects and estimates their kinematic properties.

Conclusion: KineDiff3D provides an effective solution for category-level articulated object reconstruction and pose estimation from single-view inputs while preserving articulation constraints.

Abstract: Articulated objects, such as laptops and drawers, exhibit significant challenges for 3D reconstruction and pose estimation due to their multi-part geometries and variable joint configurations, which introduce structural diversity across different states. To address these challenges, we propose KineDiff3D: Kinematic-Aware Diffusion for Category-Level Articulated Object Shape Reconstruction and Generation, a unified framework for reconstructing diverse articulated instances and pose estimation from single view input. Specifically, we first encode complete geometry (SDFs), joint angles, and part segmentation into a structured latent space via a novel Kinematic-Aware VAE (KA-VAE). In addition, we employ two conditional diffusion models: one for regressing global pose (SE(3)) and joint parameters, and another for generating the kinematic-aware latent code from partial observations. Finally, we produce an iterative optimization module that bidirectionally refines reconstruction accuracy and kinematic parameters via Chamfer-distance minimization while preserving articulation constraints. Experimental results on synthetic, semi-synthetic, and real-world datasets demonstrate the effectiveness of our approach in accurately reconstructing articulated objects and estimating their kinematic properties.

[313] GACO-CAD: Geometry-Augmented and Conciseness-Optimized CAD Model Generation from Single Image

Yinghui Wang, Xinyu Zhang, Peng Du

Main category: cs.CV

TL;DR: GACO-CAD is a two-stage post-training framework that improves geometric accuracy and modeling conciseness in generating parametric CAD models from single images using MLLMs.

DetailsMotivation: Current MLLMs struggle with accurate 3D geometry inference from 2D images due to limited spatial reasoning capabilities, which hinders their application in industrial concept design.

Method: Two-stage framework: 1) Supervised fine-tuning using depth and surface normal maps as geometric priors combined with RGB images, 2) Reinforcement learning with group length reward to promote compact parametric modeling sequences.

Result: State-of-the-art performance on DeepCAD and Fusion360 datasets, outperforming existing methods in code validity, geometric accuracy, and modeling conciseness.

Conclusion: GACO-CAD effectively addresses spatial reasoning limitations in MLLMs for CAD generation, achieving both high geometric fidelity and modeling efficiency through complementary geometric priors and compact sequence optimization.

Abstract: Generating editable, parametric CAD models from a single image holds great potential to lower the barriers of industrial concept design. However, current multi-modal large language models (MLLMs) still struggle with accurately inferring 3D geometry from 2D images due to limited spatial reasoning capabilities. We address this limitation by introducing GACO-CAD, a novel two-stage post-training framework. It is designed to achieve a joint objective: simultaneously improving the geometric accuracy of the generated CAD models and encouraging the use of more concise modeling procedures. First, during supervised fine-tuning, we leverage depth and surface normal maps as dense geometric priors, combining them with the RGB image to form a multi-channel input. In the context of single-view reconstruction, these priors provide complementary spatial cues that help the MLLM more reliably recover 3D geometry from 2D observations. Second, during reinforcement learning, we introduce a group length reward that, while preserving high geometric fidelity, promotes the generation of more compact and less redundant parametric modeling sequences. A simple dynamic weighting strategy is adopted to stabilize training. Experiments on the DeepCAD and Fusion360 datasets show that GACO-CAD achieves state-of-the-art performance under the same MLLM backbone, consistently outperforming existing methods in terms of code validity, geometric accuracy, and modeling conciseness.

[314] Investigating Adversarial Robustness against Preprocessing used in Blackbox Face Recognition

Roland Croft, Brian Du, Darcy Joseph, Sharath Kumar

Main category: cs.CV

TL;DR: Face recognition systems are vulnerable to adversarial attacks, but face preprocessing techniques significantly impact attack success rates. Different face detection models can reduce attack effectiveness by up to 78%, while preprocessing-invariant methods can improve transferability by 27%.

DetailsMotivation: To investigate how face preprocessing in FR systems affects adversarial attack transferability in blackbox settings, as preprocessing is often overlooked but plays a critical role in FR security.

Method: Studied transferability of state-of-the-art adversarial attacks against different preprocessing techniques, analyzed impact of face detection models and interpolation methods, and proposed preprocessing-invariant method using input transformations.

Result: Face detection model choice degrades attack success rate by up to 78%, while interpolation methods have minimal impact. Preprocessing degrades attack strength even in whitebox settings. The proposed preprocessing-invariant method improves transferability by up to 27%.

Conclusion: Face preprocessing is crucial in FR systems and must be considered to improve adversarial generalization of facial adversarial examples.

Abstract: Face Recognition (FR) models have been shown to be vulnerable to adversarial examples that subtly alter benign facial images, exposing blind spots in these systems, as well as protecting user privacy. End-to-end FR systems first obtain preprocessed faces from diverse facial imagery prior to computing the similarity of the deep feature embeddings. Whilst face preprocessing is a critical component of FR systems, and hence adversarial attacks against them, we observe that this preprocessing is often overlooked in blackbox settings. Our study seeks to investigate the transferability of several out-of-the-box state-of-the-art adversarial attacks against FR when applied against different preprocessing techniques used in a blackbox setting. We observe that the choice of face detection model can degrade the attack success rate by up to 78%, whereas choice of interpolation method during downsampling has relatively minimal impacts. Furthermore, we find that the requirement for facial preprocessing even degrades attack strength in a whitebox setting, due to the unintended interaction of produced noise vectors against face detection models. Based on these findings, we propose a preprocessing-invariant method using input transformations that improves the transferability of the studied attacks by up to 27%. Our findings highlight the importance of preprocessing in FR systems, and the need for its consideration towards improving the adversarial generalisation of facial adversarial examples.

[315] Generation then Reconstruction: Accelerating Masked Autoregressive Models via Two-Stage Sampling

Feihong Yan, Peiru Wang, Yao Zhu, Kaiyu Pang, Qingyan Wei, Huiqi Li, Linfeng Zhang

Main category: cs.CV

TL;DR: GtR is a training-free hierarchical sampling strategy that accelerates masked autoregressive models by decomposing generation into structure generation and detail reconstruction stages, achieving 3.72x speedup while maintaining quality.

DetailsMotivation: Masked autoregressive models have constrained acceleration potential due to modeling complexity of spatially correlated visual tokens in single step generation.

Method: Two-stage approach: structure generation for global semantic scaffolding followed by detail reconstruction for completing remaining tokens, plus Frequency-Weighted Token Selection to allocate more computation to detail tokens based on high frequency energy.

Result: Achieves 3.72x speedup on MAR-H while maintaining comparable quality (FID: 1.59, IS: 304.4 vs. original 1.59, 299.1), outperforming existing acceleration methods across various model scales and generation tasks.

Conclusion: GtR effectively accelerates masked autoregressive models through hierarchical generation strategy and intelligent token selection, demonstrating significant speed improvements without quality degradation.

Abstract: Masked Autoregressive (MAR) models promise better efficiency in visual generation than autoregressive (AR) models for the ability of parallel generation, yet their acceleration potential remains constrained by the modeling complexity of spatially correlated visual tokens in a single step. To address this limitation, we introduce Generation then Reconstruction (GtR), a training-free hierarchical sampling strategy that decomposes generation into two stages: structure generation establishing global semantic scaffolding, followed by detail reconstruction efficiently completing remaining tokens. Assuming that it is more difficult to create an image from scratch than to complement images based on a basic image framework, GtR is designed to achieve acceleration by computing the reconstruction stage quickly while maintaining the generation quality by computing the generation stage slowly. Moreover, observing that tokens on the details of an image often carry more semantic information than tokens in the salient regions, we further propose Frequency-Weighted Token Selection (FTS) to offer more computation budget to tokens on image details, which are localized based on the energy of high frequency information. Extensive experiments on ImageNet class-conditional and text-to-image generation demonstrate 3.72x speedup on MAR-H while maintaining comparable quality (e.g., FID: 1.59, IS: 304.4 vs. original 1.59, 299.1), substantially outperforming existing acceleration methods across various model scales and generation tasks. Our codes will be released in https://github.com/feihongyan1/GtR.

[316] Benchmarking Out-of-Distribution Detection for Plankton Recognition: A Systematic Evaluation of Advanced Methods in Marine Ecological Monitoring

Yingzi Han, Jiakai He, Chuanlong Xie, Jianping Li

Main category: cs.CV

TL;DR: This paper presents the first large-scale systematic evaluation of Out-of-Distribution (OoD) detection methods for plankton recognition, identifying ViM as the best-performing approach.

DetailsMotivation: Plankton recognition models face challenges from distribution shifts due to complex morphologies, species diversity, and novel species discovery, leading to unpredictable errors. The field lacks systematic integration of latest computer vision developments and unified benchmarks.

Method: Created OoD benchmarks simulating various distribution shift scenarios using DYB-PlanktonNet dataset, and systematically evaluated twenty-two OoD detection methods.

Result: ViM method significantly outperforms other approaches, particularly excelling in Far-OoD scenarios with substantial improvements in key metrics.

Conclusion: This comprehensive evaluation provides reliable reference for algorithm selection in automated plankton recognition and lays foundation for future research in plankton OoD detection.

Abstract: Automated plankton recognition models face significant challenges during real-world deployment due to distribution shifts (Out-of-Distribution, OoD) between training and test data. This stems from plankton’s complex morphologies, vast species diversity, and the continuous discovery of novel species, which leads to unpredictable errors during inference. Despite rapid advancements in OoD detection methods in recent years, the field of plankton recognition still lacks a systematic integration of the latest computer vision developments and a unified benchmark for large-scale evaluation. To address this, this paper meticulously designed a series of OoD benchmarks simulating various distribution shift scenarios based on the DYB-PlanktonNet dataset \cite{875n-f104-21}, and systematically evaluated twenty-two OoD detection methods. Extensive experimental results demonstrate that the ViM \cite{wang2022vim} method significantly outperforms other approaches in our constructed benchmarks, particularly excelling in Far-OoD scenarios with substantial improvements in key metrics. This comprehensive evaluation not only provides a reliable reference for algorithm selection in automated plankton recognition but also lays a solid foundation for future research in plankton OoD detection. To our knowledge, this study marks the first large-scale, systematic evaluation and analysis of Out-of-Distribution data detection methods in plankton recognition. Code is available at https://github.com/BlackJack0083/PlanktonOoD.

[317] Capturing Head Avatar with Hand Contacts from a Monocular Video

Haonan He, Yufeng Zheng, Jie Song

Main category: cs.CV

TL;DR: A framework that jointly learns photorealistic 3D head avatars and non-rigid facial deformations caused by hand-face interactions, addressing challenges in pose tracking and deformation learning.

DetailsMotivation: Most existing methods focus only on facial regions and ignore natural hand-face interactions, which are important for conveying cognitive states like pondering in applications such as telepresence, gaming, and VR.

Method: Proposes depth order loss with contact regularization for pose tracking, learns a PCA basis for hand-induced facial deformations from a face-hand interaction dataset, and incorporates contact loss inspired by physics-based simulation to reduce interpenetration artifacts.

Result: Evaluated on RGB(D) videos from iPhone and synthetic datasets, showing better appearance and more accurate deforming geometry of the face compared to state-of-the-art surface reconstruction methods.

Conclusion: The method successfully captures detailed head avatars with physically plausible hand-induced facial deformations, overcoming challenges in pose tracking and deformation learning from monocular videos.

Abstract: Photorealistic 3D head avatars are vital for telepresence, gaming, and VR. However, most methods focus solely on facial regions, ignoring natural hand-face interactions, such as a hand resting on the chin or fingers gently touching the cheek, which convey cognitive states like pondering. In this work, we present a novel framework that jointly learns detailed head avatars and the non-rigid deformations induced by hand-face interactions. There are two principal challenges in this task. First, naively tracking hand and face separately fails to capture their relative poses. To overcome this, we propose to combine depth order loss with contact regularization during pose tracking, ensuring correct spatial relationships between the face and hand. Second, no publicly available priors exist for hand-induced deformations, making them non-trivial to learn from monocular videos. To address this, we learn a PCA basis specific to hand-induced facial deformations from a face-hand interaction dataset. This reduces the problem to estimating a compact set of PCA parameters rather than a full spatial deformation field. Furthermore, inspired by physics-based simulation, we incorporate a contact loss that provides additional supervision, significantly reducing interpenetration artifacts and enhancing the physical plausibility of the results. We evaluate our approach on RGB(D) videos captured by an iPhone. Additionally, to better evaluate the reconstructed geometry, we construct a synthetic dataset of avatars with various types of hand interactions. We show that our method can capture better appearance and more accurate deforming geometry of the face than SOTA surface reconstruction methods.

[318] HIDISC: A Hyperbolic Framework for Domain Generalization with Generalized Category Discovery

Vaibhav Rathore, Divyam Gupta, Biplab Banerjee

Main category: cs.CV

TL;DR: HIDISC is a hyperbolic representation learning framework for Domain Generalization with Generalized Category Discovery (DG-GCD) that achieves domain and category-level generalization without episodic training, using GPT-guided diffusion for domain augmentation and Tangent CutMix for curvature-aware interpolation.

DetailsMotivation: Existing GCD methods assume simultaneous access to labeled and unlabeled data from the same domain, limiting applicability in open-world scenarios with distribution shifts. DG-GCD aims to generalize to unseen domains containing novel categories without accessing target domain data during training.

Method: Uses hyperbolic representation learning with GPT-guided diffusion for domain augmentation, Tangent CutMix for curvature-aware interpolation in tangent space, and a unified loss combining penalized Busemann alignment, hybrid hyperbolic contrastive regularization, and adaptive outlier repulsion. Includes learnable curvature parameter.

Result: Achieves state-of-the-art results on PACS, Office-Home, and DomainNet datasets, consistently outperforming existing Euclidean and hyperbolic (DG)-GCD baselines.

Conclusion: HIDISC provides an efficient and effective framework for DG-GCD that avoids computational costs of episodic training while achieving superior generalization across domains and categories through hyperbolic geometry and novel augmentation techniques.

Abstract: Generalized Category Discovery (GCD) aims to classify test-time samples into either seen categories** – available during training – or novel ones, without relying on label supervision. Most existing GCD methods assume simultaneous access to labeled and unlabeled data during training and arising from the same domain, limiting applicability in open-world scenarios involving distribution shifts. Domain Generalization with GCD (DG-GCD) lifts this constraint by requiring models to generalize to unseen domains containing novel categories, without accessing targetdomain data during training. The only prior DG-GCD method, DG2CD-Net, relies on episodic training with multiple synthetic domains and task vector aggregation, incurring high computational cost and error accumulation. We propose HIDISC, a hyperbolic representation learning framework that achieves domain and category-level generalization without episodic simulation. To expose the model to minimal but diverse domain variations, we augment the source domain using GPT-guided diffusion, avoiding overfitting while maintaining efficiency. To structure the representation space, we introduce Tangent CutMix, a curvature-aware interpolation that synthesizes pseudo-novel samples in tangent space, preserving manifold consistency. A unified loss – combining penalized Busemann alignment, hybrid hyperbolic contrastive regularization, and adaptive outlier repulsion – **facilitates compact, semantically structured embeddings. A learnable curvature parameter further adapts the geometry to dataset complexity. HIDISC achieves state-of-the-art results on PACS , Office-Home , and DomainNet, consistently outperforming the existing Euclidean and hyperbolic (DG)-GCD baselines.

[319] $\mathcal{V}isi\mathcal{P}runer$: Decoding Discontinuous Cross-Modal Dynamics for Efficient Multimodal LLMs

Yingqi Fan, Anhao Zhao, Jinlan Fu, Junlong Tong, Hui Su, Yijie Pan, Wei Zhang, Xiaoyu Shen

Main category: cs.CV

TL;DR: VisiPruner is a training-free pruning framework that reduces vision-related attention computations in MLLMs by up to 99% and FLOPs by 53.9%, based on insights about MLLMs’ three-stage cross-modal interaction process.

DetailsMotivation: MLLMs suffer from significant computational overhead due to quadratic attention growth with multimodal tokens, and existing pruning methods lack understanding of how MLLMs process and fuse multimodal information.

Method: Systematic analysis revealed a three-stage cross-modal interaction: shallow layers recognize task intent, middle layers fuse cross-modal information, and deep layers focus on linguistic refinement. Based on this, VisiPruner prunes vision tokens at appropriate stages without training.

Result: VisiPruner reduces up to 99% of vision-related attention computations and 53.9% of FLOPs on LLaVA-v1.5 7B, outperforming existing token pruning methods and generalizing across diverse MLLMs.

Conclusion: The insights provide actionable guidelines for training efficient MLLMs by aligning model architecture with intrinsic layer-wise processing dynamics, and VisiPruner offers an effective training-free pruning solution.

Abstract: Multimodal Large Language Models (MLLMs) have achieved strong performance across vision-language tasks, but suffer from significant computational overhead due to the quadratic growth of attention computations with the number of multimodal tokens. Though efforts have been made to prune tokens in MLLMs, \textit{they lack a fundamental understanding of how MLLMs process and fuse multimodal information.} Through systematic analysis, we uncover a \textbf{three-stage} cross-modal interaction process: (1) Shallow layers recognize task intent, with visual tokens acting as passive attention sinks; (2) Cross-modal fusion occurs abruptly in middle layers, driven by a few critical visual tokens; (3) Deep layers discard vision tokens, focusing solely on linguistic refinement. Based on these findings, we propose \emph{VisiPruner}, a training-free pruning framework that reduces up to 99% of vision-related attention computations and 53.9% of FLOPs on LLaVA-v1.5 7B. It significantly outperforms existing token pruning methods and generalizes across diverse MLLMs. Beyond pruning, our insights further provide actionable guidelines for training efficient MLLMs by aligning model architecture with its intrinsic layer-wise processing dynamics. Our code is available at: https://github.com/EIT-NLP/VisiPruner.

[320] ZSPAPrune: Zero-Shot Prompt-Aware Token Pruning for Vision-Language Models

Pu Zhang, Yuwei Li, Xingyuan Xian, Guoming Tang

Main category: cs.CV

TL;DR: A zero-shot method for reducing visual token redundancy in VLMs by balancing task relevance and information diversity through hierarchical token pruning.

DetailsMotivation: Vision-Language Models face prohibitive inference costs due to visual token redundancy, and existing pruning methods neglect text prompt guidance, failing to prioritize task relevance.

Method: Hierarchical approach that first selects task-relevant visual tokens based on prompt guidance, then supplements with diversity tokens to preserve broader context.

Result: Achieves performance matching or surpassing state-of-the-art with minimal accuracy loss even when pruning up to 90% of tokens, with significant reductions in GPU memory and inference latency.

Conclusion: The proposed prompt-aware token pruning method effectively balances task relevance and information diversity, enabling efficient VLM inference without compromising performance.

Abstract: As the capabilities of Vision-Language Models (VLMs) advance, they can process increasingly large inputs, which, unlike in LLMs, generates significant visual token redundancy and leads to prohibitive inference costs. While many methods aim to reduce these costs by pruning visual tokens, existing approaches, whether based on attention or diversity, typically neglect the guidance of the text prompt and thus fail to prioritize task relevance. In this work, we propose a novel, zero-shot method that reframes the problem by introducing a prompt-aware perspective, explicitly modeling visual token pruning as a balance between task relevance and information diversity. Our hierarchical approach first selects a core set of task-relevant visual tokens and then supplements them with diversity tokens to preserve broader context. Experiments across multiple models and benchmarks show that our method achieves performance that matches or surpasses the state-of-the-art with only minimal accuracy loss, even when pruning up to 90% of the tokens. Furthermore, these gains are accompanied by significant reductions in GPU memory footprint and inference latency.

[321] From Pixels to People: Satellite-Based Mapping and Quantification of Riverbank Erosion and Lost Villages in Bangladesh

M Saifuzzaman Rafat, Mohd Ruhul Ameen, Akif Islam, Abu Saleh Musa Miah, Jungpil Shin

Main category: cs.CV

TL;DR: Adapted Segment Anything Model (SAM) for riverbank erosion detection in Bangladesh using historical Google Earth imagery, achieving 86.30% mIoU and 92.60% Dice score.

DetailsMotivation: To address the challenge of tracking river erosion that destroys villages and displaces thousands in Bangladesh, which was previously difficult for human analysts.

Method: Used color-channel analysis for rough land/water segmentation, then fine-tuned SAM’s mask decoder to recognize riverbank erosion patterns using a new annotated dataset of disappeared settlements.

Result: Achieved mean Intersection over Union of 86.30% and Dice score of 92.60%, significantly outperforming traditional methods and off-the-shelf deep learning models.

Conclusion: Provides a powerful tool for policymakers to monitor erosion, anticipate trajectories, and protect vulnerable communities through specialized AI model and annotated dataset.

Abstract: The great rivers of Bangladesh, arteries of commerce and sustenance, are also agents of relentless destruction. Each year, they swallow whole villages and vast tracts of farmland, erasing communities from the map and displacing thousands of families. To track this slow-motion catastrophe has, until now, been a Herculean task for human analysts. Here we show how a powerful general-purpose vision model, the Segment Anything Model (SAM), can be adapted to this task with remarkable precision. To do this, we assembled a new dataset

  • a digital chronicle of loss compiled from historical Google Earth imagery of Bangladesh’s most vulnerable regions, including Mokterer Char Union, Kedarpur Union, Balchipara village, and Chowhali Upazila, from 2003 to 2025. Crucially, this dataset is the first to include manually annotated data on the settlements that have vanished beneath the water. Our method first uses a simple color-channel analysis to provide a rough segmentation of land and water, and then fine-tunes SAM’s mask decoder to recognize the subtle signatures of riverbank erosion. The resulting model demonstrates a keen eye for this destructive process, achieving a mean Intersection over Union of 86.30% and a Dice score of 92.60% - a performance that significantly surpasses traditional methods and off-the-shelf deep learning models. This work delivers three key contributions: the first annotated dataset of disappeared settlements in Bangladesh due to river erosion; a specialized AI model fine-tuned for this critical task; and a method for quantifying land loss with compelling visual evidence. Together, these tools provide a powerful new lens through which policymakers and disaster management agencies can monitor erosion, anticipate its trajectory, and ultimately protect the vulnerable communities in its path.

[322] Round Outcome Prediction in VALORANT Using Tactical Features from Video Analysis

Nirai Hayakawa, Kazumasa Shimari, Kazuma Yamasaki, Hirotatsu Hoshikawa, Rikuto Tsuchida, Kenichi Matsumoto

Main category: cs.CV

TL;DR: This paper presents a round outcome prediction model for VALORANT using minimap information from match footage, achieving 81% accuracy by incorporating tactical features like character positions and in-game events.

DetailsMotivation: Most existing esports match prediction research relies on match log data and statistics, but this study focuses on analyzing visual minimap information from match footage to capture complex strategies in FPS games like VALORANT.

Method: Based on the TimeSformer video recognition model, the approach incorporates detailed tactical features extracted from minimap information, including character position data and other in-game events, using an augmented dataset with tactical event labels.

Result: The model achieved approximately 81% prediction accuracy, particularly from the middle phases of a round onward, significantly outperforming a baseline model trained only on raw minimap information without tactical features.

Conclusion: Leveraging tactical features extracted from match footage is highly effective for predicting round outcomes in VALORANT, demonstrating the value of visual analysis over traditional statistical approaches.

Abstract: Recently, research on predicting match outcomes in esports has been actively conducted, but much of it is based on match log data and statistical information. This research targets the FPS game VALORANT, which requires complex strategies, and aims to build a round outcome prediction model by analyzing minimap information in match footage. Specifically, based on the video recognition model TimeSformer, we attempt to improve prediction accuracy by incorporating detailed tactical features extracted from minimap information, such as character position information and other in-game events. This paper reports preliminary results showing that a model trained on a dataset augmented with such tactical event labels achieved approximately 81% prediction accuracy, especially from the middle phases of a round onward, significantly outperforming a model trained on a dataset with the minimap information itself. This suggests that leveraging tactical features from match footage is highly effective for predicting round outcomes in VALORANT.

[323] EndoCIL: A Class-Incremental Learning Framework for Endoscopic Image Classification

Bingrong Liu, Jun Shi, Yushan Zheng

Main category: cs.CV

TL;DR: EndoCIL is a novel class-incremental learning framework for endoscopic image analysis that addresses catastrophic forgetting through distribution-aligned exemplar selection, class-balanced loss, and gradient calibration.

DetailsMotivation: Existing replay-based CIL methods fail to effectively mitigate catastrophic forgetting in endoscopic imaging due to severe domain discrepancies and class imbalance inherent in clinical data.

Method: EndoCIL incorporates three key components: Maximum Mean Discrepancy Based Replay for diverse exemplar selection, Prior Regularized Class Balanced Loss to address class imbalance, and Calibration of Fully-Connected Gradients to mitigate bias toward new classes.

Result: Extensive experiments on four public endoscopic datasets demonstrate that EndoCIL generally outperforms state-of-the-art CIL methods across varying buffer sizes and evaluation metrics.

Conclusion: The framework effectively balances stability and plasticity in lifelong endoscopic diagnosis, showing promising potential for clinical scalability and deployment.

Abstract: Class-incremental learning (CIL) for endoscopic image analysis is crucial for real-world clinical applications, where diagnostic models should continuously adapt to evolving clinical data while retaining performance on previously learned ones. However, existing replay-based CIL methods fail to effectively mitigate catastrophic forgetting due to severe domain discrepancies and class imbalance inherent in endoscopic imaging. To tackle these challenges, we propose EndoCIL, a novel and unified CIL framework specifically tailored for endoscopic image diagnosis. EndoCIL incorporates three key components: Maximum Mean Discrepancy Based Replay (MDBR), employing a distribution-aligned greedy strategy to select diverse and representative exemplars, Prior Regularized Class Balanced Loss (PRCBL), designed to alleviate both inter-phase and intra-phase class imbalance by integrating prior class distributions and balance weights into the loss function, and Calibration of Fully-Connected Gradients (CFG), which adjusts the classifier gradients to mitigate bias toward new classes. Extensive experiments conducted on four public endoscopic datasets demonstrate that EndoCIL generally outperforms state-of-the-art CIL methods across varying buffer sizes and evaluation metrics. The proposed framework effectively balances stability and plasticity in lifelong endoscopic diagnosis, showing promising potential for clinical scalability and deployment.

[324] Optimizing DINOv2 with Registers for Face Anti-Spoofing

Mika Feng, Pierre Gallin-Martel, Koichi Ito, Takafumi Aoki

Main category: cs.CV

TL;DR: A DINOv2-based method for detecting face spoofing attacks by identifying subtle differences between live and spoofed face images, using registers to enhance feature extraction and attention mechanisms.

DetailsMotivation: Face recognition systems are vulnerable to spoofing attacks using photos of registered users, requiring robust detection methods before authentication.

Method: Uses DINOv2 with registers to extract generalizable features and suppress perturbations in attention mechanisms, focusing on essential minute features for spoof detection.

Result: Demonstrated effectiveness through experiments on datasets from The 6th Face Anti-Spoofing Workshop and SiW dataset.

Conclusion: The proposed DINOv2-based approach successfully detects face spoofing attacks by leveraging enhanced feature extraction and attention mechanisms.

Abstract: Face recognition systems are designed to be robust against variations in head pose, illumination, and image blur during capture. However, malicious actors can exploit these systems by presenting a face photo of a registered user, potentially bypassing the authentication process. Such spoofing attacks must be detected prior to face recognition. In this paper, we propose a DINOv2-based spoofing attack detection method to discern minute differences between live and spoofed face images. Specifically, we employ DINOv2 with registers to extract generalizable features and to suppress perturbations in the attention mechanism, which enables focused attention on essential and minute features. We demonstrate the effectiveness of the proposed method through experiments conducted on the dataset provided by ``The 6th Face Anti-Spoofing Workshop: Unified Physical-Digital Attacks Detection@ICCV2025’’ and SiW dataset.

[325] UltraCUA: A Foundation Model for Computer Use Agents with Hybrid Action

Yuhao Yang, Zhen Yang, Zi-Yi Dou, Anh Nguyen, Keen You, Omar Attia, Andrew Szot, Michael Feng, Ram Ramrakhya, Alexander Toshev, Chao Huang, Yinfei Yang, Zhe Gan

Main category: cs.CV

TL;DR: UltraCUA is a foundation model that bridges multimodal computer-use agents with programmatic tools through hybrid actions, combining GUI primitives with high-level tool calls to reduce cascading failures and improve performance.

DetailsMotivation: Current computer-use agents rely exclusively on primitive GUI actions (click, type, scroll) which lead to cascading failures and performance bottlenecks, while being isolated from rich programmatic interfaces available to other agents.

Method: Four key components: (1) automated pipeline scaling programmatic tools from documentation and code, (2) synthetic data engine with 17,000+ verifiable tasks, (3) large-scale hybrid action trajectory collection, and (4) two-stage training combining supervised fine-tuning with online reinforcement learning.

Result: 7B and 32B models show 22% relative improvement on OSWorld, 11% faster execution, and 21.7% success rate on WindowsAgentArena (outperforming Windows-trained baselines). Hybrid action reduces error propagation while maintaining efficiency.

Conclusion: Hybrid action mechanism successfully bridges GUI primitives with programmatic tools, enabling strategic alternation between low-level and high-level actions for improved performance and reduced cascading failures in computer-use agents.

Abstract: Multimodal agents for computer use rely exclusively on primitive actions (click, type, scroll) that require accurate visual grounding and lengthy execution chains, leading to cascading failures and performance bottlenecks. While other agents leverage rich programmatic interfaces (APIs, MCP servers, tools), computer-use agents (CUAs) remain isolated from these capabilities. We present UltraCUA, a foundation model that bridges this gap through hybrid action – seamlessly integrating GUI primitives with high-level programmatic tool calls. To achieve this, our approach comprises four key components: (1) an automated pipeline that scales programmatic tools from software documentation, open-source repositories, and code generation; (2) a synthetic data engine producing over 17,000 verifiable tasks spanning real-world computer-use scenarios; (3) a large-scale high-quality hybrid action trajectory collection with both low-level GUI actions and high-level programmatic tool calls; and (4) a two-stage training pipeline combining supervised fine-tuning with online reinforcement learning, enabling strategic alternation between low-level and high-level actions. Experiments with our 7B and 32B models demonstrate substantial improvements over state-of-the-art agents. On OSWorld, UltraCUA models achieve an average 22% relative improvement over base models, while being 11% faster in terms of steps. Out-of-domain evaluation on WindowsAgentArena shows our model reaches 21.7% success rate, outperforming baselines trained on Windows data. The hybrid action mechanism proves critical, reducing error propagation while maintaining execution efficiency.

[326] When One Moment Isn’t Enough: Multi-Moment Retrieval with Cross-Moment Interactions

Zhuo Cao, Heming Du, Bingqing Zhang, Xin Yu, Xue Li, Sen Wang

Main category: cs.CV

TL;DR: This paper introduces QV-M^2 dataset and FlashMMR framework for multi-moment retrieval, addressing limitations of existing single-moment retrieval methods in real-world video temporal grounding applications.

DetailsMotivation: Existing moment retrieval methods focus on single-moment retrieval, but real-world applications often require retrieving multiple relevant moments per query, making current datasets and methods insufficient.

Method: Proposed FlashMMR framework with Multi-moment Post-verification module using constrained temporal adjustment and verification module to refine moment boundaries and filter low-confidence proposals.

Result: FlashMMR achieves improvements over prior SOTA by 3.00% on G-mAP, 2.70% on mAP@3+tgt, and 2.56% on mR@3 on QV-M^2 dataset. The dataset contains 2,212 annotations covering 6,384 video segments.

Conclusion: QV-M^2 serves as an effective benchmark for training and evaluating MMR models, while FlashMMR provides a strong baseline, establishing foundation for advancing research in realistic video temporal grounding scenarios.

Abstract: Existing Moment retrieval (MR) methods focus on Single-Moment Retrieval (SMR). However, one query can correspond to multiple relevant moments in real-world applications. This makes the existing datasets and methods insufficient for video temporal grounding. By revisiting the gap between current MR tasks and real-world applications, we introduce a high-quality datasets called QVHighlights Multi-Moment Dataset (QV-M$^2$), along with new evaluation metrics tailored for multi-moment retrieval (MMR). QV-M$^2$ consists of 2,212 annotations covering 6,384 video segments. Building on existing efforts in MMR, we propose a framework called FlashMMR. Specifically, we propose a Multi-moment Post-verification module to refine the moment boundaries. We introduce constrained temporal adjustment and subsequently leverage a verification module to re-evaluate the candidate segments. Through this sophisticated filtering pipeline, low-confidence proposals are pruned, and robust multi-moment alignment is achieved. We retrain and evaluate 6 existing MR methods on QV-M$^2$ and QVHighlights under both SMR and MMR settings. Results show that QV-M$^2$ serves as an effective benchmark for training and evaluating MMR models, while FlashMMR provides a strong baseline. Specifically, on QV-M$^2$, it achieves improvements over prior SOTA method by 3.00% on G-mAP, 2.70% on mAP@3+tgt, and 2.56% on mR@3. The proposed benchmark and method establish a foundation for advancing research in more realistic and challenging video temporal grounding scenarios. Code is released at https://github.com/Zhuo-Cao/QV-M2.

[327] Glyph: Scaling Context Windows via Visual-Text Compression

Jiale Cheng, Yusen Liu, Xinyu Zhang, Yulin Fei, Wenyi Hong, Ruiliang Lyu, Weihan Wang, Zhe Su, Xiaotao Gu, Xiao Liu, Yushi Bai, Jie Tang, Hongning Wang, Minlie Huang

Main category: cs.CV

TL;DR: Glyph converts long text into images using vision-language models for 3-4x token compression while maintaining accuracy, enabling efficient million-token processing.

DetailsMotivation: Scaling LLMs to million-token contexts is computationally prohibitive, requiring alternative approaches to handle long documents efficiently.

Method: Render long texts into images processed by VLMs, with LLM-driven genetic search to optimize visual rendering configurations for accuracy-compression balance.

Result: Achieves 3-4x token compression with comparable accuracy to Qwen3-8B, 4x faster prefilling/decoding, 2x faster SFT training, and enables 128K-context VLM to handle 1M-token tasks.

Conclusion: Visual context scaling via Glyph provides efficient long-context processing while benefiting multimodal tasks like document understanding.

Abstract: Large language models (LLMs) increasingly rely on long-context modeling for tasks such as document understanding, code analysis, and multi-step reasoning. However, scaling context windows to the million-token level brings prohibitive computational and memory costs, limiting the practicality of long-context LLMs. In this work, we take a different perspective-visual context scaling-to tackle this challenge. Instead of extending token-based sequences, we propose Glyph, a framework that renders long texts into images and processes them with vision-language models (VLMs). This approach substantially compresses textual input while preserving semantic information, and we further design an LLM-driven genetic search to identify optimal visual rendering configurations for balancing accuracy and compression. Through extensive experiments, we demonstrate that our method achieves 3-4x token compression while maintaining accuracy comparable to leading LLMs such as Qwen3-8B on various long-context benchmarks. This compression also leads to around 4x faster prefilling and decoding, and approximately 2x faster SFT training. Furthermore, under extreme compression, a 128K-context VLM could scale to handle 1M-token-level text tasks. In addition, the rendered text data benefits real-world multimodal tasks, such as document understanding. Our code and model are released at https://github.com/thu-coai/Glyph.

[328] Fair and Interpretable Deepfake Detection in Videos

Akihito Yoshii, Ryosuke Sonoda, Ramya Srinivasan

Main category: cs.CV

TL;DR: A fairness-aware deepfake detection framework that integrates temporal feature learning and demographic-aware data augmentation to address bias and improve interpretability.

DetailsMotivation: Existing deepfake detection methods exhibit bias, lack transparency, and fail to capture temporal information, leading to biased decisions and unreliable results across different demographic groups.

Method: Uses sequence-based clustering for temporal modeling, concept extraction for interpretability, and demographic-aware data augmentation with frequency-domain transformations to preserve deepfake artifacts and balance underrepresented groups.

Result: Extensive experiments on FaceForensics++, DFD, Celeb-DF, and DFDC datasets using Xception and ResNet architectures demonstrate the method achieves the best tradeoff between fairness and accuracy compared to state-of-the-art approaches.

Conclusion: The proposed framework effectively mitigates bias in deepfake detection while maintaining high accuracy and providing interpretable decisions for non-expert users.

Abstract: Existing deepfake detection methods often exhibit bias, lack transparency, and fail to capture temporal information, leading to biased decisions and unreliable results across different demographic groups. In this paper, we propose a fairness-aware deepfake detection framework that integrates temporal feature learning and demographic-aware data augmentation to enhance fairness and interpretability. Our method leverages sequence-based clustering for temporal modeling of deepfake videos and concept extraction to improve detection reliability while also facilitating interpretable decisions for non-expert users. Additionally, we introduce a demography-aware data augmentation method that balances underrepresented groups and applies frequency-domain transformations to preserve deepfake artifacts, thereby mitigating bias and improving generalization. Extensive experiments on FaceForensics++, DFD, Celeb-DF, and DFDC datasets using state-of-the-art (SoTA) architectures (Xception, ResNet) demonstrate the efficacy of the proposed method in obtaining the best tradeoff between fairness and accuracy when compared to SoTA.

[329] FineVision: Open Data Is All You Need

Luis Wiedmann, Orr Zohar, Amir Mahla, Xiaohan Wang, Rui Li, Thibaud Frere, Leandro von Werra, Aritra Roy Gosthipaty, Andrés Marafioti

Main category: cs.CV

TL;DR: FineVision is a large-scale, meticulously curated vision-language dataset of 24 million samples that addresses data fragmentation and contamination issues through a semi-automated human-in-the-loop curation pipeline.

DetailsMotivation: To overcome the fragmented landscape of inconsistent and contaminated public datasets that hinder vision-language model advancement.

Method: Semi-automated human-in-the-loop pipeline that unifies 200+ sources into 185 subsets, with automation for bulk ingestion and human reviewers for auditing, verification, and quality control, plus rigorous de-duplication and decontamination.

Result: Models trained on FineVision consistently outperform those trained on existing open mixtures across a broad evaluation suite, demonstrating benefits of scale, data hygiene, and balanced automation with human oversight.

Conclusion: The FineVision corpus and curation tools are released to accelerate data-centric VLM research, showing that careful data curation with human oversight significantly improves model performance.

Abstract: The advancement of vision-language models (VLMs) is hampered by a fragmented landscape of inconsistent and contaminated public datasets. We introduce FineVision, a meticulously collected, curated, and unified corpus of 24 million samples - the largest open resource of its kind. We unify more than 200 sources into 185 subsets via a semi-automated, human-in-the-loop pipeline: automation performs bulk ingestion and schema mapping, while reviewers audit mappings and spot-check outputs to verify faithful consumption of annotations, appropriate formatting and diversity, and safety; issues trigger targeted fixes and re-runs. The workflow further applies rigorous de-duplication within and across sources and decontamination against 66 public benchmarks. FineVision also encompasses agentic/GUI tasks with a unified action space; reviewers validate schemas and inspect a sample of trajectories to confirm executable fidelity. Models trained on FineVision consistently outperform those trained on existing open mixtures across a broad evaluation suite, underscoring the benefits of scale, data hygiene, and balanced automation with human oversight. We release the corpus and curation tools to accelerate data-centric VLM research.

[330] Enhanced Motion Forecasting with Plug-and-Play Multimodal Large Language Models

Katie Luo, Jingwei Ji, Tong He, Runsheng Xu, Yichen Xie, Dragomir Anguelov, Mingxing Tan

Main category: cs.CV

TL;DR: Plug-and-Forecast (PnF) is a plug-and-play method that enhances existing motion forecasting models using multimodal large language models (MLLMs) to improve generalization in autonomous driving without requiring fine-tuning.

DetailsMotivation: Current autonomous driving systems struggle to generalize cost-effectively to diverse real-world scenarios, as specialized models for perception and motion prediction perform well in standard conditions but lack adaptability.

Method: PnF extracts structured scene understanding from MLLMs using designed prompts, distills this information into learnable embeddings, and augments existing behavior prediction models with these embeddings to leverage MLLMs’ zero-shot reasoning capabilities.

Result: The approach achieves significant improvements in motion prediction performance on both Waymo Open Motion Dataset and nuScenes Dataset, demonstrating consistent performance gains across state-of-the-art motion forecasting models.

Conclusion: PnF provides an effective plug-and-play solution that enables quick adaptation to complex scenarios using natural language descriptions, making it practical to adopt without fine-tuning while improving motion forecasting performance.

Abstract: Current autonomous driving systems rely on specialized models for perceiving and predicting motion, which demonstrate reliable performance in standard conditions. However, generalizing cost-effectively to diverse real-world scenarios remains a significant challenge. To address this, we propose Plug-and-Forecast (PnF), a plug-and-play approach that augments existing motion forecasting models with multimodal large language models (MLLMs). PnF builds on the insight that natural language provides a more effective way to describe and handle complex scenarios, enabling quick adaptation to targeted behaviors. We design prompts to extract structured scene understanding from MLLMs and distill this information into learnable embeddings to augment existing behavior prediction models. Our method leverages the zero-shot reasoning capabilities of MLLMs to achieve significant improvements in motion prediction performance, while requiring no fine-tuning – making it practical to adopt. We validate our approach on two state-of-the-art motion forecasting models using the Waymo Open Motion Dataset and the nuScenes Dataset, demonstrating consistent performance improvements across both benchmarks.

[331] SG-CLDFF: A Novel Framework for Automated White Blood Cell Classification and Segmentation

Mehdi Zekriyapanah Gashti, Mostafa Mohammadpour, Ghasem Farjamnia

Main category: cs.CV

TL;DR: SG-CLDFF is a saliency-guided cross-layer deep feature fusion framework that improves white blood cell segmentation and classification through saliency-driven preprocessing and multi-scale feature aggregation, achieving better performance and interpretability.

DetailsMotivation: Accurate white blood cell analysis is crucial for hematological diagnosis but faces challenges from staining variability, complex backgrounds, and class imbalance in microscopic images.

Method: Uses saliency priors to highlight WBC regions, a hybrid EfficientSwin-style backbone for multi-resolution features, cross-layer fusion module, multi-task training with segmentation and classification heads, class-aware weighted losses, and saliency-alignment regularization.

Result: Consistent improvements in IoU, F1, and classification accuracy on standard benchmarks (BCCD, LISC, ALL-IDB) compared to CNN and transformer baselines, with ablation studies confirming contributions of saliency preprocessing and cross-layer fusion.

Conclusion: SG-CLDFF provides a practical and explainable solution for reliable automated WBC analysis in clinical workflows, with enhanced robustness and interpretability through saliency guidance and feature fusion.

Abstract: Accurate segmentation and classification of white blood cells (WBCs) in microscopic images are essential for diagnosis and monitoring of many hematological disorders, yet remain challenging due to staining variability, complex backgrounds, and class imbalance. In this paper, we introduce a novel Saliency-Guided Cross-Layer Deep Feature Fusion framework (SG-CLDFF) that tightly integrates saliency-driven preprocessing with multi-scale deep feature aggregation to improve both robustness and interpretability for WBC analysis. SG-CLDFF first computes saliency priors to highlight candidate WBC regions and guide subsequent feature extraction. A lightweight hybrid backbone (EfficientSwin-style) produces multi-resolution representations, which are fused by a ResNeXt-CC-inspired cross-layer fusion module to preserve complementary information from shallow and deep layers. The network is trained in a multi-task setup with concurrent segmentation and cell-type classification heads, using class-aware weighted losses and saliency-alignment regularization to mitigate imbalance and suppress background activation. Interpretability is enforced through Grad-CAM visualizations and saliency consistency checks, allowing model decisions to be inspected at the regional level. We validate the framework on standard public benchmarks (BCCD, LISC, ALL-IDB), reporting consistent gains in IoU, F1, and classification accuracy compared to strong CNN and transformer baselines. An ablation study also demonstrates the individual contributions of saliency preprocessing and cross-layer fusion. SG-CLDFF offers a practical and explainable path toward more reliable automated WBC analysis in clinical workflows.

[332] Machine Vision-Based Surgical Lighting System:Design and Implementation

Amir Gharghabi, Mahdi Hakiminezhad, Maryam Shafaei, Shaghayegh Gharghabi

Main category: cs.CV

TL;DR: A novel surgical lighting system using YOLOv11 to detect blue markers and automatically direct LED lighting via servomotors, reducing surgeon fatigue and improving illumination consistency.

DetailsMotivation: Traditional surgical lighting requires manual adjustments causing surgeon fatigue, neck strain, and inconsistent illumination due to drift and shadowing.

Method: Uses YOLOv11 object detection to identify blue markers above surgical sites, then directs high-power LED lighting using servomotors with tilt-pan brackets.

Result: YOLO model achieves 96.7% mAP@50 on validation set with annotated surgical scene images containing blue spherical markers.

Conclusion: The automated machine vision-based lighting system reduces physical strain on surgeons, improves illumination consistency, and supports better surgical outcomes.

Abstract: Effortless and ergonomically designed surgical lighting is critical for precision and safety during procedures. However, traditional systems often rely on manual adjustments, leading to surgeon fatigue, neck strain, and inconsistent illumination due to drift and shadowing. To address these challenges, we propose a novel surgical lighting system that leverages the YOLOv11 object detection algorithm to identify a blue marker placed above the target surgical site. A high-power LED light source is then directed to the identified location using two servomotors equipped with tilt-pan brackets. The YOLO model achieves 96.7% mAP@50 on the validation set consisting of annotated images simulating surgical scenes with the blue spherical marker. By automating the lighting process, this machine vision-based solution reduces physical strain on surgeons, improves consistency in illumination, and supports improved surgical outcomes.

[333] Exploring Structural Degradation in Dense Representations for Self-supervised Learning

Siran Dai, Qianqian Xu, Peisong Wen, Yang Liu, Qingming Huang

Main category: cs.CV

TL;DR: Longer self-supervised learning training can degrade dense prediction performance (Self-supervised Dense Degradation). The paper introduces DSE metric for model selection and regularization to address this issue.

DetailsMotivation: To address the counterintuitive phenomenon where longer self-supervised learning training impairs dense prediction tasks like semantic segmentation, and to provide effective evaluation without annotations.

Method: Proposes Dense representation Structure Estimator (DSE) with class-relevance and effective dimensionality measures, plus model selection strategy and DSE-based regularization.

Result: Model selection improves mIoU by 3.0% on average across 16 SSL methods, and DSE regularization consistently mitigates dense degradation effects.

Conclusion: The DSE metric effectively identifies optimal training checkpoints for dense tasks and provides a regularization method to prevent performance degradation in self-supervised learning.

Abstract: In this work, we observe a counterintuitive phenomenon in self-supervised learning (SSL): longer training may impair the performance of dense prediction tasks (e.g., semantic segmentation). We refer to this phenomenon as Self-supervised Dense Degradation (SDD) and demonstrate its consistent presence across sixteen state-of-the-art SSL methods with various losses, architectures, and datasets. When the model performs suboptimally on dense tasks at the end of training, measuring the performance during training becomes essential. However, evaluating dense performance effectively without annotations remains an open challenge. To tackle this issue, we introduce a Dense representation Structure Estimator (DSE), composed of a class-relevance measure and an effective dimensionality measure. The proposed DSE is both theoretically grounded and empirically validated to be closely correlated with the downstream performance. Based on this metric, we introduce a straightforward yet effective model selection strategy and a DSE-based regularization method. Experiments on sixteen SSL methods across four benchmarks confirm that model selection improves mIoU by $3.0%$ on average with negligible computational cost. Additionally, DSE regularization consistently mitigates the effects of dense degradation. Code is available at https://github.com/EldercatSAM/SSL-Degradation.

[334] CausalMamba: Scalable Conditional State Space Models for Neural Causal Inference

Sangyoon Bae, Jiook Cha

Main category: cs.CV

TL;DR: CausalMamba is a scalable framework that improves fMRI-based causal inference by addressing hemodynamic distortion and computational limitations of existing methods through a two-stage approach.

DetailsMotivation: To overcome fundamental limitations in fMRI-based causal inference, specifically the ill-posed nature of inferring neural causality from hemodynamically distorted BOLD signals and the computational intractability of existing methods like Dynamic Causal Modeling.

Method: Decomposes the problem into two stages: BOLD deconvolution to recover latent neural activity, followed by causal graph inference using a novel Conditional Mamba architecture.

Result: Achieves 37% higher accuracy than DCM on simulated data, recovers well-established neural pathways with 88% fidelity in real task fMRI data, and reveals strategic brain network reconfigurations during working memory tasks that traditional methods miss.

Conclusion: Provides neuroscientists with a practical tool for large-scale causal inference that captures both fundamental circuit motifs and flexible network dynamics underlying cognitive function.

Abstract: We introduce CausalMamba, a scalable framework that addresses fundamental limitations in fMRI-based causal inference: the ill-posed nature of inferring neural causality from hemodynamically distorted BOLD signals and the computational intractability of existing methods like Dynamic Causal Modeling (DCM). Our approach decomposes this complex inverse problem into two tractable stages: BOLD deconvolution to recover latent neural activity, followed by causal graph inference using a novel Conditional Mamba architecture. On simulated data, CausalMamba achieves 37% higher accuracy than DCM. Critically, when applied to real task fMRI data, our method recovers well-established neural pathways with 88% fidelity, whereas conventional approaches fail to identify these canonical circuits in over 99% of subjects. Furthermore, our network analysis of working memory data reveals that the brain strategically shifts its primary causal hub-recruiting executive or salience networks depending on the stimulus-a sophisticated reconfiguration that remains undetected by traditional methods. This work provides neuroscientists with a practical tool for large-scale causal inference that captures both fundamental circuit motifs and flexible network dynamics underlying cognitive function.

[335] A Single Set of Adversarial Clothes Breaks Multiple Defense Methods in the Physical World

Wei Zhang, Zhanhao Hu, Xiao Li, Xiaopei Zhu, Xiaolin Hu

Main category: cs.CV

TL;DR: Adversarial clothes with large coverage can break existing defense methods against adversarial patches, revealing common vulnerabilities in current adversarial defense approaches.

DetailsMotivation: To evaluate the robustness of existing adversarial defense methods against large-scale physical attacks, specifically adversarial clothes that cover large areas of the human body, as simple patch size enlargement was found to bypass current defenses.

Method: Evaluated various defense methods against adversarial clothes in both digital and physical worlds, crafting a single set of adversarial clothes that could break multiple defense methods on Faster R-CNN detector.

Result: All defense methods performed poorly against adversarial clothes. A single adversarial clothing set achieved 96.06% ASR against undefended detector and over 64.84% ASR against nine defended models in physical world.

Conclusion: Existing adversarial defense methods have common vulnerabilities against large-scale adversarial attacks like adversarial clothes, highlighting the need for more robust defense strategies.

Abstract: In recent years, adversarial attacks against deep learning-based object detectors in the physical world have attracted much attention. To defend against these attacks, researchers have proposed various defense methods against adversarial patches, a typical form of physically-realizable attack. However, our experiments showed that simply enlarging the patch size could make these defense methods fail. Motivated by this, we evaluated various defense methods against adversarial clothes which have large coverage over the human body. Adversarial clothes provide a good test case for adversarial defense against patch-based attacks because they not only have large sizes but also look more natural than a large patch on humans. Experiments show that all the defense methods had poor performance against adversarial clothes in both the digital world and the physical world. In addition, we crafted a single set of clothes that broke multiple defense methods on Faster R-CNN. The set achieved an Attack Success Rate (ASR) of 96.06% against the undefended detector and over 64.84% ASRs against nine defended models in the physical world, unveiling the common vulnerability of existing adversarial defense methods against adversarial clothes. Code is available at: https://github.com/weiz0823/adv-clothes-break-multiple-defenses.

[336] CharDiff: A Diffusion Model with Character-Level Guidance for License Plate Image Restoration

Gyuhwan Park, Kihyun Na, Injung Kim

Main category: cs.CV

TL;DR: CharDiff is a diffusion-based framework with character-level guidance that effectively restores and recognizes severely degraded license plate images using fine-grained character priors and region-wise masking.

DetailsMotivation: License plate image restoration is important not only for LPR preprocessing but also for increasing evidential value, enhancing visual clarity, and facilitating further utilization of license plate images.

Method: Uses diffusion-based framework with character-level guidance, leveraging fine-grained character priors from external segmentation and OCR modules. Incorporates CHARM module for precise region-wise masking to restrict character guidance to their own regions.

Result: Significantly outperformed baseline restoration models, achieving 28% relative reduction in CER on Roboflow-LP dataset compared to best-performing baseline model.

Conclusion: Structured character-guided conditioning effectively enhances robustness of diffusion-based license plate restoration and recognition in practical deployment scenarios.

Abstract: The significance of license plate image restoration goes beyond the preprocessing stage of License Plate Recognition (LPR) systems, as it also serves various purposes, including increasing evidential value, enhancing the clarity of visual interface, and facilitating further utilization of license plate images. We propose a novel diffusion-based framework with character-level guidance, CharDiff, which effectively restores and recognizes severely degraded license plate images captured under realistic conditions. CharDiff leverages fine-grained character-level priors extracted through external segmentation and Optical Character Recognition (OCR) modules tailored for low-quality license plate images. For precise and focused guidance, CharDiff incorporates a novel Character-guided Attention through Region-wise Masking (CHARM) module, which ensures that each character’s guidance is restricted to its own region, thereby avoiding interference with other regions. In experiments, CharDiff significantly outperformed the baseline restoration models in both restoration quality and recognition accuracy, achieving a 28% relative reduction in CER on the Roboflow-LP dataset, compared to the best-performing baseline model. These results indicate that the structured character-guided conditioning effectively enhances the robustness of diffusion-based license plate restoration and recognition in practical deployment scenarios.

[337] iDETEX: Empowering MLLMs for Intelligent DETailed EXplainable IQA

Zhaoran Zhao, Xinli Yue, Jianhui Sun, Yuhao Xie, Tao Shao, Liangchao Yao, Fan Xia, Yuetang Deng

Main category: cs.CV

TL;DR: iDETEX is a unified multimodal large language model that performs quality grounding, perception, and description for detailed and explainable image quality assessment, achieving state-of-the-art results on the ViDA-UGC benchmark and winning the ICCV MIPI 2025 challenge.

DetailsMotivation: To address the emerging challenge of detailed and explainable image quality assessment beyond simple scalar quality prediction, moving towards more human-aligned evaluation paradigms.

Method: Proposed iDETEX - a unified multimodal large language model with task-specific offline augmentation modules, data mixing strategy, and online enhancement strategies to exploit multi-sourced supervision across three key tasks: quality grounding, perception, and description.

Result: Achieved state-of-the-art performance across all subtasks on the ViDA-UGC benchmark and ranked first in the ICCV MIPI 2025 Detailed Image Quality Assessment Challenge.

Conclusion: iDETEX demonstrates effectiveness and robustness in delivering accurate and interpretable quality assessments through its unified multimodal approach.

Abstract: Image Quality Assessment (IQA) has progressed from scalar quality prediction to more interpretable, human-aligned evaluation paradigms. In this work, we address the emerging challenge of detailed and explainable IQA by proposing iDETEX-a unified multimodal large language model (MLLM) capable of simultaneously performing three key tasks: quality grounding, perception, and description. To facilitate efficient and generalizable training across these heterogeneous subtasks, we design a suite of task-specific offline augmentation modules and a data mixing strategy. These are further complemented by online enhancement strategies to fully exploit multi-sourced supervision. We validate our approach on the large-scale ViDA-UGC benchmark, where iDETEX achieves state-of-the-art performance across all subtasks. Our model ranks first in the ICCV MIPI 2025 Detailed Image Quality Assessment Challenge, demonstrating its effectiveness and robustness in delivering accurate and interpretable quality assessments.

[338] Nearest-Class Mean and Logits Agreement for Wildlife Open-Set Recognition

Jiahao Huo, Mufhumudzi Muthivhi, Terence L. van Zyl, Fredrik Gustafsson

Main category: cs.CV

TL;DR: A post-processing open-set recognition method that measures agreement between feature-based NCM distribution and softmax probabilities, achieving top performance on wildlife datasets without retraining.

DetailsMotivation: Current OSR methods require retraining pre-trained models and perform inconsistently across datasets. Wildlife classification needs reliable unknown class detection without model modification.

Method: Proposes measuring agreement between NCM-based probability distribution (from feature space) and softmax probabilities (from logit space) as a post-processing step.

Result: Achieved AUROC of 93.41 and 95.35 on African and Swedish animal datasets, ranking top three consistently across both datasets.

Conclusion: The method provides consistent open-set recognition performance across different datasets without requiring model retraining, outperforming state-of-the-art approaches that excel on only single datasets.

Abstract: Current state-of-the-art Wildlife classification models are trained under the closed world setting. When exposed to unknown classes, they remain overconfident in their predictions. Open-set Recognition (OSR) aims to classify known classes while rejecting unknown samples. Several OSR methods have been proposed to model the closed-set distribution by observing the feature, logit, or softmax probability space. A significant drawback of many existing approaches is the requirement to retrain the pre-trained classification model with the OSR-specific strategy. This study contributes a post-processing OSR method that measures the agreement between the models’ features and predicted logits. We propose a probability distribution based on an input’s distance to its Nearest Class Mean (NCM). The NCM-based distribution is then compared with the softmax probabilities from the logit space to measure agreement between the NCM and the classification head. Our proposed strategy ranks within the top three on two evaluated datasets, showing consistent performance across the two datasets. In contrast, current state-of-the-art methods excel on a single dataset. We achieve an AUROC of 93.41 and 95.35 for African and Swedish animals. The code can be found https://github.com/Applied-Representation-Learning-Lab/OSR.

[339] Exploring The Missing Semantics In Event Modality

Jingqian Wu, Shengpeng Xu, Yunbo Jia, Edmund Y. Lam

Main category: cs.CV

TL;DR: Semantic-E2VID is an event-to-video reconstruction framework that enhances reconstruction quality by incorporating semantic information from vision foundation models, outperforming state-of-the-art methods.

DetailsMotivation: Event cameras capture only intensity changes, lacking semantic information about static objects and backgrounds, which limits the quality of event-to-video reconstruction. Existing approaches often overlook semantic information that is crucial for video reconstruction.

Method: Proposes Semantic-E2VID with: 1) Cross-modal feature alignment module to transfer visual semantics from SAM to event encoder, 2) Semantic-aware feature fusion block to integrate learned semantics, and 3) Semantic Perceptual E2V Supervision using SAM-generated labels.

Result: Extensive experiments show Semantic-E2VID significantly enhances frame quality and outperforms state-of-the-art E2V methods across multiple benchmarks.

Conclusion: The proposed framework successfully bridges the semantic gap in event-to-video reconstruction by leveraging vision foundation models, demonstrating improved reconstruction quality through semantic-aware processing.

Abstract: Event cameras offer distinct advantages such as low latency, high dynamic range, and efficient motion capture. However, event-to-video reconstruction (E2V), a fundamental event-based vision task, remains challenging, particularly for reconstructing and recovering semantic information. This is primarily due to the nature of the event camera, as it only captures intensity changes, ignoring static objects and backgrounds, resulting in a lack of semantic information in captured event modality. Further, semantic information plays a crucial role in video and frame reconstruction, yet is often overlooked by existing E2V approaches. To bridge this gap, we propose Semantic-E2VID, an E2V framework that explores the missing visual semantic knowledge in event modality and leverages it to enhance event-to-video reconstruction. Specifically, Semantic-E2VID introduces a cross-modal feature alignment (CFA) module to transfer the robust visual semantics from a frame-based vision foundation model, the Segment Anything Model (SAM), to the event encoder, while aligning the high-level features from distinct modalities. To better utilize the learned semantic feature, we further propose a semantic-aware feature fusion (SFF) block to integrate learned semantics in frame modality to form event representations with rich semantics that can be decoded by the event decoder. Further, to facilitate the reconstruction of semantic information, we propose a novel Semantic Perceptual E2V Supervision that helps the model to reconstruct semantic details by leveraging SAM-generated categorical labels. Extensive experiments demonstrate that Semantic-E2VID significantly enhances frame quality, outperforming state-of-the-art E2V methods across multiple benchmarks. The sample code is included in the supplementary material.

[340] M2H: Multi-Task Learning with Efficient Window-Based Cross-Task Attention for Monocular Spatial Perception

U. V. B. L Udugama, George Vosselman, Francesco Nex

Main category: cs.CV

TL;DR: M2H is a multi-task learning framework for real-time spatial perception that performs semantic segmentation, depth, edge, and surface normal estimation from a single monocular image using a Window-Based Cross-Task Attention Module and lightweight ViT-based backbone.

DetailsMotivation: Need for efficient multi-task models that leverage complementary task information while minimizing computational overhead for real-time spatial perception deployment on edge devices.

Method: Uses Window-Based Cross-Task Attention Module for structured feature exchange while preserving task-specific details, built on lightweight ViT-based DINOv2 backbone optimized for real-time deployment.

Result: Outperforms state-of-the-art multi-task models on NYUDv2, surpasses single-task depth and semantic baselines on Hypersim, achieves superior performance on Cityscapes, while maintaining computational efficiency on laptop hardware.

Conclusion: M2H demonstrates practicality in real-world spatial perception tasks and serves as foundation for monocular spatial perception systems supporting 3D scene graph construction in dynamic environments.

Abstract: Deploying real-time spatial perception on edge devices requires efficient multi-task models that leverage complementary task information while minimizing computational overhead. This paper introduces Multi-Mono-Hydra (M2H), a novel multi-task learning framework designed for semantic segmentation and depth, edge, and surface normal estimation from a single monocular image. Unlike conventional approaches that rely on independent single-task models or shared encoder-decoder architectures, M2H introduces a Window-Based Cross-Task Attention Module that enables structured feature exchange while preserving task-specific details, improving prediction consistency across tasks. Built on a lightweight ViT-based DINOv2 backbone, M2H is optimized for real-time deployment and serves as the foundation for monocular spatial perception systems supporting 3D scene graph construction in dynamic environments. Comprehensive evaluations show that M2H outperforms state-of-the-art multi-task models on NYUDv2, surpasses single-task depth and semantic baselines on Hypersim, and achieves superior performance on the Cityscapes dataset, all while maintaining computational efficiency on laptop hardware. Beyond benchmarks, M2H is validated on real-world data, demonstrating its practicality in spatial perception tasks.

[341] Recurrent Attention-based Token Selection for Efficient Streaming Video-LLMs

Vaggelis Dorovatas, Soroush Seifi, Gunshi Gupta, Rahaf Aljundi

Main category: cs.CV

TL;DR: A training-free method for Video-LLMs that enables efficient streaming video processing by selecting important visual tokens, using recurrent processing, and caption-based QA.

DetailsMotivation: Video-LLMs struggle with streaming scenarios where hour-long videos need online processing and timely responses, requiring efficient methods to handle continuous video streams.

Method: Three key concepts: 1) LLM-informed selection of visual tokens based on attention to discard ~95% unimportant tokens, 2) Recurrent processing of past selected tokens for temporal coherence, 3) Caption-based question answering for lightweight responses.

Result: Achieves state-of-the-art performance on streaming video benchmarks with minimal performance loss while maintaining high efficiency.

Conclusion: The proposed training-free approach effectively balances efficiency and effectiveness for streaming video understanding in Video-LLMs.

Abstract: Video Large Language Models (Video-LLMs) excel at understanding videos in-context, provided they have full access to the video when answering queries. However, these models face challenges in streaming scenarios where hour-long videos must be processed online, and questions need timely responses. In this work, we propose a training-free approach compatible with standard Video-LLMs, leveraging three key concepts: 1) LLM-informed selection of visual tokens to identify those that the LLM has attended to and contributed to its understanding of each short clip. Our attention-based selection allows us to discard up to ~95% of unimportant visual tokens with minimal performance loss; 2) Recurrent processing of past selected tokens to generate temporally coherent understanding of each processed clip; 3) Caption-based question answering for lightweight and accurate responses. Our method achieves state-of-the-art performance on streaming video benchmarks, striking a balance between efficiency and effectiveness.

[342] Beyond Real Faces: Synthetic Datasets Can Achieve Reliable Recognition Performance without Privacy Compromise

Paweł Borsukiewicz, Fadi Boutros, Iyiola E. Olatunji, Charles Beumier, Wendkûuni C. Ouedraogo, Jacques Klein, Tegawendé F. Bissyandé

Main category: cs.CV

TL;DR: Synthetic facial recognition datasets can achieve comparable or better accuracy than real datasets while addressing privacy concerns, with top performers reaching 95.67% accuracy and offering unprecedented bias mitigation control.

DetailsMotivation: Real facial datasets collected without consent create ethical dilemmas and legal liabilities under regulations like GDPR, while synthetic data offers privacy-preserving alternatives but lacks comprehensive empirical validation.

Method: Systematic literature review of 25 synthetic facial recognition datasets (2018-2025) combined with experimental validation of 10M+ synthetic samples, evaluating 7 key privacy requirements and comparing results on 5 standard benchmarks.

Result: Best synthetic datasets (VariFace, VIGFace) achieved 95.67% and 94.91% accuracy respectively, surpassing CASIA-WebFace (94.70%). Synthetic data ensures proper intra-class variability and identity separability while offering unprecedented bias mitigation control.

Conclusion: Synthetic facial data is scientifically viable and ethically imperative for facial recognition research, providing privacy-preserving alternatives that can replace real datasets while enabling better bias control.

Abstract: The deployment of facial recognition systems has created an ethical dilemma: achieving high accuracy requires massive datasets of real faces collected without consent, leading to dataset retractions and potential legal liabilities under regulations like GDPR. While synthetic facial data presents a promising privacy-preserving alternative, the field lacks comprehensive empirical evidence of its viability. This study addresses this critical gap through extensive evaluation of synthetic facial recognition datasets. We present a systematic literature review identifying 25 synthetic facial recognition datasets (2018-2025), combined with rigorous experimental validation. Our methodology examines seven key requirements for privacy-preserving synthetic data: identity leakage prevention, intra-class variability, identity separability, dataset scale, ethical data sourcing, bias mitigation, and benchmark reliability. Through experiments involving over 10 million synthetic samples, extended by a comparison of results reported on five standard benchmarks, we provide the first comprehensive empirical assessment of synthetic data’s capability to replace real datasets. Best-performing synthetic datasets (VariFace, VIGFace) achieve recognition accuracies of 95.67% and 94.91% respectively, surpassing established real datasets including CASIA-WebFace (94.70%). While those images remain private, publicly available alternatives Vec2Face (93.52%) and CemiFace (93.22%) come close behind. Our findings reveal that they ensure proper intra-class variability while maintaining identity separability. Demographic bias analysis shows that, even though synthetic data inherits limited biases, it offers unprecedented control for bias mitigation through generation parameters. These results establish synthetic facial data as a scientifically viable and ethically imperative alternative for facial recognition research.

[343] Facial Expression-based Parkinson’s Disease Severity Diagnosis via Feature Fusion and Adaptive Class Balancing

Yintao Zhou, Wei Huang, Zhengyu Li, Jing Huang, Meng Pang

Main category: cs.CV

TL;DR: A new method for Parkinson’s disease severity diagnosis using multiple facial expression features with attention-based fusion and adaptive class balancing to address class imbalance issues.

DetailsMotivation: Current PD diagnosis methods rely on single facial expressions leading to misdiagnosis, ignore class imbalance across PD stages, and mostly focus on binary classification rather than severity diagnosis.

Method: Integrates multiple facial expression features through attention-based feature fusion and uses adaptive class balancing strategy that dynamically adjusts training sample contributions based on class distribution and classification difficulty.

Result: Experimental results demonstrate promising performance for PD severity diagnosis and confirm the efficacy of both attention-based feature fusion and adaptive class balancing.

Conclusion: The proposed method effectively addresses limitations of existing approaches by leveraging multiple facial expressions and handling class imbalance, showing strong potential for PD severity diagnosis.

Abstract: Parkinson’s disease (PD) severity diagnosis is crucial for early detecting potential patients and adopting tailored interventions. Diagnosing PD based on facial expression is grounded in PD patients’ “masked face” symptom and gains growing interest recently for its convenience and affordability. However, current facial expression-based approaches often rely on single type of expression which can lead to misdiagnosis, and ignore the class imbalance across different PD stages which degrades the prediction performance. Moreover, most existing methods focus on binary classification (i.e., PD / non-PD) rather than diagnosing the severity of PD. To address these issues, we propose a new facial expression-based method for PD severity diagnosis which integrates multiple facial expression features through attention-based feature fusion. Moreover, we mitigate the class imbalance problem via an adaptive class balancing strategy which dynamically adjusts the contribution of training samples based on their class distribution and classification difficulty. Experimental results demonstrate the promising performance of the proposed method for PD severity diagnosis, as well as the efficacy of attention-based feature fusion and adaptive class balancing.

[344] Closed-Loop Transfer for Weakly-supervised Affordance Grounding

Jiajin Tang, Zhengxuan Wei, Ge Zheng, Sibei Yang

Main category: cs.CV

TL;DR: LoopTrans is a closed-loop framework for affordance grounding that transfers knowledge bidirectionally between exocentric and egocentric views, improving performance in complex interaction scenarios.

DetailsMotivation: Previous weakly-supervised affordance grounding methods only transfer knowledge one-way from exocentric to egocentric images, limiting applicability in complex interaction scenarios where object interaction regions may be occluded.

Method: LoopTrans introduces a closed-loop framework with unified cross-modal localization and denoising knowledge distillation to bridge domain gaps between object-centered egocentric and interaction-centered exocentric images.

Result: Experiments show consistent improvements across all metrics on image and video benchmarks, even handling challenging scenarios where object interaction regions are fully occluded by the human body.

Conclusion: The bidirectional knowledge transfer in LoopTrans enhances affordance grounding performance and addresses limitations of previous one-way transfer approaches.

Abstract: Humans can perform previously unexperienced interactions with novel objects simply by observing others engage with them. Weakly-supervised affordance grounding mimics this process by learning to locate object regions that enable actions on egocentric images, using exocentric interaction images with image-level annotations. However, extracting affordance knowledge solely from exocentric images and transferring it one-way to egocentric images limits the applicability of previous works in complex interaction scenarios. Instead, this study introduces LoopTrans, a novel closed-loop framework that not only transfers knowledge from exocentric to egocentric but also transfers back to enhance exocentric knowledge extraction. Within LoopTrans, several innovative mechanisms are introduced, including unified cross-modal localization and denoising knowledge distillation, to bridge domain gaps between object-centered egocentric and interaction-centered exocentric images while enhancing knowledge transfer. Experiments show that LoopTrans achieves consistent improvements across all metrics on image and video benchmarks, even handling challenging scenarios where object interaction regions are fully occluded by the human body.

[345] Monitoring Horses in Stalls: From Object to Event Detection

Dmitrii Galimzianov, Viacheslav Vyshegorodtsev, Ivan Nezhivykh

Main category: cs.CV

TL;DR: A vision-based system using YOLOv11 and BoT-SORT automates horse behavior monitoring in stables by detecting/tracking horses and people, classifying five event types while accounting for camera blind spots.

DetailsMotivation: Manual monitoring of stalled horses is labor-intensive and time-consuming, creating a need for automated systems to detect health and welfare issues early.

Method: Uses object detection (YOLOv11) and multi-object tracking (BoT-SORT) to track horses and people in stables, with event classification based on object trajectories and spatial relations. Custom dataset created using CLIP and GroundingDINO foundation models.

Result: System reliably detects horse-related events but struggles with people detection due to data scarcity. Successfully distinguishes five event types and handles camera blind spots.

Conclusion: Provides foundation for real-time behavioral monitoring in equine facilities, with potential applications in animal welfare and stable management.

Abstract: Monitoring the behavior of stalled horses is essential for early detection of health and welfare issues but remains labor-intensive and time-consuming. In this study, we present a prototype vision-based monitoring system that automates the detection and tracking of horses and people inside stables using object detection and multi-object tracking techniques. The system leverages YOLOv11 and BoT-SORT for detection and tracking, while event states are inferred based on object trajectories and spatial relations within the stall. To support development, we constructed a custom dataset annotated with assistance from foundation models CLIP and GroundingDINO. The system distinguishes between five event types and accounts for the camera’s blind spots. Qualitative evaluation demonstrated reliable performance for horse-related events, while highlighting limitations in detecting people due to data scarcity. This work provides a foundation for real-time behavioral monitoring in equine facilities, with implications for animal welfare and stable management.

[346] DeepDetect: Learning All-in-One Dense Keypoints

Shaharyar Ahmed Khan Tareen, Filza Khan Tareen

Main category: cs.CV

TL;DR: DeepDetect is a dense keypoint detector that combines classical detectors with deep learning to overcome limitations of existing methods in photometric sensitivity, keypoint density, and semantic understanding.

DetailsMotivation: Traditional and learning-based keypoint detectors suffer from sensitivity to photometric changes, low keypoint density and repeatability, limited adaptability to challenging scenes, and lack of semantic understanding.

Method: Fuses outputs from 7 keypoint and 2 edge detectors to create ground-truth masks, then trains a lightweight ESPNet model using these masks to produce dense, semantically-aware keypoints.

Result: Outperforms other detectors on Oxford Affine Covariant Regions dataset with maximum values of 0.5143 (keypoint density), 0.9582 (repeatability), and 59,003 (correct matches).

Conclusion: DeepDetect successfully unifies classical detector strengths with deep learning to achieve superior performance in keypoint detection across diverse and challenging conditions.

Abstract: Keypoint detection is the foundation of many computer vision tasks, including image registration, structure-from motion, 3D reconstruction, visual odometry, and SLAM. Traditional detectors (SIFT, SURF, ORB, BRISK, etc.) and learning based methods (SuperPoint, R2D2, LF-Net, D2-Net, etc.) have shown strong performance yet suffer from key limitations: sensitivity to photometric changes, low keypoint density and repeatability, limited adaptability to challenging scenes, and lack of semantic understanding, often failing to prioritize visually important regions. We present DeepDetect, an intelligent, all-in-one, dense keypoint detector that unifies the strengths of classical detectors using deep learning. Firstly, we create ground-truth masks by fusing outputs of 7 keypoint and 2 edge detectors, extracting diverse visual cues from corners and blobs to prominent edges and textures in the images. Afterwards, a lightweight and efficient model: ESPNet, is trained using these masks as labels, enabling DeepDetect to focus semantically on images while producing highly dense keypoints, that are adaptable to diverse and visually degraded conditions. Evaluations on the Oxford Affine Covariant Regions dataset demonstrate that DeepDetect surpasses other detectors in keypoint density, repeatability, and the number of correct matches, achieving maximum values of 0.5143 (average keypoint density), 0.9582 (average repeatability), and 59,003 (correct matches).

[347] Leveraging AV1 motion vectors for Fast and Dense Feature Matching

Julien Zouein, Hossein Javidnia, François Pitié, Anil Kokaram

Main category: cs.CV

TL;DR: Repurposing AV1 motion vectors for dense sub-pixel correspondences and filtered tracks, achieving comparable performance to SIFT with less CPU usage and denser matches.

DetailsMotivation: To create a resource-efficient compressed-domain front end for computer vision pipelines by leveraging existing motion vectors from video compression.

Method: Using AV1 motion vectors to generate dense sub-pixel correspondences and short tracks filtered by cosine consistency, with evaluation on short videos.

Result: Comparable performance to sequential SIFT with far less CPU usage, denser matches with competitive pairwise geometry, and successful SfM demo on 117-frame clip with 0.46-0.62M points reconstructed at 0.51-0.53px reprojection error.

Conclusion: Compressed-domain correspondences are a practical, resource-efficient front end with clear paths to scaling in full pipelines.

Abstract: We repurpose AV1 motion vectors to produce dense sub-pixel correspondences and short tracks filtered by cosine consistency. On short videos, this compressed-domain front end runs comparably to sequential SIFT while using far less CPU, and yields denser matches with competitive pairwise geometry. As a small SfM demo on a 117-frame clip, MV matches register all images and reconstruct 0.46-0.62M points at 0.51-0.53,px reprojection error; BA time grows with match density. These results show compressed-domain correspondences are a practical, resource-efficient front end with clear paths to scaling in full pipelines.

[348] Rethinking Nighttime Image Deraining via Learnable Color Space Transformation

Qiyuan Guan, Xiang Chen, Guiyue Jin, Jiyu Jin, Shumin Fan, Tianyu Song, Jinshan Pan

Main category: cs.CV

TL;DR: This paper introduces HQ-NightRain, a high-quality benchmark for nighttime image deraining, and proposes CST-Net with learnable color space conversion and implicit illumination guidance to effectively remove rain from nighttime scenes.

DetailsMotivation: Nighttime image deraining is more challenging than daytime due to complex nighttime scenarios and lack of high-quality datasets that accurately represent the coupling between rain and illumination effects.

Method: Proposes Color Space Transformation Network (CST-Net) with learnable color space converter to facilitate rain removal in Y channel where rain is more pronounced, and introduces implicit illumination guidance to capture illumination information for better robustness.

Result: Extensive experiments demonstrate the value of the new HQ-NightRain dataset and the effectiveness of the proposed CST-Net method for nighttime image deraining.

Conclusion: The paper provides a high-quality benchmark and an effective network architecture that addresses the unique challenges of nighttime image deraining through color space transformation and illumination guidance.

Abstract: Compared to daytime image deraining, nighttime image deraining poses significant challenges due to inherent complexities of nighttime scenarios and the lack of high-quality datasets that accurately represent the coupling effect between rain and illumination. In this paper, we rethink the task of nighttime image deraining and contribute a new high-quality benchmark, HQ-NightRain, which offers higher harmony and realism compared to existing datasets. In addition, we develop an effective Color Space Transformation Network (CST-Net) for better removing complex rain from nighttime scenes. Specifically, we propose a learnable color space converter (CSC) to better facilitate rain removal in the Y channel, as nighttime rain is more pronounced in the Y channel compared to the RGB color space. To capture illumination information for guiding nighttime deraining, implicit illumination guidance is introduced enabling the learned features to improve the model’s robustness in complex scenarios. Extensive experiments show the value of our dataset and the effectiveness of our method. The source code and datasets are available at https://github.com/guanqiyuan/CST-Net.

[349] Initialize to Generalize: A Stronger Initialization Pipeline for Sparse-View 3DGS

Feng Zhou, Wenkai Guo, Pu Cao, Zhicheng Zhang, Jianqin Yin

Main category: cs.CV

TL;DR: Sparse-view 3D Gaussian Splatting suffers from overfitting to training views. The paper shows initialization is the key factor, not training-time constraints, and proposes three methods to improve SfM-based initialization for better sparse-view 3D reconstruction.

DetailsMotivation: Sparse-view 3D Gaussian Splatting overfits to training views, causing blurring in novel views. Prior work focused on initialization or training constraints, but controlled experiments reveal initialization is the decisive factor that determines performance limits.

Method: Three initialization improvements: (1) frequency-aware SfM using low-frequency view augmentation for better low-texture coverage, (2) 3DGS self-initialization that lifts photometric supervision to add points in SfM-sparse regions, (3) point-cloud regularization with geometric/visibility priors for multi-view consistency and uniform coverage.

Result: Experiments on LLFF and Mip-NeRF360 datasets show consistent improvements in sparse-view settings, establishing the approach as a stronger initialization strategy for 3D Gaussian Splatting.

Conclusion: Initialization is the primary factor in sparse-view 3DGS performance. The proposed three-pronged initialization approach effectively supplements SfM’s coverage gaps and provides a more reliable foundation for 3D Gaussian Splatting in sparse-view scenarios.

Abstract: Sparse-view 3D Gaussian Splatting (3DGS) often overfits to the training views, leading to artifacts like blurring in novel view rendering. Prior work addresses it either by enhancing the initialization (\emph{i.e.}, the point cloud from Structure-from-Motion (SfM)) or by adding training-time constraints (regularization) to the 3DGS optimization. Yet our controlled ablations reveal that initialization is the decisive factor: it determines the attainable performance band in sparse-view 3DGS, while training-time constraints yield only modest within-band improvements at extra cost. Given initialization’s primacy, we focus our design there. Although SfM performs poorly under sparse views due to its reliance on feature matching, it still provides reliable seed points. Thus, building on SfM, our effort aims to supplement the regions it fails to cover as comprehensively as possible. Specifically, we design: (i) frequency-aware SfM that improves low-texture coverage via low-frequency view augmentation and relaxed multi-view correspondences; (ii) 3DGS self-initialization that lifts photometric supervision into additional points, compensating SfM-sparse regions with learned Gaussian centers; and (iii) point-cloud regularization that enforces multi-view consistency and uniform spatial coverage through simple geometric/visibility priors, yielding a clean and reliable point cloud. Our experiments on LLFF and Mip-NeRF360 demonstrate consistent gains in sparse-view settings, establishing our approach as a stronger initialization strategy. Code is available at https://github.com/zss171999645/ItG-GS.

[350] SparseWorld: A Flexible, Adaptive, and Efficient 4D Occupancy World Model Powered by Sparse and Dynamic Queries

Chenxu Dang, Haiyan Liu, Guangjun Bao, Pei An, Xinyue Tang, Jie Ma, Bingchuan Sun, Yan Wang

Main category: cs.CV

TL;DR: SparseWorld is a novel 4D occupancy world model that uses sparse dynamic queries for flexible, adaptive, and efficient perception and forecasting in autonomous driving scenarios.

DetailsMotivation: Existing occupancy world models rely on static embeddings/grids that limit perception flexibility and have misalignment with real-world dynamics due to their "in-place classification" approach.

Method: Proposes Range-Adaptive Perception module with learnable queries modulated by ego vehicle states and temporal-spatial associations, and State-Conditioned Forecasting module using regression-guided formulation instead of classification.

Result: Achieves state-of-the-art performance across perception, forecasting, and planning tasks, with advantages in flexibility, adaptability, and efficiency.

Conclusion: SparseWorld demonstrates superior performance through its sparse dynamic query approach and regression-based forecasting, effectively addressing limitations of traditional occupancy models.

Abstract: Semantic occupancy has emerged as a powerful representation in world models for its ability to capture rich spatial semantics. However, most existing occupancy world models rely on static and fixed embeddings or grids, which inherently limit the flexibility of perception. Moreover, their ``in-place classification" over grids exhibits a potential misalignment with the dynamic and continuous nature of real scenarios.In this paper, we propose SparseWorld, a novel 4D occupancy world model that is flexible, adaptive, and efficient, powered by sparse and dynamic queries. We propose a Range-Adaptive Perception module, in which learnable queries are modulated by the ego vehicle states and enriched with temporal-spatial associations to enable extended-range perception. To effectively capture the dynamics of the scene, we design a State-Conditioned Forecasting module, which replaces classification-based forecasting with regression-guided formulation, precisely aligning the dynamic queries with the continuity of the 4D environment. In addition, We specifically devise a Temporal-Aware Self-Scheduling training strategy to enable smooth and efficient training. Extensive experiments demonstrate that SparseWorld achieves state-of-the-art performance across perception, forecasting, and planning tasks. Comprehensive visualizations and ablation studies further validate the advantages of SparseWorld in terms of flexibility, adaptability, and efficiency. The code is available at https://github.com/MSunDYY/SparseWorld.

[351] Split-Fuse-Transport: Annotation-Free Saliency via Dual Clustering and Optimal Transport Alignment

Muhammad Umer Ramzan, Ali Zia, Abdelwahed Khamis, Noman Ali, Usman Ali, Wei Xiang

Main category: cs.CV

TL;DR: AutoSOD is an unsupervised salient object detection method that uses POTNet to generate high-quality pseudo-masks without labels, achieving near-supervised performance.

DetailsMotivation: Salient object detection can reach near-supervised accuracy without labels if reliable pseudo-masks are available. Current methods underutilize global consistency when prototype quality is weak.

Method: Introduces POTNet with entropy-guided dual-clustering: spectral clustering for high-entropy pixels and k-means for low-entropy pixels, aligned by optimal transport. This generates sharp pseudo-masks that supervise a MaskFormer-style encoder-decoder.

Result: AutoSOD outperforms unsupervised methods by up to 26% and weakly supervised methods by up to 36% in F-measure, narrowing the gap to fully supervised models.

Conclusion: The proposed split-fuse-transport design enables high-quality pseudo-mask generation without handcrafted priors, making unsupervised SOD more accurate and efficient.

Abstract: Salient object detection (SOD) aims to segment visually prominent regions in images and serves as a foundational task for various computer vision applications. We posit that SOD can now reach near-supervised accuracy without a single pixel-level label, but only when reliable pseudo-masks are available. We revisit the prototype-based line of work and make two key observations. First, boundary pixels and interior pixels obey markedly different geometry; second, the global consistency enforced by optimal transport (OT) is underutilized if prototype quality is weak. To address this, we introduce POTNet, an adaptation of Prototypical Optimal Transport that replaces POT’s single k-means step with an entropy-guided dual-clustering head: high-entropy pixels are organized by spectral clustering, low-entropy pixels by k-means, and the two prototype sets are subsequently aligned by OT. This split-fuse-transport design yields sharper, part-aware pseudo-masks in a single forward pass, without handcrafted priors. Those masks supervise a standard MaskFormer-style encoder-decoder, giving rise to AutoSOD, an end-to-end unsupervised SOD pipeline that eliminates SelfMask’s offline voting yet improves both accuracy and training efficiency. Extensive experiments on five benchmarks show that AutoSOD outperforms unsupervised methods by up to 26% and weakly supervised methods by up to 36% in F-measure, further narrowing the gap to fully supervised models.

[352] Context-Aware Pseudo-Label Scoring for Zero-Shot Video Summarization

Yuanli Wu, Long Zhang, Yue Du, Bin Li

Main category: cs.CV

TL;DR: A rubric-guided pseudo-labeling framework that transforms limited ground-truth annotations into structured scoring rubrics to guide LLM-based video summarization, achieving strong zero-shot performance approaching supervised methods.

DetailsMotivation: Overcome limitations of supervised methods (high labeling costs, poor generalization) and unsupervised approaches (poor semantic capture) by creating a training-free zero-shot framework that stabilizes LLM prompting for video summarization.

Method: Transform small subset of ground-truth annotations into pseudo labels aggregated into dataset-adaptive scoring rubrics. Use contextual prompting where first/last segments are scored based on descriptions while intermediate ones incorporate adjacent scene summaries to assess narrative progression and redundancy.

Result: Achieved F1 scores of 57.58 on SumMe and 63.05 on TVSum, surpassing unsupervised and prior zero-shot baselines while approaching supervised performance.

Conclusion: Rubric-guided pseudo labeling effectively stabilizes LLM-based scoring and establishes a general, interpretable zero-shot paradigm for video summarization.

Abstract: With the rapid proliferation of video content across social media, surveillance, and education platforms, efficiently summarizing long videos into concise yet semantically faithful surrogates has become increasingly vital. Existing supervised methods achieve strong in-domain accuracy by learning from dense annotations but suffer from high labeling costs and limited cross-dataset generalization, while unsupervised approaches, though label-free, often fail to capture high-level human semantics and fine-grained narrative cues. More recently, zero-shot prompting pipelines have leveraged large language models (LLMs) for training-free video summarization, yet remain highly sensitive to handcrafted prompt templates and dataset-specific score normalization. To overcome these limitations, we introduce a rubric-guided, pseudo-labeled prompting framework that transforms a small subset of ground-truth annotations into high-confidence pseudo labels, which are aggregated into structured, dataset-adaptive scoring rubrics guiding interpretable scene evaluation. During inference, first and last segments are scored based solely on their descriptions, whereas intermediate ones incorporate brief contextual summaries of adjacent scenes to assess narrative progression and redundancy. This contextual prompting enables the LLM to balance local salience and global coherence without parameter tuning. On SumMe and TVSum, our method achieves F1 scores of \textbf{57.58} and \textbf{63.05}, surpassing unsupervised and prior zero-shot baselines while approaching supervised performance. The results demonstrate that rubric-guided pseudo labeling effectively stabilizes LLM-based scoring and establishes a general, interpretable zero-shot paradigm for video summarization.

[353] MUG-V 10B: High-efficiency Training Pipeline for Large Video Generation Models

Yongshun Zhang, Zhongyi Fan, Yonghang Zhang, Zhangzikang Li, Weifeng Chen, Zhongwei Feng, Chaoyue Wang, Peng Hou, Anxiang Zeng

Main category: cs.CV

TL;DR: A training framework for large-scale video generation models that optimizes data processing, model architecture, training strategy, and infrastructure, resulting in MUG-V 10B model that matches state-of-the-art performance and is fully open-sourced.

DetailsMotivation: Training large-scale video generation models is challenging due to cross-modal text-video alignment, long sequences, and complex spatiotemporal dependencies, requiring efficient solutions.

Method: Optimized four pillars: data processing, model architecture, training strategy, and infrastructure. Used curriculum-based pretraining, alignment-focused post-training, and Megatron-Core for efficient multi-node scaling.

Result: MUG-V 10B matches state-of-the-art video generators overall and surpasses leading open-source baselines on e-commerce video generation tasks in human evaluations. Achieved significant efficiency gains and performance improvements.

Conclusion: Successfully developed and open-sourced a complete large-scale video generation training stack including model weights, training code, and inference pipelines, enabling efficient training of video generation models.

Abstract: In recent years, large-scale generative models for visual content (\textit{e.g.,} images, videos, and 3D objects/scenes) have made remarkable progress. However, training large-scale video generation models remains particularly challenging and resource-intensive due to cross-modal text-video alignment, the long sequences involved, and the complex spatiotemporal dependencies. To address these challenges, we present a training framework that optimizes four pillars: (i) data processing, (ii) model architecture, (iii) training strategy, and (iv) infrastructure for large-scale video generation models. These optimizations delivered significant efficiency gains and performance improvements across all stages of data preprocessing, video compression, parameter scaling, curriculum-based pretraining, and alignment-focused post-training. Our resulting model, MUG-V 10B, matches recent state-of-the-art video generators overall and, on e-commerce-oriented video generation tasks, surpasses leading open-source baselines in human evaluations. More importantly, we open-source the complete stack, including model weights, Megatron-Core-based large-scale training code, and inference pipelines for video generation and enhancement. To our knowledge, this is the first public release of large-scale video generation training code that exploits Megatron-Core to achieve high training efficiency and near-linear multi-node scaling, details are available in \href{https://github.com/Shopee-MUG/MUG-V}{our webpage}.

[354] MambaX-Net: Dual-Input Mamba-Enhanced Cross-Attention Network for Longitudinal MRI Segmentation

Yovin Yahathugoda, Davide Prezzi, Piyalitt Ittichaiwong, Vicky Goh, Sebastien Ourselin, Michela Antonelli

Main category: cs.CV

TL;DR: MambaX-Net is a semi-supervised 3D segmentation model for longitudinal prostate cancer surveillance that leverages previous time point MRI and segmentation masks, using Mamba-enhanced cross-attention and shape encoding to improve temporal analysis with limited expert labels.

DetailsMotivation: Active Surveillance for prostate cancer requires accurate longitudinal segmentation, but existing models trained on single-time-point expert annotations are unsuitable for multi-time-point analysis with scarce labels.

Method: Proposed MambaX-Net with Mamba-enhanced Cross-Attention Module for temporal evolution and long-range dependencies, Shape Extractor Module for anatomical representation, and semi-supervised self-training using pseudo-labels from pre-trained nnU-Net.

Result: MambaX-Net significantly outperforms state-of-the-art U-Net and Transformer-based models on longitudinal AS dataset, achieving superior prostate zone segmentation with limited and noisy data.

Conclusion: The proposed architecture effectively addresses longitudinal segmentation challenges in Active Surveillance by leveraging temporal information and semi-supervised learning, enabling accurate prostate monitoring without extensive expert annotations.

Abstract: Active Surveillance (AS) is a treatment option for managing low and intermediate-risk prostate cancer (PCa), aiming to avoid overtreatment while monitoring disease progression through serial MRI and clinical follow-up. Accurate prostate segmentation is an important preliminary step for automating this process, enabling automated detection and diagnosis of PCa. However, existing deep-learning segmentation models are often trained on single-time-point and expertly annotated datasets, making them unsuitable for longitudinal AS analysis, where multiple time points and a scarcity of expert labels hinder their effective fine-tuning. To address these challenges, we propose MambaX-Net, a novel semi-supervised, dual-scan 3D segmentation architecture that computes the segmentation for time point t by leveraging the MRI and the corresponding segmentation mask from the previous time point. We introduce two new components: (i) a Mamba-enhanced Cross-Attention Module, which integrates the Mamba block into cross attention to efficiently capture temporal evolution and long-range spatial dependencies, and (ii) a Shape Extractor Module that encodes the previous segmentation mask into a latent anatomical representation for refined zone delination. Moreover, we introduce a semi-supervised self-training strategy that leverages pseudo-labels generated from a pre-trained nnU-Net, enabling effective learning without expert annotations. MambaX-Net was evaluated on a longitudinal AS dataset, and results showed that it significantly outperforms state-of-the-art U-Net and Transformer-based models, achieving superior prostate zone segmentation even when trained on limited and noisy data.

[355] WP-CrackNet: A Collaborative Adversarial Learning Framework for End-to-End Weakly-Supervised Road Crack Detection

Nachuan Ma, Zhengfei Song, Qiang Hu, Xiaoyu Tang, Chengxi Zhang, Rui Fan, Lihua Xie

Main category: cs.CV

TL;DR: WP-CrackNet is a weakly-supervised road crack detection method that uses only image-level labels for pixel-wise detection, achieving comparable performance to supervised methods through adversarial learning between classifier and reconstructor components.

DetailsMotivation: To reduce reliance on costly pixel-level annotations for road crack detection in smart cities, enabling more scalable infrastructure maintenance.

Method: End-to-end weakly-supervised framework with three components: classifier generating CAMs, reconstructor measuring feature inferability, and detector producing pixel-wise results. Uses adversarial learning between classifier and reconstructor, path-aware attention module (PAAM), and center-enhanced CAM consistency module (CECCM).

Result: Achieves comparable results to supervised methods and outperforms existing weakly-supervised methods on three created image-level datasets.

Conclusion: WP-CrackNet significantly advances scalable road inspection by enabling effective crack detection with minimal annotation requirements, making infrastructure maintenance more practical and cost-effective.

Abstract: Road crack detection is essential for intelligent infrastructure maintenance in smart cities. To reduce reliance on costly pixel-level annotations, we propose WP-CrackNet, an end-to-end weakly-supervised method that trains with only image-level labels for pixel-wise crack detection. WP-CrackNet integrates three components: a classifier generating class activation maps (CAMs), a reconstructor measuring feature inferability, and a detector producing pixel-wise road crack detection results. During training, the classifier and reconstructor alternate in adversarial learning to encourage crack CAMs to cover complete crack regions, while the detector learns from pseudo labels derived from post-processed crack CAMs. This mutual feedback among the three components improves learning stability and detection accuracy. To further boost detection performance, we design a path-aware attention module (PAAM) that fuses high-level semantics from the classifier with low-level structural cues from the reconstructor by modeling spatial and channel-wise dependencies. Additionally, a center-enhanced CAM consistency module (CECCM) is proposed to refine crack CAMs using center Gaussian weighting and consistency constraints, enabling better pseudo-label generation. We create three image-level datasets and extensive experiments show that WP-CrackNet achieves comparable results to supervised methods and outperforms existing weakly-supervised methods, significantly advancing scalable road inspection. The source code package and datasets are available at https://mias.group/WP-CrackNet/.

[356] PAGE-4D: Disentangled Pose and Geometry Estimation for 4D Perception

Kaichen Zhou, Yuhan Wang, Grace Chen, Xinhai Chang, Gaspard Beaudouin, Fangneng Zhan, Paul Pu Liang, Mengyu Wang

Main category: cs.CV

TL;DR: PAGE-4D extends VGGT to handle dynamic scenes by introducing a dynamics-aware aggregator that disentangles static and dynamic information, enabling simultaneous camera pose estimation, depth prediction, and point cloud reconstruction without post-processing.

DetailsMotivation: Existing 3D feed-forward models like VGGT struggle with dynamic elements in real-world scenarios because they are trained on static datasets, limiting their effectiveness in scenes with moving humans or deformable objects.

Method: Proposes a dynamics-aware aggregator that predicts a dynamics-aware mask to resolve task conflicts - suppressing motion cues for pose estimation while amplifying them for geometry reconstruction, enabling multi-task 4D reconstruction.

Result: Extensive experiments show PAGE-4D consistently outperforms VGGT in dynamic scenarios, achieving superior results in camera pose estimation, monocular/video depth estimation, and dense point map reconstruction.

Conclusion: PAGE-4D successfully extends static 3D models to dynamic scenes by addressing the inherent conflict between pose estimation and geometry reconstruction through dynamics-aware information disentanglement.

Abstract: Recent 3D feed-forward models, such as the Visual Geometry Grounded Transformer (VGGT), have shown strong capability in inferring 3D attributes of static scenes. However, since they are typically trained on static datasets, these models often struggle in real-world scenarios involving complex dynamic elements, such as moving humans or deformable objects like umbrellas. To address this limitation, we introduce PAGE-4D, a feedforward model that extends VGGT to dynamic scenes, enabling camera pose estimation, depth prediction, and point cloud reconstruction – all without post-processing. A central challenge in multi-task 4D reconstruction is the inherent conflict between tasks: accurate camera pose estimation requires suppressing dynamic regions, while geometry reconstruction requires modeling them. To resolve this tension, we propose a dynamics-aware aggregator that disentangles static and dynamic information by predicting a dynamics-aware mask – suppressing motion cues for pose estimation while amplifying them for geometry reconstruction. Extensive experiments show that PAGE-4D consistently outperforms the original VGGT in dynamic scenarios, achieving superior results in camera pose estimation, monocular and video depth estimation, and dense point map reconstruction.

[357] Expose Camouflage in the Water: Underwater Camouflaged Instance Segmentation and Dataset

Chuhong Wang, Hua Li, Chongyi Li, Huazhong Liu, Xiongxin Tang, Sam Kwong

Main category: cs.CV

TL;DR: The paper introduces UCIS4K, the first underwater camouflaged instance segmentation dataset, and proposes UCIS-SAM, a novel network with three key modules to address underwater vision challenges including color distortion, low contrast, and blurring.

DetailsMotivation: Traditional camouflaged instance segmentation methods trained on terrestrial datasets perform poorly in underwater environments due to degraded conditions like color distortion, low contrast, and blurring, which make it challenging to segment objects that blend closely with their surroundings.

Method: Proposed UCIS-SAM network with three modules: Channel Balance Optimization Module (CBOM) for enhanced underwater feature learning, Frequency Domain True Integration Module (FDTIM) to emphasize intrinsic object features and reduce camouflage interference, and Multi-scale Feature Frequency Aggregation Module (MFFAM) to strengthen boundaries of low-contrast camouflaged instances.

Result: Extensive experiments on the proposed UCIS4K dataset and public benchmarks show that UCIS-SAM outperforms state-of-the-art approaches in underwater camouflaged instance segmentation.

Conclusion: The UCIS4K dataset and UCIS-SAM network effectively address the challenges of underwater camouflaged instance segmentation, providing superior performance compared to existing methods through specialized modules for underwater feature enhancement and camouflage pattern handling.

Abstract: With the development of underwater exploration and marine protection, underwater vision tasks are widespread. Due to the degraded underwater environment, characterized by color distortion, low contrast, and blurring, camouflaged instance segmentation (CIS) faces greater challenges in accurately segmenting objects that blend closely with their surroundings. Traditional camouflaged instance segmentation methods, trained on terrestrial-dominated datasets with limited underwater samples, may exhibit inadequate performance in underwater scenes. To address these issues, we introduce the first underwater camouflaged instance segmentation (UCIS) dataset, abbreviated as UCIS4K, which comprises 3,953 images of camouflaged marine organisms with instance-level annotations. In addition, we propose an Underwater Camouflaged Instance Segmentation network based on Segment Anything Model (UCIS-SAM). Our UCIS-SAM includes three key modules. First, the Channel Balance Optimization Module (CBOM) enhances channel characteristics to improve underwater feature learning, effectively addressing the model’s limited understanding of underwater environments. Second, the Frequency Domain True Integration Module (FDTIM) is proposed to emphasize intrinsic object features and reduce interference from camouflage patterns, enhancing the segmentation performance of camouflaged objects blending with their surroundings. Finally, the Multi-scale Feature Frequency Aggregation Module (MFFAM) is designed to strengthen the boundaries of low-contrast camouflaged instances across multiple frequency bands, improving the model’s ability to achieve more precise segmentation of camouflaged objects. Extensive experiments on the proposed UCIS4K and public benchmarks show that our UCIS-SAM outperforms state-of-the-art approaches.

[358] ShapeCraft: LLM Agents for Structured, Textured and Interactive 3D Modeling

Shuyuan Zhang, Chenhan Jiang, Zuoou Li, Jiankang Deng

Main category: cs.CV

TL;DR: ShapeCraft is a multi-agent framework that generates structured, interactive 3D assets from natural language using Graph-based Procedural Shape representation and LLM agents.

DetailsMotivation: To address limitations of existing text-to-3D methods that produce unstructured meshes with poor interactivity, making them impractical for artistic workflows.

Method: Proposes Graph-based Procedural Shape (GPS) representation that decomposes natural language into structured sub-task graphs, using LLM agents to hierarchically parse input and iteratively refine procedural modeling and painting.

Result: Qualitative and quantitative experiments show superior performance in generating geometrically accurate and semantically rich 3D assets compared to existing LLM-based methods, with demonstrated versatility in animation and user-customized editing.

Conclusion: ShapeCraft enables practical text-to-3D generation with structured, textured, and interactive assets, showing potential for broader interactive applications.

Abstract: 3D generation from natural language offers significant potential to reduce expert manual modeling efforts and enhance accessibility to 3D assets. However, existing methods often yield unstructured meshes and exhibit poor interactivity, making them impractical for artistic workflows. To address these limitations, we represent 3D assets as shape programs and introduce ShapeCraft, a novel multi-agent framework for text-to-3D generation. At its core, we propose a Graph-based Procedural Shape (GPS) representation that decomposes complex natural language into a structured graph of sub-tasks, thereby facilitating accurate LLM comprehension and interpretation of spatial relationships and semantic shape details. Specifically, LLM agents hierarchically parse user input to initialize GPS, then iteratively refine procedural modeling and painting to produce structured, textured, and interactive 3D assets. Qualitative and quantitative experiments demonstrate ShapeCraft’s superior performance in generating geometrically accurate and semantically rich 3D assets compared to existing LLM-based agents. We further show the versatility of ShapeCraft through examples of animated and user-customized editing, highlighting its potential for broader interactive applications.

[359] Integrating BIM and UAV-based photogrammetry for Automated 3D Structure Model Segmentation

Siqi Chen, Shanyue Guan

Main category: cs.CV

TL;DR: Proposes a machine learning framework using UAV-scanned point clouds and synthetic BIM data for automated segmentation of 3D infrastructure models, reducing manual labeling needs and improving efficiency.

DetailsMotivation: To overcome the time-consuming and error-prone manual labeling process required for segmenting structural components from UAV-captured 3D models in structural health monitoring.

Method: Combines real-world UAV-scanned point clouds with synthetic data generated from Building Information Modeling (BIM) to train machine learning models for automated 3D point cloud segmentation.

Result: Achieved high accuracy in identifying and segmenting railroad track components (rails and crossties) and significantly reduced training time while maintaining reasonable segmentation accuracy using smaller datasets supplemented with BIM data.

Conclusion: The automated approach improves precision and efficiency of 3D infrastructure model segmentation and advances integration of UAV and BIM technologies for structural health monitoring and infrastructure management.

Abstract: The advancement of UAV technology has enabled efficient, non-contact structural health monitoring. Combined with photogrammetry, UAVs can capture high-resolution scans and reconstruct detailed 3D models of infrastructure. However, a key challenge remains in segmenting specific structural components from these models-a process traditionally reliant on time-consuming and error-prone manual labeling. To address this issue, we propose a machine learning-based framework for automated segmentation of 3D point clouds. Our approach uses the complementary strengths of real-world UAV-scanned point clouds and synthetic data generated from Building Information Modeling (BIM) to overcome the limitations associated with manual labeling. Validation on a railroad track dataset demonstrated high accuracy in identifying and segmenting major components such as rails and crossties. Moreover, by using smaller-scale datasets supplemented with BIM data, the framework significantly reduced training time while maintaining reasonable segmentation accuracy. This automated approach improves the precision and efficiency of 3D infrastructure model segmentation and advances the integration of UAV and BIM technologies in structural health monitoring and infrastructure management.

[360] One Dinomaly2 Detect Them All: A Unified Framework for Full-Spectrum Unsupervised Anomaly Detection

Jia Guo, Shuai Lu, Lei Fan, Zelin Li, Donglin Di, Yang Song, Weihang Zhang, Wenbing Zhu, Hong Yan, Fang Chen, Huiqi Li, Hongen Liao

Main category: cs.CV

TL;DR: Dinomaly2 is a unified framework for unsupervised anomaly detection that bridges performance gaps in multi-class models while extending across diverse data modalities and task settings through simple, minimalistic design.

DetailsMotivation: Existing multi-class anomaly detection models significantly underperform specialized single-class models, and the field has fragmented into scenario-specific methods, creating deployment barriers that highlight the need for a unified solution.

Method: A reconstruction-based framework guided by ’less is more’ philosophy, using orchestration of five simple elements to achieve superior performance without modification across diverse tasks.

Result: Achieves unprecedented 99.9% and 99.3% image-level AUROC on MVTec-AD and VisA for multi-class models, state-of-the-art performance in multi-view/multi-modal inspection, and surpasses previous full-shot models using only 8 normal examples per class (98.7% and 97.4% I-AUROC).

Conclusion: The combination of minimalistic design, computational scalability, and universal applicability positions Dinomaly2 as a unified solution for the full spectrum of real-world anomaly detection applications.

Abstract: Unsupervised anomaly detection (UAD) has evolved from building specialized single-class models to unified multi-class models, yet existing multi-class models significantly underperform the most advanced one-for-one counterparts. Moreover, the field has fragmented into specialized methods tailored to specific scenarios (multi-class, 3D, few-shot, etc.), creating deployment barriers and highlighting the need for a unified solution. In this paper, we present Dinomaly2, the first unified framework for full-spectrum image UAD, which bridges the performance gap in multi-class models while seamlessly extending across diverse data modalities and task settings. Guided by the “less is more” philosophy, we demonstrate that the orchestration of five simple element achieves superior performance in a standard reconstruction-based framework. This methodological minimalism enables natural extension across diverse tasks without modification, establishing that simplicity is the foundation of true universality. Extensive experiments on 12 UAD benchmarks demonstrate Dinomaly2’s full-spectrum superiority across multiple modalities (2D, multi-view, RGB-3D, RGB-IR), task settings (single-class, multi-class, inference-unified multi-class, few-shot) and application domains (industrial, biological, outdoor). For example, our multi-class model achieves unprecedented 99.9% and 99.3% image-level (I-) AUROC on MVTec-AD and VisA respectively. For multi-view and multi-modal inspection, Dinomaly2 demonstrates state-of-the-art performance with minimum adaptations. Moreover, using only 8 normal examples per class, our method surpasses previous full-shot models, achieving 98.7% and 97.4% I-AUROC on MVTec-AD and VisA. The combination of minimalistic design, computational scalability, and universal applicability positions Dinomaly2 as a unified solution for the full spectrum of real-world anomaly detection applications.

[361] CaMiT: A Time-Aware Car Model Dataset for Classification and Generation

Frédéric LIN, Biruk Abere Ambaw, Adrian Popescu, Hejer Ammar, Romaric Audigier, Hervé Le Borgne

Main category: cs.CV

TL;DR: CaMiT dataset tracks car model evolution over time (2005-2023) with 787K labeled and 5.1M unlabeled samples. Methods include time-incremental pretraining and classifier learning to improve temporal robustness, plus time-aware image generation.

DetailsMotivation: AI systems need to adapt to evolving visual environments where object appearances change over time, particularly for technological artifacts like cars.

Method: Static pretraining on in-domain data; time-incremental classification with two strategies: updating backbone or final layer only; time-aware image generation using temporal metadata.

Result: Static pretraining achieves competitive performance but accuracy declines across years. Time-incremental approaches improve temporal robustness. Time-aware generation yields more realistic outputs.

Conclusion: CaMiT provides a benchmark for studying temporal adaptation in fine-grained visual recognition and generation, addressing the challenge of evolving object appearances over time.

Abstract: AI systems must adapt to evolving visual environments, especially in domains where object appearances change over time. We introduce Car Models in Time (CaMiT), a fine-grained dataset capturing the temporal evolution of car models, a representative class of technological artifacts. CaMiT includes 787K labeled samples of 190 car models (2007-2023) and 5.1M unlabeled samples (2005-2023), supporting both supervised and self-supervised learning. Static pretraining on in-domain data achieves competitive performance with large-scale generalist models while being more resource-efficient, yet accuracy declines when models are tested across years. To address this, we propose a time-incremental classification setting, a realistic continual learning scenario with emerging, evolving, and disappearing classes. We evaluate two strategies: time-incremental pretraining, which updates the backbone, and time-incremental classifier learning, which updates only the final layer, both improving temporal robustness. Finally, we explore time-aware image generation that leverages temporal metadata during training, yielding more realistic outputs. CaMiT offers a rich benchmark for studying temporal adaptation in fine-grained visual recognition and generation.

[362] Self-supervised Pre-training for Mapping of Archaeological Stone Wall in Historic Landscapes Using High-Resolution DEM Derivatives

Zexian Huang, Mashnoon Islam, Brian Armstrong, Kourosh Khoshelham, Martin Tomko

Main category: cs.CV

TL;DR: DINO-CV is a self-supervised segmentation framework that automatically maps dry-stone walls using LiDAR-derived DEMs, overcoming vegetation occlusion and data scarcity through cross-view pre-training.

DetailsMotivation: Dry-stone walls have heritage and environmental value but remain unmapped due to inaccessibility and high manual mapping costs. Visual occlusion by vegetation and limited labeled data hinder deep learning approaches.

Method: Uses high-resolution Airborne LiDAR-derived DEMs to capture terrain structures beneath vegetation. Introduces self-supervised cross-view pre-training based on knowledge distillation to learn invariant representations across multiple DEM derivatives, supporting various vision backbones.

Result: Achieved 68.6% mIoU on test areas in Budj Bim cultural landscape, maintaining 63.8% mIoU with only 10% labeled data. Successfully identified dense colonial dry-stone walls beyond Indigenous heritage contexts.

Conclusion: Self-supervised learning on high-resolution DEM derivatives enables effective automated dry-stone wall mapping in vegetated, heritage-rich environments with limited annotations.

Abstract: Dry-stone walls hold significant heritage and environmental value. Mapping these structures is essential for ecosystem preservation and wildfire management in Australia. Yet, many walls remain unidentified due to their inaccessibility and the high cost of manual mapping. Deep learning-based segmentation offers a scalable solution, but two major challenges persist: (1) visual occlusion of low-lying walls by dense vegetation, and (2) limited labeled data for supervised training. We propose DINO-CV, a segmentation framework for automatic mapping of low-lying dry-stone walls using high-resolution Airborne LiDAR-derived digital elevation models (DEMs). DEMs overcome visual occlusion by capturing terrain structures hidden beneath vegetation, enabling analysis of structural rather than spectral cues. DINO-CV introduces a self-supervised cross-view pre-training strategy based on knowledge distillation to mitigate data scarcity. It learns invariant visual and geometric representations across multiple DEM derivatives, supporting various vision backbones including ResNet, Wide ResNet, and Vision Transformers. Applied to the UNESCO World Heritage cultural landscape of Budj Bim, Victoria, the method identifies one of Australia’s densest collections of colonial dry-stone walls beyond Indigenous heritage contexts. DINO-CV achieves a mean Intersection over Union (mIoU) of 68.6% on test areas and maintains 63.8% mIoU when fine-tuned with only 10% labeled data. These results demonstrate the potential of self-supervised learning on high-resolution DEM derivatives for automated dry-stone wall mapping in vegetated and heritage-rich environments with scarce annotations.

[363] Frugal Federated Learning for Violence Detection: A Comparison of LoRA-Tuned VLMs and Personalized CNNs

Sébastien Thuau, Siba Haidar, Ayush Bajracharya, Rachid Chelouah

Main category: cs.CV

TL;DR: This paper compares two frugal federated learning approaches for violence detection: zero-shot/fine-tuned vision-language models (VLMs) and personalized compact 3D CNNs, evaluating accuracy, calibration, energy usage, and CO2 emissions under non-IID settings.

DetailsMotivation: To develop energy-efficient and environmentally sustainable AI solutions for violence detection in video surveillance, addressing the need for responsible, resource-aware systems that can operate effectively in federated learning settings.

Method: Two complementary strategies: (1) zero-shot and federated fine-tuning of vision-language models (LLaVA-7B) using LoRA, and (2) personalized training of a compact 3D CNN (65.8M parameters). Evaluation includes accuracy, calibration, energy usage, and CO2 emissions under realistic non-IID conditions.

Result: Both approaches exceed 90% accuracy. CNN3D slightly outperforms LoRA-tuned VLMs in ROC AUC and log loss while using less energy. VLMs remain better for contextual reasoning and multimodal inference. Energy and CO2 emissions were quantified across training and inference phases.

Conclusion: A hybrid model is recommended: lightweight CNNs for routine classification with selective VLM activation for complex scenarios. The framework provides a reproducible baseline for responsible, resource-aware AI in video surveillance with extensions toward real-time, multimodal systems.

Abstract: We examine frugal federated learning approaches to violence detection by comparing two complementary strategies: (i) zero-shot and federated fine-tuning of vision-language models (VLMs), and (ii) personalized training of a compact 3D convolutional neural network (CNN3D). Using LLaVA-7B and a 65.8M parameter CNN3D as representative cases, we evaluate accuracy, calibration, and energy usage under realistic non-IID settings. Both approaches exceed 90% accuracy. CNN3D slightly outperforms Low-Rank Adaptation(LoRA)-tuned VLMs in ROC AUC and log loss, while using less energy. VLMs remain favorable for contextual reasoning and multimodal inference. We quantify energy and CO$_2$ emissions across training and inference, and analyze sustainability trade-offs for deployment. To our knowledge, this is the first comparative study of LoRA-tuned vision-language models and personalized CNNs for federated violence detection, with an emphasis on energy efficiency and environmental metrics. These findings support a hybrid model: lightweight CNNs for routine classification, with selective VLM activation for complex or descriptive scenarios. The resulting framework offers a reproducible baseline for responsible, resource-aware AI in video surveillance, with extensions toward real-time, multimodal, and lifecycle-aware systems.

[364] 4DSegStreamer: Streaming 4D Panoptic Segmentation via Dual Threads

Ling Liu, Jun Tian, Li Yi

Main category: cs.CV

TL;DR: 4DSegStreamer is a dual-thread framework for real-time 4D panoptic segmentation in streaming scenarios, enabling efficient processing of dynamic environments through predictive forecasting and timely inference.

DetailsMotivation: Address the need for real-time, fine-grained perception in highly dynamic environments like crowd evacuation and autonomous driving, where constrained time budgets require efficient streaming processing.

Method: Uses a Dual-Thread System with a predictive thread that leverages historical motion/geometric data to forecast future dynamics, and an inference thread that ensures timely predictions by aligning with latest memory while compensating for ego-motion and object movements.

Result: Demonstrates superior robustness under high FPS conditions, accurately predicts dynamic objects in complex scenes, and shows effectiveness on both indoor (HOI4D) and outdoor (SemanticKITTI, nuScenes) datasets.

Conclusion: 4DSegStreamer provides a general framework that can be integrated into existing 3D/4D segmentation methods to enable real-time capability while maintaining robust performance in dynamic streaming environments.

Abstract: 4D panoptic segmentation in a streaming setting is critical for highly dynamic environments, such as evacuating dense crowds and autonomous driving in complex scenarios, where real-time, fine-grained perception within a constrained time budget is essential. In this paper, we introduce 4DSegStreamer, a novel framework that employs a Dual-Thread System to efficiently process streaming frames. The framework is general and can be seamlessly integrated into existing 3D and 4D segmentation methods to enable real-time capability. It also demonstrates superior robustness compared to existing streaming perception approaches, particularly under high FPS conditions. The system consists of a predictive thread and an inference thread. The predictive thread leverages historical motion and geometric information to extract features and forecast future dynamics. The inference thread ensures timely prediction for incoming frames by aligning with the latest memory and compensating for ego-motion and dynamic object movements. We evaluate 4DSegStreamer on the indoor HOI4D dataset and the outdoor SemanticKITTI and nuScenes datasets. Comprehensive experiments demonstrate the effectiveness of our approach, particularly in accurately predicting dynamic objects in complex scenes.

[365] PICABench: How Far Are We from Physically Realistic Image Editing?

Yuandong Pu, Le Zhuo, Songhao Han, Jinbo Xing, Kaiwen Zhu, Shuo Cao, Bin Fu, Si Liu, Hongsheng Li, Yu Qiao, Wenlong Zhang, Xi Chen, Yihao Liu

Main category: cs.CV

TL;DR: The paper introduces PICABench, a benchmark for evaluating physical realism in image editing, and PICAEval, an evaluation protocol using VLM-as-a-judge with human annotations. It also explores learning physics from videos and constructs PICA-100K dataset.

DetailsMotivation: Existing image editing models focus on instruction completion but overlook physical effects like shadows, reflections, and object interactions, which are crucial for realism.

Method: Systematic evaluation across eight physical sub-dimensions (optics, mechanics, state transitions) for common editing operations. Uses VLM-as-a-judge with region-level human annotations and questions. Explores learning physics from videos and builds PICA-100K dataset.

Result: Evaluation of mainstream models shows physical realism remains challenging with significant room for improvement.

Conclusion: Physical realism is a key challenge in image editing. The benchmark and proposed solutions provide a foundation for moving from naive content editing to physically consistent realism.

Abstract: Image editing has achieved remarkable progress recently. Modern editing models could already follow complex instructions to manipulate the original content. However, beyond completing the editing instructions, the accompanying physical effects are the key to the generation realism. For example, removing an object should also remove its shadow, reflections, and interactions with nearby objects. Unfortunately, existing models and benchmarks mainly focus on instruction completion but overlook these physical effects. So, at this moment, how far are we from physically realistic image editing? To answer this, we introduce PICABench, which systematically evaluates physical realism across eight sub-dimension (spanning optics, mechanics, and state transitions) for most of the common editing operations (add, remove, attribute change, etc). We further propose the PICAEval, a reliable evaluation protocol that uses VLM-as-a-judge with per-case, region-level human annotations and questions. Beyond benchmarking, we also explore effective solutions by learning physics from videos and construct a training dataset PICA-100K. After evaluating most of the mainstream models, we observe that physical realism remains a challenging problem with large rooms to explore. We hope that our benchmark and proposed solutions can serve as a foundation for future work moving from naive content editing toward physically consistent realism.

[366] Intelligent Communication Mixture-of-Experts Boosted-Medical Image Segmentation Foundation Model

Xinwei Zhang, Hu Chen, Zhe Yuan, Sukun Tian, Peng Feng

Main category: cs.CV

TL;DR: IC-MoE is a medical image segmentation foundation model that addresses limitations in existing fine-tuning methods by using mixture-of-experts with adaptive voting and semantic-guided contrastive learning to enhance high-level feature representation while preserving pretrained weight integrity.

DetailsMotivation: Existing fine-tuning methods for medical image segmentation have two main limitations: insufficient representation of high-level features and disruption of pretrained weight structural integrity during fine-tuning.

Method: 1) Constructs three types of experts (basic, semantic, adaptive) with pixel probability adaptive voting for expert selection and fusion. 2) Uses semantic-guided contrastive learning to address weak supervision in contrastive learning.

Result: Extensive experiments on three public medical image segmentation datasets show IC-MoE outperforms other state-of-the-art models and demonstrates superior generalizability across diverse medical image segmentation scenarios.

Conclusion: IC-MoE effectively supplements foundational medical image segmentation models with enhanced high-level features and preserved pretrained structural integrity.

Abstract: Foundation models for medical image segmentation have achieved remarkable performance. Adaptive fine-tuning of natural image segmentation foundation models is crucial for medical image segmentation tasks. However, some limitations exist in existing fine-tuning methods: 1) insufficient representation of high-level features and 2) the fine-tuning process disrupts the structural integrity of pretrained weights. Inspired by these critical problems, we propose an intelligent communication mixture-of-experts boosted-medical image segmentation foundation model, named IC-MoE, with twofold ideas: 1) We construct basic experts, semantic experts, and adaptive experts. Moreover, we implement a pixel probability adaptive voting strategy, which enables expert selection and fusion through label consistency and load balancing. This approach preliminarily enhances the representation capability of high-level features while preserving the structural integrity of pretrained weights. 2) We propose a semantic-guided contrastive learning method to address the issue of weak supervision in contrastive learning. This method further enhances the representation capability of high-level features while preserving the structural integrity of pretrained weights. Extensive experiments across three public medical image segmentation datasets demonstrate that the IC-MoE outperforms other SOTA models. Consequently, the proposed IC-MoE effectively supplements foundational medical image segmentation models with high-level features and pretrained structural integrity. We also validate the superior generalizability of the IC-MoE across diverse medical image segmentation scenarios.

[367] Multilingual Text-to-Image Person Retrieval via Bidirectional Relation Reasoning and Aligning

Min Cao, Xinyu Zhou, Ding Jiang, Bo Du, Mang Ye, Min Zhang

Main category: cs.CV

TL;DR: Proposes Bi-IRRA framework for multilingual text-to-image person retrieval with bidirectional implicit relation reasoning and multi-dimensional global alignment, achieving SOTA results.

DetailsMotivation: Address limitations in current TIPR methods: global methods overlook fine-grained differences, local methods need prior information, and existing approaches are English-centric, restricting multilingual applications.

Method: Develops Bi-IRRA framework with bidirectional implicit relation reasoning module for masked image/text prediction and multi-dimensional global alignment module to bridge modality heterogeneity. Also creates multilingual TIPR benchmark using LLMs with domain-specific refinement.

Result: Achieves new state-of-the-art results on all multilingual TIPR datasets.

Conclusion: Bi-IRRA effectively addresses multilingual TIPR challenges through implicit relation reasoning and global alignment, providing a robust solution for cross-modal person retrieval in multilingual contexts.

Abstract: Text-to-image person retrieval (TIPR) aims to identify the target person using textual descriptions, facing challenge in modality heterogeneity. Prior works have attempted to address it by developing cross-modal global or local alignment strategies. However, global methods typically overlook fine-grained cross-modal differences, whereas local methods require prior information to explore explicit part alignments. Additionally, current methods are English-centric, restricting their application in multilingual contexts. To alleviate these issues, we pioneer a multilingual TIPR task by developing a multilingual TIPR benchmark, for which we leverage large language models for initial translations and refine them by integrating domain-specific knowledge. Correspondingly, we propose Bi-IRRA: a Bidirectional Implicit Relation Reasoning and Aligning framework to learn alignment across languages and modalities. Within Bi-IRRA, a bidirectional implicit relation reasoning module enables bidirectional prediction of masked image and text, implicitly enhancing the modeling of local relations across languages and modalities, a multi-dimensional global alignment module is integrated to bridge the modality heterogeneity. The proposed method achieves new state-of-the-art results on all multilingual TIPR datasets. Data and code are presented in https://github.com/Flame-Chasers/Bi-IRRA.

[368] Towards 3D Objectness Learning in an Open World

Taichi Liu, Zhenyu Wang, Ruofeng Liu, Guang Wang, Desheng Zhang

Main category: cs.CV

TL;DR: OP3Det is a class-agnostic open-world 3D detector that detects any objects in 3D scenes without text prompts, leveraging 2D foundation models and cross-modal fusion to achieve generalized 3D object discovery.

DetailsMotivation: Traditional closed-set 3D detectors struggle with open-world scenarios, and existing 3D open-vocabulary models face challenges with vocabulary expansion and semantic overlap. There's insufficient research on learning generalized 3D objectness for detecting novel objects unseen during training.

Method: Proposes OP3Det using 2D semantic priors and 3D geometric priors for class-agnostic proposals. Integrates complementary information from point cloud and RGB images through cross-modal mixture of experts, dynamically routing uni-modal and multi-modal features to learn generalized 3D objectness.

Result: Significantly surpasses existing open-world 3D detectors by up to 16.0% in AR and achieves 13.5% improvement compared to closed-world 3D detectors.

Conclusion: OP3Det demonstrates extraordinary performance in open-world 3D object detection, effectively addressing the limitations of traditional closed-set detectors and existing open-vocabulary approaches through its prompt-free, class-agnostic design and cross-modal fusion strategy.

Abstract: Recent advancements in 3D object detection and novel category detection have made significant progress, yet research on learning generalized 3D objectness remains insufficient. In this paper, we delve into learning open-world 3D objectness, which focuses on detecting all objects in a 3D scene, including novel objects unseen during training. Traditional closed-set 3D detectors struggle to generalize to open-world scenarios, while directly incorporating 3D open-vocabulary models for open-world ability struggles with vocabulary expansion and semantic overlap. To achieve generalized 3D object discovery, We propose OP3Det, a class-agnostic Open-World Prompt-free 3D Detector to detect any objects within 3D scenes without relying on hand-crafted text prompts. We introduce the strong generalization and zero-shot capabilities of 2D foundation models, utilizing both 2D semantic priors and 3D geometric priors for class-agnostic proposals to broaden 3D object discovery. Then, by integrating complementary information from point cloud and RGB image in the cross-modal mixture of experts, OP3Det dynamically routes uni-modal and multi-modal features to learn generalized 3D objectness. Extensive experiments demonstrate the extraordinary performance of OP3Det, which significantly surpasses existing open-world 3D detectors by up to 16.0% in AR and achieves a 13.5% improvement compared to closed-world 3D detectors.

[369] GAS: Improving Discretization of Diffusion ODEs via Generalized Adversarial Solver

Aleksandr Oganov, Ilya Bykov, Eva Neudachina, Mishan Aliev, Alexander Tolmachev, Alexander Sidorov, Aleksandr Zuev, Andrey Okhotin, Denis Rakitin, Aibek Alanov

Main category: cs.CV

TL;DR: The paper introduces Generalized Adversarial Solver (GAS), a simple parameterization of ODE diffusion solvers that combines distillation loss with adversarial training to improve sampling efficiency while preserving fine-grained details.

DetailsMotivation: Diffusion models achieve state-of-the-art generation quality but suffer from computationally expensive sampling, requiring dozens of function evaluations. Existing gradient-based optimization methods for distillation often rely on intricate training techniques and don't explicitly preserve fine-grained details.

Method: Proposes Generalized Solver - a simple parameterization of ODE sampler without additional training tricks, combined with adversarial training to mitigate artifacts and enhance detail fidelity. The method uses both original distillation loss and adversarial training.

Result: The Generalized Adversarial Solver demonstrates superior performance compared to existing solver training methods under similar resource constraints, reducing function evaluations from dozens to just a few while improving quality.

Conclusion: The proposed method provides an effective solution for efficient diffusion model sampling that maintains generation quality and preserves fine details through a combination of simple parameterization and adversarial training.

Abstract: While diffusion models achieve state-of-the-art generation quality, they still suffer from computationally expensive sampling. Recent works address this issue with gradient-based optimization methods that distill a few-step ODE diffusion solver from the full sampling process, reducing the number of function evaluations from dozens to just a few. However, these approaches often rely on intricate training techniques and do not explicitly focus on preserving fine-grained details. In this paper, we introduce the Generalized Solver: a simple parameterization of the ODE sampler that does not require additional training tricks and improves quality over existing approaches. We further combine the original distillation loss with adversarial training, which mitigates artifacts and enhances detail fidelity. We call the resulting method the Generalized Adversarial Solver and demonstrate its superior performance compared to existing solver training methods under similar resource constraints. Code is available at https://github.com/3145tttt/GAS.

[370] Improving Cross-Patient Generalization in Parkinson’s Disease Detection through Chunk-Based Analysis of Hand-Drawn Patterns

Mhd Adnan Albani, Riad Sonbol

Main category: cs.CV

TL;DR: A two-stage approach for Parkinson’s disease detection from hand-drawn images using chunking and ensemble methods, achieving high accuracy with minimal performance drop on unseen patient data.

DetailsMotivation: Address limitations in existing Parkinson's disease detection methods: lack of sufficient datasets and poor robustness with unseen patient data.

Method: Two-stage approach: first classifies drawing types (circle, meander, spiral), then extracts features from 2x2 image chunks processed separately. Uses ensemble method to merge decisions from each chunk.

Result: Achieved 97.08% accuracy for seen patients and 94.91% for unseen patients on NewHandPD dataset, with only 2.17 percentage point gap compared to 4.76-point drop in prior work.

Conclusion: The proposed chunking and ensemble approach significantly improves robustness and generalization for Parkinson’s disease detection, particularly on unseen patient data.

Abstract: Parkinson’s disease (PD) is a neurodegenerative disease affecting about 1% of people over the age of 60, causing motor impairments that impede hand coordination activities such as writing and drawing. Many approaches have tried to support early detection of Parkinson’s disease based on hand-drawn images; however, we identified two major limitations in the related works: (1) the lack of sufficient datasets, (2) the robustness when dealing with unseen patient data. In this paper, we propose a new approach to detect Parkinson’s disease that consists of two stages: The first stage classifies based on their drawing type(circle, meander, spiral), and the second stage extracts the required features from the images and detects Parkinson’s disease. We overcame the previous two limitations by applying a chunking strategy where we divide each image into 2x2 chunks. Each chunk is processed separately when extracting features and recognizing Parkinson’s disease indicators. To make the final classification, an ensemble method is used to merge the decisions made from each chunk. Our evaluation shows that our proposed approach outperforms the top performing state-of-the-art approaches, in particular on unseen patients. On the NewHandPD dataset our approach, it achieved 97.08% accuracy for seen patients and 94.91% for unseen patients, our proposed approach maintained a gap of only 2.17 percentage points, compared to the 4.76-point drop observed in prior work.

[371] Elastic ViTs from Pretrained Models without Retraining

Walter Simoncini, Michael Dorkenwald, Tijmen Blankevoort, Cees G. M. Snoek, Yuki M. Asano

Main category: cs.CV

TL;DR: SnapViT is a post-pretraining structured pruning method that enables elastic inference across compute budgets for Vision Transformers, using evolutionary algorithms and gradient information without requiring retraining or labeled data.

DetailsMotivation: Vision foundation models are only available in limited pre-determined sizes, forcing sub-optimal deployment choices under real-world constraints.

Method: Combines gradient information with cross-network structure correlations approximated via evolutionary algorithm, uses self-supervised importance scoring, and is retraining-free.

Result: Superior performance over state-of-the-art methods across various sparsities, generates elastic models in <5 minutes on single A100 GPU that can be adjusted to any computational budget.

Conclusion: SnapViT provides an efficient pruning strategy for pretrained Vision Transformers with novel evolutionary approximation of Hessian structures, enabling flexible deployment without performance degradation.

Abstract: Vision foundation models achieve remarkable performance but are only available in a limited set of pre-determined sizes, forcing sub-optimal deployment choices under real-world constraints. We introduce SnapViT: Single-shot network approximation for pruned Vision Transformers, a new post-pretraining structured pruning method that enables elastic inference across a continuum of compute budgets. Our approach efficiently combines gradient information with cross-network structure correlations, approximated via an evolutionary algorithm, does not require labeled data, generalizes to models without a classification head, and is retraining-free. Experiments on DINO, SigLIPv2, DeIT, and AugReg models demonstrate superior performance over state-of-the-art methods across various sparsities, requiring less than five minutes on a single A100 GPU to generate elastic models that can be adjusted to any computational budget. Our key contributions include an efficient pruning strategy for pretrained Vision Transformers, a novel evolutionary approximation of Hessian off-diagonal structures, and a self-supervised importance scoring mechanism that maintains strong performance without requiring retraining or labels. Code and pruned models are available at: https://elastic.ashita.nl/

[372] Automatic Classification of Circulating Blood Cell Clusters based on Multi-channel Flow Cytometry Imaging

Suqiang Ma, Subhadeep Sengupta, Yao Lee, Beikang Gu, Xianyan Chen, Xianqiao Wang, Yang Liu, Mengjia Xu, Galit H. Frydman, He Li

Main category: cs.CV

TL;DR: A new computational framework using YOLOv11 and multi-channel fluorescence analysis achieves over 95% accuracy in automatically identifying circulating blood cell clusters and their cell types from flow cytometry images.

DetailsMotivation: Current computational approaches focus on single-cell analysis but lack tools for analyzing irregular-shaped cell clusters, which are important biomarkers for thrombosis, infection, and inflammation.

Method: Two-step analysis: 1) Fine-tuned YOLOv11 model classifies images into cell cluster vs non-cluster groups, 2) Cell type identification by overlaying cluster contours with multi-channel fluorescence stain regions to handle debris and artifacts.

Result: Achieved over 95% accuracy in both cluster classification and phenotype identification, outperforming traditional CNNs and Vision Transformers.

Conclusion: The automated framework effectively analyzes CCC images and has potential for broader applications in analyzing immune and tumor cell clusters across various diseases.

Abstract: Circulating blood cell clusters (CCCs) containing red blood cells (RBCs), white blood cells(WBCs), and platelets are significant biomarkers linked to conditions like thrombosis, infection, and inflammation. Flow cytometry, paired with fluorescence staining, is commonly used to analyze these cell clusters, revealing cell morphology and protein profiles. While computational approaches based on machine learning have advanced the automatic analysis of single-cell flow cytometry images, there is a lack of effort to build tools to automatically analyze images containing CCCs. Unlike single cells, cell clusters often exhibit irregular shapes and sizes. In addition, these cell clusters often consist of heterogeneous cell types, which require multi-channel staining to identify the specific cell types within the clusters. This study introduces a new computational framework for analyzing CCC images and identifying cell types within clusters. Our framework uses a two-step analysis strategy. First, it categorizes images into cell cluster and non-cluster groups by fine-tuning the You Only Look Once(YOLOv11) model, which outperforms traditional convolutional neural networks (CNNs), Vision Transformers (ViT). Then, it identifies cell types by overlaying cluster contours with regions from multi-channel fluorescence stains, enhancing accuracy despite cell debris and staining artifacts. This approach achieved over 95% accuracy in both cluster classification and phenotype identification. In summary, our automated framework effectively analyzes CCC images from flow cytometry, leveraging both bright-field and fluorescence data. Initially tested on blood cells, it holds potential for broader applications, such as analyzing immune and tumor cell clusters, supporting cellular research across various diseases.

[373] MT-Video-Bench: A Holistic Video Understanding Benchmark for Evaluating Multimodal LLMs in Multi-Turn Dialogues

Yaning Pan, Zekun Wang, Qianqian Xie, Yongqian Wen, Yuanxing Zhang, Guohui Zhang, Haoxuan Hu, Zhiyu Pan, Yibing Huang, Zhidong Gan, Yonghong Lin, An Ping, Tianhao Peng, Jiaheng Liu

Main category: cs.CV

TL;DR: MT-Video-Bench is a new benchmark for evaluating Multimodal Large Language Models (MLLMs) in multi-turn video dialogues, addressing limitations of existing single-turn QA benchmarks.

DetailsMotivation: Existing MLLM evaluation benchmarks are limited to single-turn question answering, which doesn't capture the complexity of real-world multi-turn dialogues in video understanding scenarios.

Method: Created MT-Video-Bench with 987 meticulously curated multi-turn dialogues from diverse domains, assessing six core competencies focused on perceptivity and interactivity, aligned with real-world applications like sports analysis and video tutoring.

Result: Extensive evaluation of state-of-the-art open-source and closed-source MLLMs revealed significant performance discrepancies and limitations in handling multi-turn video dialogues.

Conclusion: The benchmark exposes current MLLMs’ shortcomings in multi-turn video understanding and will be publicly available to advance future research in this area.

Abstract: The recent development of Multimodal Large Language Models (MLLMs) has significantly advanced AI’s ability to understand visual modalities. However, existing evaluation benchmarks remain limited to single-turn question answering, overlooking the complexity of multi-turn dialogues in real-world scenarios. To bridge this gap, we introduce MT-Video-Bench, a holistic video understanding benchmark for evaluating MLLMs in multi-turn dialogues. Specifically, our MT-Video-Bench mainly assesses six core competencies that focus on perceptivity and interactivity, encompassing 987 meticulously curated multi-turn dialogues from diverse domains. These capabilities are rigorously aligned with real-world applications, such as interactive sports analysis and multi-turn video-based intelligent tutoring. With MT-Video-Bench, we extensively evaluate various state-of-the-art open-source and closed-source MLLMs, revealing their significant performance discrepancies and limitations in handling multi-turn video dialogues. The benchmark will be publicly available to foster future research.

[374] Raindrop GS: A Benchmark for 3D Gaussian Splatting under Raindrop Conditions

Zhiqiang Teng, Beibei Lin, Tingting Chen, Zifeng Yuan, Xuanyi Li, Xuanyu Zhang, Shunli Zhang

Main category: cs.CV

TL;DR: RaindropGS is a benchmark for evaluating 3D Gaussian Splatting under real-world raindrop conditions, addressing limitations of existing synthetic benchmarks by using unconstrained raindrop-corrupted images and analyzing the full pipeline from pose estimation to reconstruction.

DetailsMotivation: 3D Gaussian Splatting suffers from severe degradation under raindrop conditions due to occlusions and distortions. Existing benchmarks use synthetic raindrops with known poses, but real-world scenarios involve pose estimation challenges and domain gaps between synthetic and real raindrops.

Method: Created RaindropGS benchmark with three components: data preparation (collecting real-world dataset with raindrop-focused, background-focused, and rain-free images), data processing, and raindrop-aware 3DGS evaluation including pose estimation, point cloud initialization, rain removal, and Gaussian training comparison.

Result: Revealed critical insights: camera focus position significantly impacts 3DGS reconstruction performance, and inaccurate pose/point cloud initialization interferes with reconstruction quality under raindrop conditions.

Conclusion: The benchmark establishes clear directions for developing more robust 3DGS methods by identifying performance limitations and the varying impact of different pipeline components in real-world raindrop scenarios.

Abstract: 3D Gaussian Splatting (3DGS) under raindrop conditions suffers from severe occlusions and optical distortions caused by raindrop contamination on the camera lens, substantially degrading reconstruction quality. Existing benchmarks typically evaluate 3DGS using synthetic raindrop images with known camera poses (constrained images), assuming ideal conditions. However, in real-world scenarios, raindrops often interfere with accurate camera pose estimation and point cloud initialization. Moreover, a significant domain gap between synthetic and real raindrops further impairs generalization. To tackle these issues, we introduce RaindropGS, a comprehensive benchmark designed to evaluate the full 3DGS pipeline-from unconstrained, raindrop-corrupted images to clear 3DGS reconstructions. Specifically, the whole benchmark pipeline consists of three parts: data preparation, data processing, and raindrop-aware 3DGS evaluation, including types of raindrop interference, camera pose estimation and point cloud initialization, single image rain removal comparison, and 3D Gaussian training comparison. First, we collect a real-world raindrop reconstruction dataset, in which each scene contains three aligned image sets: raindrop-focused, background-focused, and rain-free ground truth, enabling a comprehensive evaluation of reconstruction quality under different focus conditions. Through comprehensive experiments and analyses, we reveal critical insights into the performance limitations of existing 3DGS methods on unconstrained raindrop images and the varying impact of different pipeline components: the impact of camera focus position on 3DGS reconstruction performance, and the interference caused by inaccurate pose and point cloud initialization on reconstruction. These insights establish clear directions for developing more robust 3DGS methods under raindrop conditions.

[375] Signature Forgery Detection: Improving Cross-Dataset Generalization

Matheus Ramos Parracho

Main category: cs.CV

TL;DR: This paper investigates feature learning strategies for offline signature verification to improve cross-dataset generalization, comparing raw image processing with shell preprocessing methods across three public benchmarks.

DetailsMotivation: Automated signature verification is critical for banking and authentication, but current deep learning methods struggle with generalization across datasets due to handwriting style variations and different acquisition protocols.

Method: Developed two experimental pipelines: one using raw signature images and another using shell preprocessing. Tested on three public benchmarks (CEDAR, ICDAR, and GPDS Synthetic) to evaluate cross-dataset performance.

Result: The raw-image model achieved higher performance across benchmarks, while the shell-based model showed promising potential but no definitive superiority was established between the two approaches.

Conclusion: Both approaches have merits, with raw-image processing currently performing better, but shell preprocessing offers potential for future refinement toward robust cross-domain signature verification.

Abstract: Automated signature verification is a critical biometric technique used in banking, identity authentication, and legal documentation. Despite the notable progress achieved by deep learning methods, most approaches in offline signature verification still struggle to generalize across datasets, as variations in handwriting styles and acquisition protocols often degrade performance. This study investigates feature learning strategies for signature forgery detection, focusing on improving cross-dataset generalization – that is, model robustness when trained on one dataset and tested on another. Using three public benchmarks – CEDAR, ICDAR, and GPDS Synthetic – two experimental pipelines were developed: one based on raw signature images and another employing a preprocessing method referred to as shell preprocessing. Several behavioral patterns were identified and analyzed; however, no definitive superiority between the two approaches was established. The results show that the raw-image model achieved higher performance across benchmarks, while the shell-based model demonstrated promising potential for future refinement toward robust, cross-domain signature verification.

[376] Can Image-To-Video Models Simulate Pedestrian Dynamics?

Aaron Appelle, Jerome P. Lynch

Main category: cs.CV

TL;DR: Investigates if image-to-video models can generate realistic pedestrian movement patterns in crowded scenes using trajectory benchmarks.

DetailsMotivation: To determine whether diffusion transformer-based I2V models trained on large video datasets can produce realistic pedestrian dynamics in crowded public environments.

Method: Conditions I2V models on keyframes from pedestrian trajectory benchmarks and evaluates trajectory prediction performance using quantitative pedestrian dynamics measures.

Result: Not specified in the abstract - appears to be a research investigation rather than reporting completed results.

Conclusion: The study explores the potential of I2V models for generating realistic pedestrian movement patterns, suggesting these models may have applications in pedestrian dynamics modeling.

Abstract: Recent high-performing image-to-video (I2V) models based on variants of the diffusion transformer (DiT) have displayed remarkable inherent world-modeling capabilities by virtue of training on large scale video datasets. We investigate whether these models can generate realistic pedestrian movement patterns in crowded public scenes. Our framework conditions I2V models on keyframes extracted from pedestrian trajectory benchmarks, then evaluates their trajectory prediction performance using quantitative measures of pedestrian dynamics.

[377] Joint Multi-Condition Representation Modelling via Matrix Factorisation for Visual Place Recognition

Timur Ismagilov, Shakaiba Majeed, Michael Milford, Tan Viet Tuyen Nguyen, Sarvapali D. Ramchurn, Shoaib Ehsan

Main category: cs.CV

TL;DR: A training-free, descriptor-agnostic approach for multi-reference visual place recognition that uses matrix decomposition to jointly model places from multiple reference descriptors, enabling projection-based residual matching.

DetailsMotivation: To improve localization performance in visual place recognition using multiple reference sets under varying conditions, while avoiding the computational costs of increased data diversity and model complexity in deep learning approaches.

Method: Proposes a training-free, descriptor-agnostic approach that jointly models places using multiple reference descriptors via matrix decomposition into basis representations, enabling projection-based residual matching.

Result: Improves Recall@1 by up to ~18% over single-reference methods and outperforms multi-reference baselines across appearance and viewpoint changes, with ~5% gains on unstructured data.

Conclusion: The method demonstrates strong generalization while remaining lightweight, providing effective multi-reference visual place recognition without requiring training.

Abstract: We address multi-reference visual place recognition (VPR), where reference sets captured under varying conditions are used to improve localisation performance. While deep learning with large-scale training improves robustness, increasing data diversity and model complexity incur extensive computational cost during training and deployment. Descriptor-level fusion via voting or aggregation avoids training, but often targets multi-sensor setups or relies on heuristics with limited gains under appearance and viewpoint change. We propose a training-free, descriptor-agnostic approach that jointly models places using multiple reference descriptors via matrix decomposition into basis representations, enabling projection-based residual matching. We also introduce SotonMV, a structured benchmark for multi-viewpoint VPR. On multi-appearance data, our method improves Recall@1 by up to ~18% over single-reference and outperforms multi-reference baselines across appearance and viewpoint changes, with gains of ~5% on unstructured data, demonstrating strong generalisation while remaining lightweight.

[378] Towards Explainable Skin Cancer Classification: A Dual-Network Attention Model with Lesion Segmentation and Clinical Metadata Fusion

Md. Enamul Atiq, Shaikh Anowarul Fattah

Main category: cs.CV

TL;DR: A dual-encoder attention framework combining segmented lesions and clinical metadata improves skin cancer classification accuracy and interpretability.

DetailsMotivation: Automated skin cancer diagnosis faces challenges with high intra-class variability, subtle inter-class differences, and lack of interpretability in deep learning models, limiting clinical trust.

Method: Uses Deep-UNet with Dual Attention Gates and ASPP for lesion segmentation, then dual DenseNet201 encoders with multi-head cross-attention for classification, plus transformer-based metadata integration.

Result: Achieves state-of-the-art segmentation performance and significantly improves classification accuracy and average AUC on HAM10000, ISIC 2018 and 2019 datasets, with Grad-CAM confirming lesion-focused predictions.

Conclusion: Integrating precise lesion segmentation and clinical data with attention-based fusion creates a more accurate and interpretable skin cancer classification model.

Abstract: Skin cancer is a life-threatening disease where early detection significantly improves patient outcomes. Automated diagnosis from dermoscopic images is challenging due to high intra-class variability and subtle inter-class differences. Many deep learning models operate as “black boxes,” limiting clinical trust. In this work, we propose a dual-encoder attention-based framework that leverages both segmented lesions and clinical metadata to enhance skin lesion classification in terms of both accuracy and interpretability. A novel Deep-UNet architecture with Dual Attention Gates (DAG) and Atrous Spatial Pyramid Pooling (ASPP) is first employed to segment lesions. The classification stage uses two DenseNet201 encoders-one on the original image and another on the segmented lesion whose features are fused via multi-head cross-attention. This dual-input design guides the model to focus on salient pathological regions. In addition, a transformer-based module incorporates patient metadata (age, sex, lesion site) into the prediction. We evaluate our approach on the HAM10000 dataset and the ISIC 2018 and 2019 challenges. The proposed method achieves state-of-the-art segmentation performance and significantly improves classification accuracy and average AUC compared to baseline models. To validate our model’s reliability, we use Gradient-weighted Class Activation Mapping (Grad-CAM) to generate heatmaps. These visualizations confirm that our model’s predictions are based on the lesion area, unlike models that rely on spurious background features. These results demonstrate that integrating precise lesion segmentation and clinical data with attention-based fusion leads to a more accurate and interpretable skin cancer classification model.

[379] SparseVILA: Decoupling Visual Sparsity for Efficient VLM Inference

Samir Khaki, Junxian Guo, Jiaming Tang, Shang Yang, Yukang Chen, Konstantinos N. Plataniotis, Yao Lu, Song Han, Zhijian Liu

Main category: cs.CV

TL;DR: SparseVILA is a new paradigm for efficient Vision Language Model inference that decouples visual sparsity across prefilling and decoding stages, achieving up to 4.0× faster prefilling and 2.5× faster decoding while maintaining or improving accuracy.

DetailsMotivation: Vision Language Models face scalability limitations due to the growing number of visual tokens that dominate inference latency, requiring more efficient inference methods.

Method: SparseVILA decouples visual sparsity by pruning redundant visual tokens during prefill and retrieving only query-relevant tokens during decoding, built on an AWQ-optimized inference pipeline.

Result: Achieves up to 4.0× faster prefilling, 2.5× faster decoding, and 2.6× overall end-to-end speedup on long-context video tasks while improving accuracy on document-understanding and reasoning tasks.

Conclusion: SparseVILA establishes a new direction for efficient multimodal inference with a training-free, architecture-agnostic framework that accelerates large VLMs without sacrificing capability.

Abstract: Vision Language Models (VLMs) have rapidly advanced in integrating visual and textual reasoning, powering applications across high-resolution image understanding, long-video analysis, and multi-turn conversation. However, their scalability remains limited by the growing number of visual tokens that dominate inference latency. We present SparseVILA, a new paradigm for efficient VLM inference that decouples visual sparsity across the prefilling and decoding stages. SparseVILA distributes sparsity across stages by pruning redundant visual tokens during prefill and retrieving only query-relevant tokens during decoding. This decoupled design matches leading prefill pruning methods while preserving multi-turn fidelity by retaining most of the visual cache so that query-aware tokens can be retrieved at each conversation round. Built on an AWQ-optimized inference pipeline, SparseVILA achieves up to 4.0 times faster prefilling, 2.5 times faster decoding, and an overall 2.6 times end-to-end speedup on long-context video tasks – while improving accuracy on document-understanding and reasoning tasks. By decoupling query-agnostic pruning and query-aware retrieval, SparseVILA establishes a new direction for efficient multimodal inference, offering a training-free, architecture-agnostic framework for accelerating large VLMs without sacrificing capability.

[380] ConsistEdit: Highly Consistent and Precise Training-free Visual Editing

Zixin Yin, Ling-Hao Chen, Lionel Ni, Xili Dai

Main category: cs.CV

TL;DR: ConsistEdit is a novel attention control method for MM-DiT models that enables consistent text-guided editing across images and videos by using vision-only attention control, mask-guided pre-attention fusion, and differentiated token manipulation.

DetailsMotivation: Current training-free attention control methods struggle to balance strong editing strength with source consistency, especially in multi-round and video editing where errors accumulate. Existing methods enforce global consistency, limiting fine-grained attribute editing.

Method: Proposes ConsistEdit with three key components: vision-only attention control, mask-guided pre-attention fusion, and differentiated manipulation of query, key, and value tokens. Works across all inference steps and attention layers without handcraft.

Result: Achieves state-of-the-art performance across image and video editing tasks in both structure-consistent and structure-inconsistent scenarios. Enables robust multi-round and multi-region editing with progressive structural consistency adjustment.

Conclusion: ConsistEdit is the first method to perform editing across all inference steps and attention layers without manual intervention, significantly enhancing reliability and consistency for complex editing tasks.

Abstract: Recent advances in training-free attention control methods have enabled flexible and efficient text-guided editing capabilities for existing generation models. However, current approaches struggle to simultaneously deliver strong editing strength while preserving consistency with the source. This limitation becomes particularly critical in multi-round and video editing, where visual errors can accumulate over time. Moreover, most existing methods enforce global consistency, which limits their ability to modify individual attributes such as texture while preserving others, thereby hindering fine-grained editing. Recently, the architectural shift from U-Net to MM-DiT has brought significant improvements in generative performance and introduced a novel mechanism for integrating text and vision modalities. These advancements pave the way for overcoming challenges that previous methods failed to resolve. Through an in-depth analysis of MM-DiT, we identify three key insights into its attention mechanisms. Building on these, we propose ConsistEdit, a novel attention control method specifically tailored for MM-DiT. ConsistEdit incorporates vision-only attention control, mask-guided pre-attention fusion, and differentiated manipulation of the query, key, and value tokens to produce consistent, prompt-aligned edits. Extensive experiments demonstrate that ConsistEdit achieves state-of-the-art performance across a wide range of image and video editing tasks, including both structure-consistent and structure-inconsistent scenarios. Unlike prior methods, it is the first approach to perform editing across all inference steps and attention layers without handcraft, significantly enhancing reliability and consistency, which enables robust multi-round and multi-region editing. Furthermore, it supports progressive adjustment of structural consistency, enabling finer control.

[381] Dress Well via Fashion Cognitive Learning

Kaicheng Pang, Xingxing Zou, Waikeung Wong

Main category: cs.CV

TL;DR: The paper proposes a Fashion Cognitive Network (FCN) for personalized fashion recommendations using personal physical information, outperforming existing methods on the new O4U dataset.

DetailsMotivation: Current fashion compatibility models provide good outfit compositions but lack personalization based on individual physical characteristics, limiting their effectiveness for precise customer recommendations.

Method: FCN uses an outfit encoder with convolutional layers to create outfit embeddings and a Multi-label Graph Neural Network (ML-GCN) to learn label classifiers via stacked GCN, modeling relationships between outfit compositions and personal appearance features.

Result: Extensive experiments on the O4U dataset show FCN provides strong qualitative and quantitative improvements over alternative methods.

Conclusion: The proposed Fashion Cognitive Network effectively addresses personalized fashion recommendation by incorporating personal physical information, demonstrating superior performance compared to existing approaches.

Abstract: Fashion compatibility models enable online retailers to easily obtain a large number of outfit compositions with good quality. However, effective fashion recommendation demands precise service for each customer with a deeper cognition of fashion. In this paper, we conduct the first study on fashion cognitive learning, which is fashion recommendations conditioned on personal physical information. To this end, we propose a Fashion Cognitive Network (FCN) to learn the relationships among visual-semantic embedding of outfit composition and appearance features of individuals. FCN contains two submodules, namely outfit encoder and Multi-label Graph Neural Network (ML-GCN). The outfit encoder uses a convolutional layer to encode an outfit into an outfit embedding. The latter module learns label classifiers via stacked GCN. We conducted extensive experiments on the newly collected O4U dataset, and the results provide strong qualitative and quantitative evidence that our framework outperforms alternative methods.

[382] Privacy-Preserving Visual Localization with Event Cameras

Junho Kim, Young Min Kim, Ramzi Zahreddine, Weston A. Welge, Gurunandan Krishnan, Sizhuo Ma, Jian Wang

Main category: cs.CV

TL;DR: A client-server localization system using event cameras that addresses computational efficiency, robustness, and privacy concerns through event-to-image conversion and multi-level privacy protection techniques.

DetailsMotivation: Traditional client-server localization systems face challenges in computational efficiency, robustness, and privacy preservation during data transmission, especially for resource-limited edge devices that cannot store large-scale 3D maps.

Method: Uses event cameras for low energy consumption and small memory bandwidth. Applies event-to-image conversion to leverage mature image-based localization. Introduces two-level privacy protection: network level (split inference to hide user’s view) and sensor level (light-weight filtering to hide sensitive details like faces).

Result: Achieves robustness in low-light or fast-moving scenes. Provides significant privacy protection with small client-side computation and minimal localization performance loss. User study shows reduced feelings of insecurity.

Conclusion: The method serves as a practical building block for location-based services using event cameras, effectively addressing efficiency, robustness, and privacy challenges in client-server localization.

Abstract: We consider the problem of client-server localization, where edge device users communicate visual data with the service provider for locating oneself against a pre-built 3D map. This localization paradigm is a crucial component for location-based services in AR/VR or mobile applications, as it is not trivial to store large-scale 3D maps and process fast localization on resource-limited edge devices. Nevertheless, conventional client-server localization systems possess numerous challenges in computational efficiency, robustness, and privacy-preservation during data transmission. Our work aims to jointly solve these challenges with a localization pipeline based on event cameras. By using event cameras, our system consumes low energy and maintains small memory bandwidth. Then during localization, we propose applying event-to-image conversion and leverage mature image-based localization, which achieves robustness even in low-light or fast-moving scenes. To further enhance privacy protection, we introduce privacy protection techniques at two levels. Network level protection aims to hide the entire user’s view in private scenes using a novel split inference approach, while sensor level protection aims to hide sensitive user details such as faces with light-weight filtering. Both methods involve small client-side computation and localization performance loss, while significantly mitigating the feeling of insecurity as revealed in our user study. We thus project our method to serve as a building block for practical location-based services using event cameras. Project page including the code is available through this link: https://82magnolia.github.io/event_localization/.

[383] FireANTs: Adaptive Riemannian Optimization for Multi-Scale Diffeomorphic Matching

Rohit Jena, Pratik Chaudhari, James C. Gee

Main category: cs.CV

TL;DR: FireANTs is a fast, training-free multi-scale adaptive Riemannian optimization algorithm for dense diffeomorphic image matching that outperforms traditional methods in speed and competes with deep learning methods while using less memory.

DetailsMotivation: Existing diffeomorphic image matching methods are slow due to inefficient implementations and ill-conditioned optimization. Deep learning methods require extensive training, substantial memory, and fail to generalize across diverse distributions and modalities.

Method: Training-free, GPU-accelerated multi-scale Adaptive Riemannian Optimization algorithm for dense diffeomorphic image matching.

Result: FireANTs runs 2.5x faster than ANTs on CPU and up to 1200x faster on GPU. On GPU, it competes with deep learning methods in runtime while using 10x less memory. Shows robustness across modalities, species, and organs without domain-specific training.

Conclusion: FireANTs provides a fast, memory-efficient alternative to both traditional and deep learning registration methods, enabling hyperparameter grid search with significantly reduced resources and time.

Abstract: The paper proposes FireANTs, a multi-scale Adaptive Riemannian Optimization algorithm for dense diffeomorphic image matching. Existing state-of-the-art methods for diffeomorphic image matching are slow due to inefficient implementations and slow convergence due to the ill-conditioned nature of the optimization problem. Deep learning methods offer fast inference but require extensive training time, substantial inference memory, and fail to generalize across long-tailed distributions or diverse image modalities, necessitating costly retraining. We address these challenges by proposing a training-free, GPU-accelerated multi-scale Adaptive Riemannian Optimization algorithm for fast and accurate dense diffeomorphic image matching. FireANTs runs about 2.5x faster than ANTs on a CPU, and upto 1200x faster on a GPU. On a single GPU, FireANTs performs competitively with deep learning methods on inference runtime while consuming upto 10x less memory. FireANTs shows remarkable robustness to a wide variety of matching problems across modalities, species, and organs without any domain-specific training or tuning. Our framework allows hyperparameter grid search studies with significantly less resources and time compared to traditional and deep learning registration algorithms alike.

[384] Predicting High-precision Depth on Low-Precision Devices Using 2D Hilbert Curves

Mykhailo Uss, Ruslan Yermolenko, Oleksii Shashko, Olena Kolodiazhna, Ivan Safonov, Volodymyr Savin, Yoonjae Yeo, Seowon Ji, Jaeyun Jeong

Main category: cs.CV

TL;DR: Proposes a method to restore high-precision depth from low-bit DNN predictions by representing depth as Hilbert curve components, enabling 8-bit quantization with minimal computational overhead.

DetailsMotivation: Dense depth prediction DNNs have high computational complexity limiting use on low-end devices, and low-bit quantization struggles to represent high dynamic range depth.

Method: Represent high dynamic range depth as two low dynamic range components of a Hilbert curve, train full-precision DNN to predict these components, then use quantization with post-processing to reconstruct depth from low-bit predictions.

Result: Method increases bit precision by up to 3 bits with little computational overhead, reduces quantization error by up to 4.6 times, and enables effective 8-bit depth prediction.

Conclusion: The proposed Hilbert curve representation enables accurate depth prediction with 8-bit quantized DNNs, overcoming limitations of low-bit precision for high dynamic range depth.

Abstract: Dense depth prediction deep neural networks (DNN) have achieved impressive results for both monocular and binocular data, but still they are limited by high computational complexity, restricting their use on low-end devices. For better on-device efficiency and hardware utilization, weights and activations of the DNN should be converted to low-bit precision. However, this precision is not sufficient to represent high dynamic range depth. In this paper, we aim to overcome this limitation and restore high-precision depth from low-bit precision predictions. To achieve this, we propose to represent high dynamic range depth as two low dynamic range components of a Hilbert curve, and to train the full-precision DNN to directly predict the latter. For on-device deployment, we use standard quantization methods and add a post-processing step that reconstructs depth from the Hilbert curve components predicted in low-bit precision. Extensive experiments demonstrate that our method increases the bit precision of predicted depth by up to three bits with little computational overhead. We also observed a positive side effect of quantization error reduction by up to 4.6 times. Our method enables effective and accurate depth prediction with DNN weights and activations quantized to eight-bit precision.

[385] Eye-for-an-eye: Appearance Transfer with Semantic Correspondence in Diffusion Models

Sooyeon Go, Kyungmook Choi, Minjung Shin, Youngjung Uh

Main category: cs.CV

TL;DR: Training-free appearance transfer method that rearranges features based on dense semantic correspondences to better preserve target structure and reference appearance.

DetailsMotivation: Existing methods for appearance transfer don't reflect semantic correspondence well because they rely on self-attention similarity, which is insufficient for establishing proper correspondences between images.

Method: Explicitly rearranging features according to dense semantic correspondences rather than using query-key similarity in self-attention layers.

Result: Method shows superiority in preserving target structure and reflecting correct colors from reference, even when images are not aligned.

Conclusion: The proposed explicit feature rearrangement based on semantic correspondences outperforms existing methods for appearance transfer tasks.

Abstract: As pre-trained text-to-image diffusion models have become a useful tool for image synthesis, people want to specify the results in various ways. This paper tackles training-free appearance transfer, which produces an image with the structure of a target image from the appearance of a reference image. Existing methods usually do not reflect semantic correspondence, as they rely on query-key similarity within the self-attention layer to establish correspondences between images. To this end, we propose explicitly rearranging the features according to the dense semantic correspondences. Extensive experiments show the superiority of our method in various aspects: preserving the structure of the target and reflecting the correct color from the reference, even when the two images are not aligned.

[386] GeoReasoner: Geo-localization with Reasoning in Street Views using a Large Vision-Language Model

Ling Li, Yu Ye, Yao Zhou, Wei Zeng

Main category: cs.CV

TL;DR: GeoReasoner is a large vision-language model enhanced with human inference knowledge for geo-localization, achieving significant improvements over existing methods while requiring fewer training resources.

DetailsMotivation: Addressing the scarcity of high-quality training data for geo-localization and the lack of reasoning inference in existing street-view datasets, which often contain low-quality images without visual clues.

Method: Uses CLIP-based network to quantify locatability of street-view images, creates new dataset of highly locatable views, integrates human inference knowledge from geo-localization games, and trains GeoReasoner through dedicated reasoning and location-tuning stages.

Result: Outperforms counterpart LVLMs by more than 25% at country-level and 38% at city-level geo-localization tasks, and surpasses StreetCLIP performance while requiring fewer training resources.

Conclusion: The proposed approach effectively addresses data quality and reasoning challenges in geo-localization, demonstrating superior performance through human inference knowledge integration and optimized training methodology.

Abstract: This work tackles the problem of geo-localization with a new paradigm using a large vision-language model (LVLM) augmented with human inference knowledge. A primary challenge here is the scarcity of data for training the LVLM - existing street-view datasets often contain numerous low-quality images lacking visual clues, and lack any reasoning inference. To address the data-quality issue, we devise a CLIP-based network to quantify the degree of street-view images being locatable, leading to the creation of a new dataset comprising highly locatable street views. To enhance reasoning inference, we integrate external knowledge obtained from real geo-localization games, tapping into valuable human inference capabilities. The data are utilized to train GeoReasoner, which undergoes fine-tuning through dedicated reasoning and location-tuning stages. Qualitative and quantitative evaluations illustrate that GeoReasoner outperforms counterpart LVLMs by more than 25% at country-level and 38% at city-level geo-localization tasks, and surpasses StreetCLIP performance while requiring fewer training resources. The data and code are available at https://github.com/lingli1996/GeoReasoner.

[387] Enhancing Test Time Adaptation with Few-shot Guidance

Siqi Luo, Yi Xin, Yuntao Du, Zhongwei Wan, Tao Tan, Guangtao Zhai, Xiaohong Liu

Main category: cs.CV

TL;DR: FS-TTA introduces few-shot support sets to improve Test Time Adaptation, reducing blind exploration in unseen target domains through a two-stage framework with feature diversity augmentation and prototype memory bank guidance.

DetailsMotivation: Existing Test Time Adaptation methods lack reliable domain shift correction mechanisms and can be erratic in real-world applications when adapting pre-trained models to out-of-distribution streaming target data.

Method: Two-stage framework: (1) fine-tuning pre-trained source model with few-shot support set using feature diversity augmentation to avoid overfitting, (2) test time adaptation with prototype memory bank guidance to produce high-quality pseudo-labels for model adaptation.

Result: Superior performance and reliability demonstrated through extensive experiments on three cross-domain classification benchmarks.

Conclusion: FS-TTA provides a practical and effective approach for domain adaptation by leveraging few-shot support sets to enhance Test Time Adaptation performance and reliability.

Abstract: Deep neural networks often encounter significant performance drops while facing with domain shifts between training (source) and test (target) data. To address this issue, Test Time Adaptation (TTA) methods have been proposed to adapt pre-trained source model to handle out-of-distribution streaming target data. Although these methods offer some relief, they lack a reliable mechanism for domain shift correction, which can often be erratic in real-world applications. In response, we develop Few-Shot Test Time Adaptation (FS-TTA), a novel and practical setting that utilizes a few-shot support set on top of TTA. Adhering to the principle of few inputs, big gains, FS-TTA reduces blind exploration in unseen target domains. Furthermore, we propose a two-stage framework to tackle FS-TTA, including (i) fine-tuning the pre-trained source model with few-shot support set, along with using feature diversity augmentation module to avoid overfitting, (ii) implementing test time adaptation based on prototype memory bank guidance to produce high quality pseudo-label for model adaptation. Through extensive experiments on three cross-domain classification benchmarks, we demonstrate the superior performance and reliability of our FS-TTA and framework.

[388] Scalable Frame Sampling for Video Classification: A Semi-Optimal Policy Approach with Reduced Search Space

Junho Lee, Jeongwoo Shin, Seung Woo Ko, Seongsu Ha, Joonseok Lee

Main category: cs.CV

TL;DR: A semi-optimal frame sampling policy that reduces search space from O(T^N) to O(T) by selecting top N frames based on independently estimated per-frame confidence values.

DetailsMotivation: Frame sampling for video classification faces computational challenges due to the vast search space of binom(T,N), especially when N is large. Existing methods struggle with this complexity.

Method: Proposed semi-optimal policy that independently estimates the value of each frame using per-frame confidence, then selects the top N frames, reducing search space from O(T^N) to O(T).

Result: The semi-optimal policy efficiently approximates the optimal policy, particularly in practical settings. Extensive experiments show stable high performance across various datasets and model architectures regardless of N and T size.

Conclusion: The proposed approach provides an efficient solution to frame sampling by significantly reducing computational complexity while maintaining performance, making it suitable for practical video classification tasks.

Abstract: Given a video with $T$ frames, frame sampling is a task to select $N \ll T$ frames, so as to maximize the performance of a fixed video classifier. Not just brute-force search, but most existing methods suffer from its vast search space of $\binom{T}{N}$, especially when $N$ gets large. To address this challenge, we introduce a novel perspective of reducing the search space from $O(T^N)$ to $O(T)$. Instead of exploring the entire $O(T^N)$ space, our proposed semi-optimal policy selects the top $N$ frames based on the independently estimated value of each frame using per-frame confidence, significantly reducing the computational complexity. We verify that our semi-optimal policy can efficiently approximate the optimal policy, particularly under practical settings. Additionally, through extensive experiments on various datasets and model architectures, we demonstrate that learning our semi-optimal policy ensures stable and high performance regardless of the size of $N$ and $T$.

[389] Large Language Model-Guided Semantic Alignment for Human Activity Recognition

Hua Yan, Heng Tan, Yi Ding, Pengfei Zhou, Vinod Namboodiri, Yu Yang

Main category: cs.CV

TL;DR: LanHAR uses LLMs to generate semantic interpretations of sensor data for cross-dataset human activity recognition, improving performance on new datasets and activities.

DetailsMotivation: Address distribution gaps in HAR caused by variations in activity patterns, device types, and sensor placements across different datasets.

Method: Uses LLMs to generate semantic interpretations of sensor readings and activity labels, employs iterative re-generation for quality, and implements two-stage training to bridge semantic spaces.

Result: Significantly outperforms state-of-the-art methods in cross-dataset HAR and new activity recognition on five public datasets.

Conclusion: LanHAR effectively mitigates cross-dataset heterogeneity and enhances recognition of new activities through semantic interpretation, producing a lightweight sensor encoder suitable for mobile deployment.

Abstract: Human Activity Recognition (HAR) using Inertial Measurement Unit (IMU) sensors is critical for applications in healthcare, safety, and industrial production. However, variations in activity patterns, device types, and sensor placements create distribution gaps across datasets, reducing the performance of HAR models. To address this, we propose LanHAR, a novel system that leverages Large Language Models (LLMs) to generate semantic interpretations of sensor readings and activity labels for cross-dataset HAR. This approach not only mitigates cross-dataset heterogeneity but also enhances the recognition of new activities. LanHAR employs an iterative re-generation method to produce high-quality semantic interpretations with LLMs and a two-stage training framework that bridges the semantic interpretations of sensor readings and activity labels. This ultimately leads to a lightweight sensor encoder suitable for mobile deployment, enabling any sensor reading to be mapped into the semantic interpretation space. Experiments on five public datasets demonstrate that our approach significantly outperforms state-of-the-art methods in both cross-dataset HAR and new activity recognition. The source code is publicly available at https://github.com/DASHLab/LanHAR.

[390] MaskControl: Spatio-Temporal Control for Masked Motion Synthesis

Ekkasit Pinyoanuntapong, Muhammad Usama Saleem, Korrawe Karunratanakul, Pu Wang, Hongfei Xue, Chen Chen, Chuan Guo, Junli Cao, Jian Ren, Sergey Tulyakov

Main category: cs.CV

TL;DR: MaskControl introduces controllability to masked motion models through Logits Regularizer and Logit Optimization with Differentiable Expectation Sampling, achieving superior motion quality and higher control precision compared to state-of-the-art methods.

DetailsMotivation: Existing motion diffusion models struggle to achieve high-precision control while maintaining high-quality motion generation, creating a need for better controllable motion generation approaches.

Method: Proposes MaskControl with two key innovations: Logits Regularizer for implicit perturbation during training to align motion tokens with controlled joint positions, and Logit Optimization for explicit optimization during inference to reshape token distribution. Also introduces Differentiable Expectation Sampling to handle non-differentiable distribution sampling.

Result: Outperforms state-of-the-art methods with FID decreasing by ~77% and achieving higher control precision (average error 0.91 vs. 1.08). Enables diverse applications including any-joint-any-frame control, body-part timeline control, and zero-shot objective control.

Conclusion: MaskControl successfully addresses the challenge of high-precision control in motion generation while maintaining motion quality, representing a significant advancement in controllable text-to-motion generation.

Abstract: Recent advances in motion diffusion models have enabled spatially controllable text-to-motion generation. However, these models struggle to achieve high-precision control while maintaining high-quality motion generation. To address these challenges, we propose MaskControl, the first approach to introduce controllability to the generative masked motion model. Our approach introduces two key innovations. First, \textit{Logits Regularizer} implicitly perturbs logits at training time to align the distribution of motion tokens with the controlled joint positions, while regularizing the categorical token prediction to ensure high-fidelity generation. Second, \textit{Logit Optimization} explicitly optimizes the predicted logits during inference time, directly reshaping the token distribution that forces the generated motion to accurately align with the controlled joint positions. Moreover, we introduce \textit{Differentiable Expectation Sampling (DES)} to combat the non-differential distribution sampling process encountered by logits regularizer and optimization. Extensive experiments demonstrate that MaskControl outperforms state-of-the-art methods, achieving superior motion quality (FID decreases by ~77%) and higher control precision (average error 0.91 vs. 1.08). Additionally, MaskControl enables diverse applications, including any-joint-any-frame control, body-part timeline control, and zero-shot objective control. Video visualization can be found at https://www.ekkasit.com/ControlMM-page/

[391] Enhancing Osteoporosis Detection: An Explainable Multi-Modal Learning Framework with Feature Fusion and Variable Clustering

Mehdi Hosseini Chagahi, Saeed Mohammadi Dashtaki, Niloufar Delfan, Nadia Mohammadi, Farshid Rostami Pouria, Behzad Moshiri, Md. Jalil Piran, Oliver Faust

Main category: cs.CV

TL;DR: A multi-modal learning framework combining clinical data and X-ray imaging features using pre-trained networks (VGG19, InceptionV3, ResNet50) with PCA and clustering for osteoporosis diagnosis, showing clinical data is more important than imaging features.

DetailsMotivation: Early osteoporosis diagnosis is crucial for fracture prevention, but healthcare faces challenges with limited labeled data and medical image processing difficulties.

Method: Uses three pre-trained networks to extract X-ray features, applies PCA for dimensionality reduction, clustering for component selection, combines with clinical data, and processes through FCN for classification.

Result: Clinical data (Medical History, BMI, Height) were main contributors to predictions, while imaging features had lower importance, demonstrating clinical data’s crucial role in accurate diagnosis.

Conclusion: The framework enables precise and interpretable osteoporosis predictions, enhancing AI transparency and trust for clinical integration.

Abstract: Osteoporosis is a common condition that increases fracture risk, especially in older adults. Early diagnosis is vital for preventing fractures, reducing treatment costs, and preserving mobility. However, healthcare providers face challenges like limited labeled data and difficulties in processing medical images. This study presents a novel multi-modal learning framework that integrates clinical and imaging data to improve diagnostic accuracy and model interpretability. The model utilizes three pre-trained networks-VGG19, InceptionV3, and ResNet50-to extract deep features from X-ray images. These features are transformed using PCA to reduce dimensionality and focus on the most relevant components. A clustering-based selection process identifies the most representative components, which are then combined with preprocessed clinical data and processed through a fully connected network (FCN) for final classification. A feature importance plot highlights key variables, showing that Medical History, BMI, and Height were the main contributors, emphasizing the significance of patient-specific data. While imaging features were valuable, they had lower importance, indicating that clinical data are crucial for accurate predictions. This framework promotes precise and interpretable predictions, enhancing transparency and building trust in AI-driven diagnoses for clinical integration.

[392] Delta-Influence: Unlearning Poisons via Influence Functions

Wenjie Li, Jiawei Li, Pengcheng Zeng, Christian Schroeder de Witt, Ameya Prabhu, Amartya Sanyal

Main category: cs.CV

TL;DR: Delta-Influence is a novel method that uses influence functions to identify poisoned training data from just one poisoned test example, enabling effective unlearning of data poisoning attacks.

DetailsMotivation: Address data integrity challenges by accurately attributing abnormal model behavior to specific poisoned training data, overcoming limitations of existing influence functions and unlearning algorithms.

Method: Leverages influence functions with data transformations that sever the link between poisoned training data and compromised test points, detecting influence collapse to identify poisoned samples.

Result: Consistently achieves the best unlearning performance across three vision-based poisoning attacks and three datasets, outperforming five detection algorithms and five unlearning strategies.

Conclusion: Demonstrates the promise of influence functions for corrective unlearning, effectively eliminating data poisoning through targeted retraining of identified poisoned subsets.

Abstract: Addressing data integrity challenges, such as unlearning the effects of data poisoning after model training, is necessary for the reliable deployment of machine learning models. State-of-the-art influence functions, such as EK-FAC and TRAK, often fail to accurately attribute abnormal model behavior to the specific poisoned training data responsible for the data poisoning attack. In addition, traditional unlearning algorithms often struggle to effectively remove the influence of poisoned samples, particularly when only a few affected examples can be identified. To address these challenge, we introduce $\Delta$-Influence, a novel approach that leverages influence functions to trace abnormal model behavior back to the responsible poisoned training data using as little as just one poisoned test example. $\Delta$-Influence applies data transformations that sever the link between poisoned training data and compromised test points without significantly affecting clean data. This allows $\Delta$-Influence to detect large negative shifts in influence scores following data transformations, a phenomenon we term as influence collapse, thereby accurately identifying poisoned training data. Unlearning this subset, e.g. through retraining, effectively eliminates the data poisoning. We validate our method across three vision-based poisoning attacks and three datasets, benchmarking against five detection algorithms and five unlearning strategies. We show that $\Delta$-Influence consistently achieves the best unlearning across all settings, showing the promise of influence functions for corrective unlearning. Our code is publicly available at: https://github.com/Ruby-a07/delta-influence

[393] VisualLens: Personalization through Task-Agnostic Visual History

Wang Bill Zhu, Deqing Fu, Kai Sun, Yi Lu, Zhaojiang Lin, Seungwhan Moon, Kanika Narang, Mustafa Canim, Yue Liu, Anuj Kumar, Xin Luna Dong

Main category: cs.CV

TL;DR: VisualLens is a framework that uses multimodal LLMs to extract user preferences from visual history (daily life images) for personalized recommendations, outperforming item-based methods and GPT-4o.

DetailsMotivation: Existing systems rely on user interaction logs or text signals, which are not always accessible or generalizable for multimodal recommendation. Visual history offers rich, task-agnostic insights into user preferences.

Method: Proposes VisualLens framework that extracts, filters, and refines user profiles from visual history using multimodal LLMs to support personalized recommendation.

Result: VisualLens improves over state-of-the-art item-based multimodal recommendations by 5-10% on Hit@3, outperforms GPT-4o by 2-5%, and shows robustness across varying history lengths and unseen content categories.

Conclusion: Visual history can be effectively leveraged for personalization, with VisualLens demonstrating superior performance and adaptability compared to existing methods.

Abstract: Existing recommendation systems either rely on user interaction logs, such as online shopping history for shopping recommendations, or focus on text signals. However, item-based histories are not always accessible, and are not generalizable for multimodal recommendation. We hypothesize that a user’s visual history – comprising images from daily life – can offer rich, task-agnostic insights into their interests and preferences, and thus be leveraged for effective personalization. To this end, we propose VisualLens, a novel framework that leverages multimodal large language models (MLLMs) to enable personalization using task-agnostic visual history. VisualLens extracts, filters, and refines a spectrum user profile from the visual history to support personalized recommendation. We created two new benchmarks, Google-Review-V and Yelp-V, with task-agnostic visual histories, and show that VisualLens improves over state-of-the-art item-based multimodal recommendations by 5-10% on Hit@3, and outperforms GPT-4o by 2-5%. Further analysis shows that VisualLens is robust across varying history lengths and excels at adapting to both longer histories and unseen content categories.

[394] Efficient Long-duration Talking Video Synthesis with Linear Diffusion Transformer under Multimodal Guidance

Haojie Zhang, Zhihao Liang, Ruibo Fu, Bingyan Liu, Zhengqi Wen, Xuefei Liu, Jianhua Tao, Yaling Liang

Main category: cs.CV

TL;DR: LetsTalk is a diffusion transformer framework with multimodal guidance and memory bank mechanism for high-quality, efficient long-duration talking video synthesis, addressing visual degradation and temporal inconsistency issues.

DetailsMotivation: Long-duration talking video synthesis faces challenges in video quality, portrait/temporal consistency, and computational efficiency. Issues like visual degradation, identity inconsistency, temporal incoherence, and error accumulation worsen with longer videos, affecting realism and reliability.

Method: Uses diffusion transformer framework with multimodal guidance and noise-regularized memory bank to maintain contextual continuity. Employs deep compression autoencoder and spatiotemporal-aware transformer with linear attention for multimodal fusion. Combines Symbiotic Fusion for portrait features and Direct Fusion for audio.

Result: Establishes new state-of-the-art in generation quality, producing temporally coherent and realistic talking videos with enhanced diversity and liveliness. Maintains remarkable efficiency with 8x fewer parameters than previous approaches.

Conclusion: LetsTalk effectively addresses long-duration talking video synthesis challenges through its memory bank mechanism and optimized fusion schemes, achieving superior visual realism, precise speech-driven motion, and computational efficiency.

Abstract: Long-duration talking video synthesis faces enduring challenges in achieving high video quality, portrait and temporal consistency, and computational efficiency. As video length increases, issues such as visual degradation, identity inconsistency, temporal incoherence, and error accumulation become increasingly problematic, severely affecting the realism and reliability of the results. To address these challenges, we present LetsTalk, a diffusion transformer framework equipped with multimodal guidance and a novel memory bank mechanism, explicitly maintaining contextual continuity and enabling robust, high-quality, and efficient generation of long-duration talking videos. In particular, LetsTalk introduces a noise-regularized memory bank to alleviate error accumulation and sampling artifacts during extended video generation. To further improve efficiency and spatiotemporal consistency, LetsTalk employs a deep compression autoencoder and a spatiotemporal-aware transformer with linear attention for effective multimodal fusion. We systematically analyze three fusion schemes and show that combining deep (Symbiotic Fusion) for portrait features and shallow (Direct Fusion) for audio achieves superior visual realism and precise speech-driven motion, while preserving diversity of movements. Extensive experiments demonstrate that LetsTalk establishes new state-of-the-art in generation quality, producing temporally coherent and realistic talking videos with enhanced diversity and liveliness, and maintains remarkable efficiency with 8x fewer parameters than previous approaches.

[395] Free$^2$Guide: Training-Free Text-to-Video Alignment using Image LVLM

Jaemin Kim, Bryan Sangwoo Kim, Jong Chul Ye

Main category: cs.CV

TL;DR: Free²Guide is a gradient-free, training-free framework that uses path integral control to align generated videos with text prompts by leveraging non-differentiable reward functions from Large Vision-Language Models.

DetailsMotivation: Existing RL-based approaches for text-to-video alignment require differentiable reward functions trained specifically for videos, limiting their scalability and applicability.

Method: Uses path integral control to approximate guidance for diffusion models with non-differentiable reward functions, employs stitching between video frames to enable image-trained LVLMs to assess video alignment, and supports flexible ensembling of multiple reward models.

Result: Free²Guide significantly improves text-to-video alignment using image-trained LVLMs, enhancing overall video quality without significant computational overhead.

Conclusion: The framework enables effective integration of powerful black-box LVLMs as reward models for video generation alignment, overcoming limitations of previous differentiable reward approaches.

Abstract: Diffusion models have achieved impressive results in generative tasks for text-to-video (T2V) synthesis. However, achieving accurate text alignment in T2V generation remains challenging due to the complex temporal dependencies across frames. Existing reinforcement learning (RL)-based approaches to enhance text alignment often require differentiable reward functions trained for videos, hindering their scalability and applicability. In this paper, we propose \textbf{Free$^2$Guide}, a novel gradient-free and training-free framework for aligning generated videos with text prompts. Specifically, leveraging principles from path integral control, Free$^2$Guide approximates guidance for diffusion models using non-differentiable reward functions, thereby enabling the integration of powerful black-box Large Vision-Language Models (LVLMs) as reward models. To enable image-trained LVLMs to assess text-to-video alignment, we leverage \textit{stitching} between video frames and use system prompts to capture sequential attributions. Our framework supports the flexible ensembling of multiple reward models to synergistically enhance alignment without significant computational overhead. Experimental results confirm that Free$^2$Guide using image-trained LVLMs significantly improves text-to-video alignment, thereby enhancing the overall video quality. Our results and code are available at https://kjm981995.github.io/free2guide/

[396] SoPo: Text-to-Motion Generation Using Semi-Online Preference Optimization

Xiaofeng Tan, Hongsong Wang, Xin Geng, Pan Zhou

Main category: cs.CV

TL;DR: SoPo is a semi-online preference optimization method that combines online and offline DPO to address limitations in text-to-motion generation, achieving state-of-the-art performance.

DetailsMotivation: Text-to-motion generation faces challenges in producing consistent, realistic motions. The paper aims to fine-tune models to consistently favor high-quality, human-preferred motions, addressing a critical but unexplored problem.

Method: Semi-online Preference Optimization (SoPo) uses semi-online data pairs consisting of unpreferred motion from online distribution and preferred motion from offline datasets. This leverages both online and offline DPO to compensate for each other’s limitations.

Result: SoPo outperforms other preference alignment methods with MM-Dist of 3.25% on MLD model and 2.91% on MDM model. The MLD model fine-tuned by SoPo surpasses state-of-the-art models in R-precision and MM Dist metrics.

Conclusion: SoPo effectively addresses limitations in both online and offline DPO through semi-online data pairing, demonstrating superior performance in text-to-motion generation preference alignment.

Abstract: Text-to-motion generation is essential for advancing the creative industry but often presents challenges in producing consistent, realistic motions. To address this, we focus on fine-tuning text-to-motion models to consistently favor high-quality, human-preferred motions, a critical yet largely unexplored problem. In this work, we theoretically investigate the DPO under both online and offline settings, and reveal their respective limitation: overfitting in offline DPO, and biased sampling in online DPO. Building on our theoretical insights, we introduce Semi-online Preference Optimization (SoPo), a DPO-based method for training text-to-motion models using “semi-online” data pair, consisting of unpreferred motion from online distribution and preferred motion in offline datasets. This method leverages both online and offline DPO, allowing each to compensate for the other’s limitations. Extensive experiments demonstrate that SoPo outperforms other preference alignment methods, with an MM-Dist of 3.25% (vs e.g. 0.76% of MoDiPO) on the MLD model, 2.91% (vs e.g. 0.66% of MoDiPO) on MDM model, respectively. Additionally, the MLD model fine-tuned by our SoPo surpasses the SoTA model in terms of R-precision and MM Dist. Visualization results also show the efficacy of our SoPo in preference alignment. Project page: https://xiaofeng-tan.github.io/projects/SoPo/ .

[397] FairGen: Enhancing Fairness in Text-to-Image Diffusion Models via Self-Discovering Latent Directions

Yilei Jiang, Weihong Li, Yiyuan Zhang, Minghong Cai, Xiangyu Yue

Main category: cs.CV

TL;DR: FairGen is a plug-and-play method that debiases Diffusion Models by learning attribute latent directions through self-discovery, eliminating the need for reference datasets or retraining.

DetailsMotivation: Diffusion Models reflect inherent biases from training data, which can perpetuate distorted worldviews and hinder opportunities for minority groups. Existing debiasing methods require expensive reference datasets or additional classifiers.

Method: FairGen uses attribute adapters that learn attribute latent directions via noise composition through self-discovery, and a distribution indicator that multiplies with adapters to guide generation towards prescribed distributions.

Result: Extensive experiments show FairGen outperforms previous state-of-the-art methods by a large margin in debiasing gender, racial, and intersectional biases.

Conclusion: FairGen provides an effective, lightweight solution for debiasing multiple attributes simultaneously in Diffusion Models without requiring retraining or reference datasets.

Abstract: While Diffusion Models (DM) exhibit remarkable performance across various image generative tasks, they nonetheless reflect the inherent bias presented in the training set. As DMs are now widely used in real-world applications, these biases could perpetuate a distorted worldview and hinder opportunities for minority groups. Existing methods on debiasing DMs usually requires model retraining with a human-crafted reference dataset or additional classifiers, which suffer from two major limitations: (1) collecting reference datasets causes expensive annotation cost; (2) the debiasing performance is heavily constrained by the quality of the reference dataset or the additional classifier. To address the above limitations, we propose FairGen, a plug-and-play method that learns attribute latent directions in a self-discovering manner, thus eliminating the reliance on such reference dataset. Specifically, FairGen consists of two parts: a set of attribute adapters and a distribution indicator. Each adapter in the set aims to learn an attribute latent direction, and is optimized via noise composition through a self-discovering process. Then, the distribution indicator is multiplied by the set of adapters to guide the generation process towards the prescribed distribution. Our method enables debiasing multiple attributes in DMs simultaneously, while remaining lightweight and easily integrable with other DMs, eliminating the need for retraining. Extensive experiments on debiasing gender, racial, and their intersectional biases show that our method outperforms previous SOTA by a large margin.

[398] NanoHTNet: Nano Human Topology Network for Efficient 3D Human Pose Estimation

Jialun Cai, Mengyuan Liu, Hong Liu, Shuheng Zhou, Wenhao Li

Main category: cs.CV

TL;DR: Proposes NanoHTNet, a tiny 3D human pose estimation network for edge devices, using hierarchical mixers to capture explicit spatio-temporal priors and PoseCLR pre-training to extract implicit human topology representations, achieving superior efficiency.

DetailsMotivation: Enable efficient 3D human pose estimation on resource-constrained edge devices by effectively utilizing structural priors in human skeletal inputs.

Method: NanoHTNet uses spatial Hierarchical Mixer for human physical topology and temporal Hierarchical Mixer with DCT and filtering for movement patterns. Includes ETST for spatio-temporal interaction. PoseCLR pre-training uses contrastive learning with 2D pose alignment.

Result: Outperforms state-of-the-art methods in efficiency, making it ideal for deployment on edge devices like Jetson Nano.

Conclusion: The combination of NanoHTNet’s efficient architecture and PoseCLR pre-training effectively captures both explicit and implicit human structural priors, enabling high-performance 3D HPE on edge devices.

Abstract: The widespread application of 3D human pose estimation (HPE) is limited by resource-constrained edge devices, requiring more efficient models. A key approach to enhancing efficiency involves designing networks based on the structural characteristics of input data. However, effectively utilizing the structural priors in human skeletal inputs remains challenging. To address this, we leverage both explicit and implicit spatio-temporal priors of the human body through innovative model design and a pre-training proxy task. First, we propose a Nano Human Topology Network (NanoHTNet), a tiny 3D HPE network with stacked Hierarchical Mixers to capture explicit features. Specifically, the spatial Hierarchical Mixer efficiently learns the human physical topology across multiple semantic levels, while the temporal Hierarchical Mixer with discrete cosine transform and low-pass filtering captures local instantaneous movements and global action coherence. Moreover, Efficient Temporal-Spatial Tokenization (ETST) is introduced to enhance spatio-temporal interaction and reduce computational complexity significantly. Second, PoseCLR is proposed as a general pre-training method based on contrastive learning for 3D HPE, aimed at extracting implicit representations of human topology. By aligning 2D poses from diverse viewpoints in the proxy task, PoseCLR aids 3D HPE encoders like NanoHTNet in more effectively capturing the high-dimensional features of the human body, leading to further performance improvements. Extensive experiments verify that NanoHTNet with PoseCLR outperforms other state-of-the-art methods in efficiency, making it ideal for deployment on edge devices like the Jetson Nano. Code and models are available at https://github.com/vefalun/NanoHTNet.

[399] DynVFX: Augmenting Real Videos with Dynamic Content

Danah Yatim, Rafail Fridman, Omer Bar-Tal, Tali Dekel

Main category: cs.CV

TL;DR: A zero-shot, training-free method for augmenting real-world videos with dynamic content using text instructions, achieving seamless integration of new objects/scene effects while handling camera motion, occlusions, and scene interactions.

DetailsMotivation: To enable users to easily add new dynamic content to existing videos through simple text instructions, creating realistic augmented videos without complex manual editing or training requirements.

Method: Uses pre-trained text-to-video diffusion transformer and vision-language model with novel inference-based feature manipulation in attention mechanism for accurate localization and seamless integration of new content.

Result: Successfully demonstrates diverse video edits on real-world videos with various objects and scenarios, handling both camera and object motion while preserving original scene integrity.

Conclusion: The proposed framework provides an effective, automated solution for video augmentation that requires only simple user instructions and produces realistic, cohesive output videos.

Abstract: We present a method for augmenting real-world videos with newly generated dynamic content. Given an input video and a simple user-provided text instruction describing the desired content, our method synthesizes dynamic objects or complex scene effects that naturally interact with the existing scene over time. The position, appearance, and motion of the new content are seamlessly integrated into the original footage while accounting for camera motion, occlusions, and interactions with other dynamic objects in the scene, resulting in a cohesive and realistic output video. We achieve this via a zero-shot, training-free framework that harnesses a pre-trained text-to-video diffusion transformer to synthesize the new content and a pre-trained vision-language model to envision the augmented scene in detail. Specifically, we introduce a novel inference-based method that manipulates features within the attention mechanism, enabling accurate localization and seamless integration of the new content while preserving the integrity of the original scene. Our method is fully automated, requiring only a simple user instruction. We demonstrate its effectiveness on a wide range of edits applied to real-world videos, encompassing diverse objects and scenarios involving both camera and object motion.

[400] Dual Caption Preference Optimization for Diffusion Models

Amir Saeidi, Yiran Luo, Agneet Chatterjee, Shamanthak Hegde, Bimsara Pathiraja, Yezhou Yang, Chitta Baral

Main category: cs.CV

TL;DR: DCPO is a framework that uses dual captions for preference pairs to enhance training signals in text-to-image diffusion models, improving image quality and prompt relevance.

DetailsMotivation: Existing preference datasets often have captions that don't clearly distinguish between preferred and less-preferred images, weakening supervision during training.

Method: Dual Caption Preference Optimization (DCPO) assigns two distinct captions to each preference pair using three strategies: captioning, perturbation, and hybrid methods. Also created Pick-Double Caption dataset.

Result: DCPO significantly outperforms Stable Diffusion 2.1, SFT_Chosen, Diffusion-DPO, and MaPO across multiple metrics including Pickscore, HPSv2.1, GenEval, CLIPscore, and ImageReward.

Conclusion: Using dual captions for preference pairs effectively reinforces learning signals and improves model performance in text-to-image generation.

Abstract: Recent advancements in human preference optimization, originally developed for Large Language Models (LLMs), have shown significant potential in improving text-to-image diffusion models. These methods aim to learn the distribution of preferred samples while distinguishing them from less preferred ones. However, within the existing preference datasets, the original caption often does not clearly favor the preferred image over the alternative, which weakens the supervision signal available during training. To address this issue, we introduce Dual Caption Preference Optimization (DCPO), a data augmentation and optimization framework that reinforces the learning signal by assigning two distinct captions to each preference pair. This encourages the model to better differentiate between preferred and less-preferred outcomes during training. We also construct Pick-Double Caption, a modified version of Pick-a-Pic v2 with separate captions for each image, and propose three different strategies for generating distinct captions: captioning, perturbation, and hybrid methods. Our experiments show that DCPO significantly improves image quality and relevance to prompts, outperforming Stable Diffusion (SD) 2.1, SFT_Chosen, Diffusion-DPO, and MaPO across multiple metrics, including Pickscore, HPSv2.1, GenEval, CLIPscore, and ImageReward, fine-tuned on SD 2.1 as the backbone.

[401] Indoor Heat Estimation from a Single Visible-Light Panorama

Guanzhou Ji, Sriram Narayanan, Azadeh Sawyer, Srinivasa Narasimhan

Main category: cs.CV

TL;DR: A novel image-based rendering technique that jointly estimates indoor lighting and thermal conditions from paired indoor-outdoor HDR panoramas, enabling integrated light and heat estimation for virtual home staging.

DetailsMotivation: To advance traditional virtual home staging by providing photorealistic and physically informed visualization through integrated light and heat estimation from panoramic imagery.

Method: Uses indoor panorama for 3D floor layout estimation and outdoor panorama as environment map to infer illumination and material properties. Models light-heat relationship assuming Lambertian surfaces and outdoor light as heat source, performing transient heat simulations to generate temperature distributions.

Result: Generated indoor temperature distributions through heat simulations, with validation against real-world thermal images captured using infrared cameras.

Conclusion: The approach successfully enables joint estimation of lighting and thermal conditions, supporting physically informed visualization for virtual home staging applications.

Abstract: This paper introduces a novel image-based rendering technique for jointly estimating indoor lighting and thermal conditions from paired indoor-outdoor high dynamic range (HDR) panoramas. Our method uses the indoor panorama to estimate the 3D floor layout, while the corresponding outdoor panorama serves as an environment map to infer spatially-varying illumination and material properties. Assuming indoor surfaces are Lambertian and that all heat originates from outdoor visible light, we model the relationship between light transport and heat transfer, and perform transient heat simulations to generate indoor temperature distributions. The simulated heat maps are validated against real-world thermal images captured with an infrared camera. This approach supports photorealistic and physically informed visualization, enabling integrated light and heat estimation to advance traditional virtual home staging.

[402] PartSDF: Part-Based Implicit Neural Representation for Composite 3D Shape Parametrization and Optimization

Nicolas Talabot, Olivier Clerc, Arda Cinar Demirtas, Alexis Goujon, Hieu Le, Doruk Oner, Pascal Fua

Main category: cs.CV

TL;DR: PartSDF is a supervised implicit representation framework that models composite shapes with independent, controllable parts while maintaining shape consistency, outperforming existing methods in reconstruction and generation tasks.

DetailsMotivation: Engineering workflows require structured, part-based representations as objects are designed as assemblies of distinct components, but existing methods either model shapes holistically or decompose them without predefined part structures, limiting real-world applicability.

Method: PartSDF uses a supervised implicit representation framework with a simple but innovative architecture that explicitly models composite shapes with independent, controllable parts while maintaining shape consistency.

Result: PartSDF outperforms both supervised and unsupervised baselines in reconstruction and generation tasks, and serves as an effective structured shape prior for engineering applications.

Conclusion: The framework enables precise control over individual components while preserving overall coherence, making it suitable for real-world engineering design tasks.

Abstract: Accurate 3D shape representation is essential in engineering applications such as design, optimization, and simulation. In practice, engineering workflows require structured, part-based representations, as objects are inherently designed as assemblies of distinct components. However, most existing methods either model shapes holistically or decompose them without predefined part structures, limiting their applicability in real-world design tasks. We propose PartSDF, a supervised implicit representation framework that explicitly models composite shapes with independent, controllable parts while maintaining shape consistency. Thanks to its simple but innovative architecture, PartSDF outperforms both supervised and unsupervised baselines in reconstruction and generation tasks. We further demonstrate its effectiveness as a structured shape prior for engineering applications, enabling precise control over individual components while preserving overall coherence. Code available at https://github.com/cvlab-epfl/PartSDF.

[403] ELIP: Enhanced Visual-Language Foundation Models for Image Retrieval

Guanqi Zhan, Yuanpei Liu, Kai Han, Weidi Xie, Andrew Zisserman

Main category: cs.CV

TL;DR: ELIP enhances text-to-image retrieval by using text queries to predict visual prompts that condition ViT image encoding, improving performance of CLIP/SigLIP/BLIP-2 models with limited computing resources.

DetailsMotivation: To improve text-to-image retrieval performance and enable large-scale pre-trained vision-language models to be used for text-to-image re-ranking.

Method: Enhanced Language-Image Pre-training (ELIP) uses text queries via MLP mapping to predict visual prompts that condition ViT image encoding. Includes global hard sample mining and dataset curation for efficient training.

Result: ELIP significantly boosts CLIP/SigLIP/BLIP-2 text-to-image retrieval performance, outperforms BLIP-2 on several benchmarks, and provides easy adaptation to out-of-distribution datasets.

Conclusion: ELIP framework effectively enhances vision-language models for text-to-image retrieval, demonstrating strong performance improvements and generalization capabilities with limited computing resources.

Abstract: The objective in this paper is to improve the performance of text-to-image retrieval. To this end, we introduce a new framework that can boost the performance of large-scale pre-trained vision-language models, so that they can be used for text-to-image re-ranking. The approach, Enhanced Language-Image Pre-training (ELIP), uses the text query, via a simple MLP mapping network, to predict a set of visual prompts to condition the ViT image encoding. ELIP can easily be applied to the commonly used CLIP, SigLIP and BLIP-2 networks. To train the architecture with limited computing resources, we develop a ‘student friendly’ best practice, involving global hard sample mining, and curation of a large-scale dataset. On the evaluation side, we set up two new out-of-distribution (OOD) benchmarks, Occluded COCO and ImageNet-R, to assess the zero-shot generalisation of the models to different domains. The results demonstrate that ELIP significantly boosts CLIP/SigLIP/SigLIP-2 text-to-image retrieval performance and outperforms BLIP-2 on several benchmarks, as well as providing an easy means to adapt to OOD datasets.

[404] Rethinking Multimodal Learning from the Perspective of Mitigating Classification Ability Disproportion

QingYuan Jiang, Longfei Huang, Yang Yang

Main category: cs.CV

TL;DR: Proposes a novel multimodal learning approach using boosting principles to dynamically balance classification ability between strong and weak modalities, addressing modality imbalance by optimizing classification and residual errors.

DetailsMotivation: Existing multimodal learning approaches overlook the inherent disproportion in model classification ability as the primary cause of modality imbalance, leading to suboptimal performance.

Method: Uses sustained boosting algorithm with simultaneous optimization of classification and residual errors, plus adaptive classifier assignment strategy to dynamically improve weak modality performance.

Result: Empirical experiments on widely used datasets show superiority over various state-of-the-art multimodal learning baselines.

Conclusion: The proposed method effectively balances classification ability between strong and weak modalities, mitigating the modality imbalance issue in multimodal learning.

Abstract: Multimodal learning (MML) is significantly constrained by modality imbalance, leading to suboptimal performance in practice. While existing approaches primarily focus on balancing the learning of different modalities to address this issue, they fundamentally overlook the inherent disproportion in model classification ability, which serves as the primary cause of this phenomenon. In this paper, we propose a novel multimodal learning approach to dynamically balance the classification ability of weak and strong modalities by incorporating the principle of boosting. Concretely, we first propose a sustained boosting algorithm in multimodal learning by simultaneously optimizing the classification and residual errors. Subsequently, we introduce an adaptive classifier assignment strategy to dynamically facilitate the classification performance of the weak modality. Furthermore, we theoretically analyze the convergence property of the cross-modal gap function, ensuring the effectiveness of the proposed boosting scheme. To this end, the classification ability of strong and weak modalities is expected to be balanced, thereby mitigating the imbalance issue. Empirical experiments on widely used datasets reveal the superiority of our method through comparison with various state-of-the-art (SOTA) multimodal learning baselines. The source code is available at https://github.com/njustkmg/NeurIPS25-AUG.

[405] Cutting-edge 3D reconstruction solutions for underwater coral reef images: A review and comparison

Jiageng Zhong, Ming Li, Armin Gruen, Konrad Schindler, Xuan Liao, Qinghua Guo

Main category: cs.CV

TL;DR: This paper provides a systematic review and evaluation of photogrammetry-based 3D reconstruction methods for underwater coral reef imaging, focusing on camera pose estimation and dense surface reconstruction stages.

DetailsMotivation: Coral reefs are fragile ecosystems that need preservation through accurate 3D modeling. While photogrammetry shows promise, there's a lack of systematic reviews evaluating cutting-edge methods specifically for underwater coral reef environments with their unique challenges.

Method: The authors systematically review classical and emerging methods for camera pose estimation and dense surface reconstruction, conducting comprehensive evaluations using both real-world and simulated datasets.

Result: The study provides reference recommendations for scientists and managers, identifying which methods work best for underwater coral reef 3D reconstruction given the challenges of underwater environments and complex coral structures.

Conclusion: This work bridges the gap between technical studies and practical applications, offering technical foundation and practical guidance for processing underwater coral reef images for 3D reconstruction, while discussing development potential and challenges of existing approaches.

Abstract: Corals serve as the foundational habitat-building organisms within reef ecosystems, constructing extensive structures that extend over vast distances. However, their inherent fragility and vulnerability to various threats render them susceptible to significant damage and destruction. The application of advanced 3D reconstruction technologies for high-quality modeling is crucial for preserving them. These technologies help scientists to accurately document and monitor the state of coral reefs, including their structure, species distribution and changes over time. Photogrammetry-based approaches stand out among existing solutions, especially with recent advancements in underwater videography, photogrammetric computer vision, and machine learning. Despite continuous progress in image-based 3D reconstruction techniques, there remains a lack of systematic reviews and comprehensive evaluations of cutting-edge solutions specifically applied to underwater coral reef images. The emerging advanced methods may have difficulty coping with underwater imaging environments, complex coral structures, and computational resource constraints. They need to be reviewed and evaluated to bridge the gap between many cutting-edge technical studies and practical applications. This paper focuses on the two critical stages of these approaches: camera pose estimation and dense surface reconstruction. We systematically review and summarize classical and emerging methods, conducting comprehensive evaluations through real-world and simulated datasets. Based on our findings, we offer reference recommendations and discuss the development potential and challenges of existing approaches in depth. This work equips scientists and managers with a technical foundation and practical guidance for processing underwater coral reef images for 3D reconstruction….

[406] NFIG: Multi-Scale Autoregressive Image Generation via Frequency Ordering

Zhihao Huang, Xi Qiu, Yukuo Ma, Yifu Zhou, Junjie Chen, Hongyuan Zhang, Chi Zhang, Xuelong Li

Main category: cs.CV

TL;DR: NFIG introduces a frequency-aware autoregressive image generation framework that decomposes generation into multiple stages guided by spectral hierarchy, improving quality while reducing inference cost.

DetailsMotivation: Standard autoregressive models generate pixels in fixed spatial order, failing to leverage the natural hierarchical structure of image information in the spectral domain where low-frequency components capture global structure efficiently.

Method: NFIG decomposes image generation into frequency-guided stages: first generating low-frequency components with fewer tokens to establish global structure, then progressively adding higher-frequency details.

Result: On ImageNet-256 benchmark, NFIG achieves superior performance with FID: 2.81 and a 1.25x speedup compared to VAR-d20 baseline.

Conclusion: The frequency-aware paradigm aligns generation with natural image structure, offering both quality improvements and computational efficiency gains in autoregressive image generation.

Abstract: Autoregressive models have achieved significant success in image generation. However, unlike the inherent hierarchical structure of image information in the spectral domain, standard autoregressive methods typically generate pixels sequentially in a fixed spatial order. To better leverage this spectral hierarchy, we introduce NextFrequency Image Generation (NFIG). NFIG is a novel framework that decomposes the image generation process into multiple frequency-guided stages. NFIG aligns the generation process with the natural image structure. It does this by first generating low-frequency components, which efficiently capture global structure with significantly fewer tokens, and then progressively adding higher-frequency details. This frequency-aware paradigm offers substantial advantages: it not only improves the quality of generated images but crucially reduces inference cost by efficiently establishing global structure early on. Extensive experiments on the ImageNet-256 benchmark validate NFIG’s effectiveness, demonstrating superior performance (FID: 2.81) and a notable 1.25x speedup compared to the strong baseline VAR-d20. The source code is available at https://github.com/Pride-Huang/NFIG.

[407] DitHub: A Modular Framework for Incremental Open-Vocabulary Object Detection

Chiara Cappellino, Gianluca Mancusi, Matteo Mosconi, Angelo Porrello, Simone Calderara, Rita Cucchiara

Main category: cs.CV

TL;DR: DitHub is a modular framework for open-vocabulary object detection that manages adaptation modules like version control branches, enabling efficient composition and achieving SOTA performance on ODinW benchmarks.

DetailsMotivation: To address the need for adapting open-vocabulary detectors to rare classes and specialized domains, moving beyond monolithic adaptation strategies to embrace modular deep learning.

Method: Introduces DitHub framework inspired by Version Control Systems, managing expert modules as branches that can be fetched and merged as needed, enabling compositional exploration of adaptation modules.

Result: Achieves state-of-the-art performance on ODinW-13 benchmark and newly introduced ODinW-O benchmark for assessing class reappearance.

Conclusion: Modular approach enables efficient adaptation and composition of expert modules, marking the first such study in Object Detection and demonstrating superior performance on challenging benchmarks.

Abstract: Open-Vocabulary object detectors can generalize to an unrestricted set of categories through simple textual prompting. However, adapting these models to rare classes or reinforcing their abilities on multiple specialized domains remains essential. While recent methods rely on monolithic adaptation strategies with a single set of weights, we embrace modular deep learning. We introduce DitHub, a framework designed to build and maintain a library of efficient adaptation modules. Inspired by Version Control Systems, DitHub manages expert modules as branches that can be fetched and merged as needed. This modular approach allows us to conduct an in-depth exploration of the compositional properties of adaptation modules, marking the first such study in Object Detection. Our method achieves state-of-the-art performance on the ODinW-13 benchmark and ODinW-O, a newly introduced benchmark designed to assess class reappearance. For more details, visit our project page: https://aimagelab.github.io/DitHub/

[408] Adaptive Label Correction for Robust Medical Image Segmentation with Noisy Labels

Chengxuan Qian, Kai Han, Jianxia Ding, Chongwen Lyu, Zhenlong Yuan, Jun Chen, Zhe Liu

Main category: cs.CV

TL;DR: Proposes a Mean Teacher-based Adaptive Label Correction framework for robust medical image segmentation with noisy labels, using adaptive label refinement and uncertainty-based sample selection to handle label noise effectively.

DetailsMotivation: Deep learning in medical imaging requires large labeled datasets, but obtaining high-quality labels is challenging. Noisy labels are easier to acquire but degrade model performance when used directly in training.

Method: Uses Mean Teacher architecture with adaptive label refinement mechanism that weights differences across disturbance versions, sample-level uncertainty-based label selection to prioritize confident samples, and consistency learning between student and teacher networks.

Result: Extensive experiments on two public datasets show significant improvements in segmentation performance, achieving competitive results compared to state-of-the-art methods.

Conclusion: The ALC framework effectively processes noisy labels, adapts to challenging scenarios, and fully exploits Mean Teacher structure strengths for robust medical image segmentation.

Abstract: Deep learning has shown remarkable success in medical image analysis, but its reliance on large volumes of high-quality labeled data limits its applicability. While noisy labeled data are easier to obtain, directly incorporating them into training can degrade model performance. To address this challenge, we propose a Mean Teacher-based Adaptive Label Correction (ALC) self-ensemble framework for robust medical image segmentation with noisy labels. The framework leverages the Mean Teacher architecture to ensure consistent learning under noise perturbations. It includes an adaptive label refinement mechanism that dynamically captures and weights differences across multiple disturbance versions to enhance the quality of noisy labels. Additionally, a sample-level uncertainty-based label selection algorithm is introduced to prioritize high-confidence samples for network updates, mitigating the impact of noisy annotations. Consistency learning is integrated to align the predictions of the student and teacher networks, further enhancing model robustness. Extensive experiments on two public datasets demonstrate the effectiveness of the proposed framework, showing significant improvements in segmentation performance. By fully exploiting the strengths of the Mean Teacher structure, the ALC framework effectively processes noisy labels, adapts to challenging scenarios, and achieves competitive results compared to state-of-the-art methods.

[409] Jasmine: Harnessing Diffusion Prior for Self-supervised Depth Estimation

Jiyuan Wang, Chunyu Lin, Cheng Guan, Lang Nie, Jing He, Haodong Li, Kang Liao, Yao Zhao

Main category: cs.CV

TL;DR: Jasmine is the first Stable Diffusion-based self-supervised framework for monocular depth estimation that leverages SD’s visual priors to improve prediction sharpness and generalization without requiring high-precision supervision.

DetailsMotivation: Previous SD-based methods require supervision for dense prediction, while self-supervised methods suffer from blurry predictions and artifacts that compromise SD's latent priors. The goal is to create a self-supervised approach that preserves SD's detail priors.

Method: Proposes a hybrid image reconstruction surrogate task that preserves SD’s detail priors by reconstructing images themselves. Introduces Scale-Shift GRU to bridge the distribution gap between SD’s scale-shift invariant estimation and self-supervised scale-invariant depth estimation.

Result: Achieves state-of-the-art performance on KITTI benchmark and demonstrates superior zero-shot generalization across multiple datasets.

Conclusion: Jasmine successfully leverages SD’s visual priors in a self-supervised framework for monocular depth estimation, overcoming limitations of both supervised SD methods and traditional self-supervised approaches.

Abstract: In this paper, we propose Jasmine, the first Stable Diffusion (SD)-based self-supervised framework for monocular depth estimation, which effectively harnesses SD’s visual priors to enhance the sharpness and generalization of unsupervised prediction. Previous SD-based methods are all supervised since adapting diffusion models for dense prediction requires high-precision supervision. In contrast, self-supervised reprojection suffers from inherent challenges (e.g., occlusions, texture-less regions, illumination variance), and the predictions exhibit blurs and artifacts that severely compromise SD’s latent priors. To resolve this, we construct a novel surrogate task of hybrid image reconstruction. Without any additional supervision, it preserves the detail priors of SD models by reconstructing the images themselves while preventing depth estimation from degradation. Furthermore, to address the inherent misalignment between SD’s scale and shift invariant estimation and self-supervised scale-invariant depth estimation, we build the Scale-Shift GRU. It not only bridges this distribution gap but also isolates the fine-grained texture of SD output against the interference of reprojection loss. Extensive experiments demonstrate that Jasmine achieves SoTA performance on the KITTI benchmark and exhibits superior zero-shot generalization across multiple datasets.

[410] Leveraging Vision-Language Models for Open-Vocabulary Instance Segmentation and Tracking

Bastian Pätzold, Jan Nogga, Sven Behnke

Main category: cs.CV

TL;DR: Integrates vision-language models with open-vocabulary detection and video segmentation to combine VLM descriptive power with reliable grounding and real-time processing.

DetailsMotivation: VLMs excel in visual understanding but lack reliable grounding capabilities and actionable inference rates. Need to leverage their strengths while mitigating drawbacks.

Method: Uses VLM-generated structured descriptions to identify objects, collects attributes, informs open-vocabulary detector for bounding boxes, then passes to video segmentation model for masks and tracking. Processes image streams in real-time with minimal overhead.

Result: Evaluation across datasets and robotics platforms demonstrates broad applicability and ability to extract task-specific attributes from non-standard objects in dynamic environments.

Conclusion: Successfully combines descriptive power of VLMs with grounding capability of OVD and pixel-level understanding/speed of video segmentation for real-time processing in dynamic environments.

Abstract: Vision-language models (VLMs) excel in visual understanding but often lack reliable grounding capabilities and actionable inference rates. Integrating them with open-vocabulary object detection (OVD), instance segmentation, and tracking leverages their strengths while mitigating these drawbacks. We utilize VLM-generated structured descriptions to identify visible object instances, collect application-relevant attributes, and inform an open-vocabulary detector to extract corresponding bounding boxes that are passed to a video segmentation model providing segmentation masks and tracking. Once initialized, this model directly extracts segmentation masks, processing image streams in real time with minimal computational overhead. Tracks can be updated online as needed by generating new structured descriptions and detections. This combines the descriptive power of VLMs with the grounding capability of OVD and the pixel-level understanding and speed of video segmentation. Our evaluation across datasets and robotics platforms demonstrates the broad applicability of this approach, showcasing its ability to extract task-specific attributes from non-standard objects in dynamic environments. Code, data, videos, and benchmarks are available at https://vlm-gist.github.io

[411] Unseen from Seen: Rewriting Observation-Instruction Using Foundation Models for Augmenting Vision-Language Navigation

Ziming Wei, Bingqian Lin, Yunshuang Nie, Jiaqi Chen, Shikui Ma, Hang Xu, Xiaodan Liang

Main category: cs.CV

TL;DR: RAM is a rewriting-driven augmentation method for Vision-Language Navigation that creates new observation-instruction pairs by rewriting human-annotated data, improving generalization without simulators or manual data cleaning.

DetailsMotivation: Data scarcity in VLN hinders agent generalization to unseen environments. Existing methods rely on limited simulator data or noisy web-collected data requiring manual cleaning.

Method: Uses object-enriched observation rewriting (VLMs+LLMs for scene descriptions, T2IMs for image synthesis) and observation-contrast instruction rewriting (LLMs reasoning differences between observations). Includes mixing-then-focusing training with random cropping.

Result: Superior performance on discrete (R2R, REVERIE, R4R) and continuous (R2R-CE) environments, showing impressive generalization ability.

Conclusion: RAM provides an effective simulator-free and labor-saving paradigm for VLN data augmentation that significantly improves generalization through rewriting-based data creation.

Abstract: Data scarcity is a long-standing challenge in the Vision-Language Navigation (VLN) field, which extremely hinders the generalization of agents to unseen environments. Previous works primarily rely on additional simulator data or web-collected images/videos to improve the generalization. However, the simulator environments still face limited diversity, and the web-collected data often requires extensive labor to remove the noise. In this paper, we propose a Rewriting-driven AugMentation (RAM) paradigm for VLN, which directly creates the unseen observation-instruction pairs via rewriting human-annotated training data. Benefiting from our rewriting mechanism, new observation-instruction pairs can be obtained in both simulator-free and labor-saving manners to promote generalization. Specifically, we first introduce Object-Enriched Observation Rewriting, where we combine Vision-Language Models (VLMs) and Large Language Models (LLMs) to derive rewritten object-enriched scene descriptions, enabling observation synthesis with diverse objects and spatial layouts via Text-to-Image Generation Models (T2IMs). Then, we propose Observation-Contrast Instruction Rewriting, which generates observation-aligned rewritten instructions by requiring LLMs to reason the difference between original and new observations. We further develop a mixing-then-focusing training strategy with a random observation cropping scheme, effectively enhancing data distribution diversity while suppressing augmentation data noise during training. Experiments on both the discrete environments (R2R, REVERIE, and R4R datasets) and continuous environments (R2R-CE dataset) show the superior performance and impressive generalization ability of our method. Code is available at https://github.com/SaDil13/VLN-RAM.

[412] Morpheus: Benchmarking Physical Reasoning of Video Generative Models with Real Physical Experiments

Chenyu Zhang, Daniil Cherniavskii, Antonios Tragoudaras, Antonios Vozikis, Thijmen Nijdam, Derck W. E. Prinzhorn, Mark Bodracska, Nicu Sebe, Andrii Zadaianchuk, Efstratios Gavves

Main category: cs.CV

TL;DR: Morpheus is a benchmark for evaluating video generation models on physical reasoning using 80 real-world videos and physics-informed metrics based on conservation laws.

DetailsMotivation: To assess whether current image and video generation models possess world modeling capabilities and adhere to physical conservation laws, which is crucial for applications in robotics, autonomous driving, and scientific simulation.

Method: Created Morpheus benchmark with 80 real-world videos capturing physical phenomena, using physics-informed metrics evaluated with respect to conservation laws through physics-informed neural networks and vision-language foundation models.

Result: Current video generation models struggle to encode physical principles despite generating aesthetically pleasing videos, even with advanced prompting and video conditioning.

Conclusion: Video generation models currently lack proper physical reasoning capabilities and cannot be reliably treated as world models for applications requiring physical plausibility.

Abstract: Recent advances in image and video generation raise hopes that these models possess world modeling capabilities, the ability to generate realistic, physically plausible videos. This could revolutionize applications in robotics, autonomous driving, and scientific simulation. However, before treating these models as world models, we must ask: Do they adhere to physical conservation laws? To answer this, we introduce Morpheus, a benchmark for evaluating video generation models on physical reasoning. It features 80 real-world videos capturing physical phenomena, guided by conservation laws. Since artificial generations lack ground truth, we assess physical plausibility using physics-informed metrics evaluated with respect to infallible conservation laws known per physical setting, leveraging advances in physics-informed neural networks and vision-language foundation models. Our findings reveal that even with advanced prompting and video conditioning, current models struggle to encode physical principles despite generating aesthetically pleasing videos. All data, leaderboard, and code are open-sourced at our project page.

[413] GaSLight: Gaussian Splats for Spatially-Varying Lighting in HDR

Christophe Bolduc, Yannick Hold-Geoffroy, Zhixin Shu, Jean-François Lalonde

Main category: cs.CV

TL;DR: GaSLight is a method that generates spatially-varying lighting from regular images using HDR Gaussian Splats as light source representation, enabling regular images to serve as light sources in 3D rendering.

DetailsMotivation: To enable regular images to serve as light sources in 3D rendering, which hasn't been possible before, and to create spatially-varying lighting from standard images.

Method: Two-stage process: 1) Enhance dynamic range of images using diffusion model priors, 2) Use Gaussian Splats to model 3D lighting for spatially variant effects. Also introduces a novel dataset for benchmarking.

Result: Achieves state-of-the-art results on HDR estimations and applications in illuminating virtual objects and scenes. Validated on both novel and existing datasets.

Conclusion: GaSLight successfully enables regular images to function as light sources in 3D rendering through HDR Gaussian Splats, with superior performance in HDR estimation and virtual scene illumination.

Abstract: We present GaSLight, a method that generates spatially-varying lighting from regular images. Our method proposes using HDR Gaussian Splats as light source representation, marking the first time regular images can serve as light sources in a 3D renderer. Our two-stage process first enhances the dynamic range of images plausibly and accurately by leveraging the priors embedded in diffusion models. Next, we employ Gaussian Splats to model 3D lighting, achieving spatially variant lighting. Our approach yields state-of-the-art results on HDR estimations and their applications in illuminating virtual objects and scenes. To facilitate the benchmarking of images as light sources, we introduce a novel dataset of calibrated and unsaturated HDR to evaluate images as light sources. We assess our method using a combination of this novel dataset and an existing dataset from the literature. Project page: https://lvsn.github.io/gaslight/

[414] Hierarchical Feature Learning for Medical Point Clouds via State Space Model

Guoqing Zhang, Jingyun Yang, Yang Li

Main category: cs.CV

TL;DR: An SSM-based hierarchical framework for medical point cloud understanding that uses coordinate-order and inside-out scanning strategies to process irregular points, achieving superior performance on anatomy classification, completion, and segmentation tasks.

DetailsMotivation: Limited research on medical point clouds despite their potential in disease diagnosis and treatment, and the promising capabilities of transformer and state space models in point cloud learning.

Method: Hierarchical feature learning with farthest point sampling, KNN queries for multi-scale information aggregation, coordinate-order and inside-out scanning for point serialization, and Point SSM blocks to capture local patterns and long-range dependencies.

Result: The method achieves superior performance across anatomy classification, completion, and segmentation tasks on the newly built MedPointS dataset.

Conclusion: The proposed SSM-based framework effectively handles medical point clouds and demonstrates strong performance on multiple medical tasks, with the dataset and code made publicly available.

Abstract: Deep learning-based point cloud modeling has been widely investigated as an indispensable component of general shape analysis. Recently, transformer and state space model (SSM) have shown promising capacities in point cloud learning. However, limited research has been conducted on medical point clouds, which have great potential in disease diagnosis and treatment. This paper presents an SSM-based hierarchical feature learning framework for medical point cloud understanding. Specifically, we down-sample input into multiple levels through the farthest point sampling. At each level, we perform a series of k-nearest neighbor (KNN) queries to aggregate multi-scale structural information. To assist SSM in processing point clouds, we introduce coordinate-order and inside-out scanning strategies for efficient serialization of irregular points. Point features are calculated progressively from short neighbor sequences and long point sequences through vanilla and group Point SSM blocks, to capture both local patterns and long-range dependencies. To evaluate the proposed method, we build a large-scale medical point cloud dataset named MedPointS for anatomy classification, completion, and segmentation. Extensive experiments conducted on MedPointS demonstrate that our method achieves superior performance across all tasks. The dataset is available at https://flemme-docs.readthedocs.io/en/latest/medpoints.html. Code is merged to a public medical imaging platform: https://github.com/wlsdzyzl/flemme.

[415] Generate, but Verify: Reducing Hallucination in Vision-Language Models with Retrospective Resampling

Tsung-Han Wu, Heekyung Lee, Jiaxin Ge, Joseph E. Gonzalez, Trevor Darrell, David M. Chan

Main category: cs.CV

TL;DR: REVERSE is a unified framework that integrates hallucination-aware training with on-the-fly self-verification to reduce visual hallucinations in Vision-Language Models, achieving state-of-the-art performance.

DetailsMotivation: Vision-Language Models suffer from visual hallucinations that generate descriptions of nonexistent objects, posing risks in safety-critical applications. Existing methods either rely on heuristics without correction mechanisms or are complicated multi-model systems that reject outputs rather than refine them.

Method: REVERSE combines hallucination-aware training using a new dataset with 1.3M semi-synthetic samples and a novel inference-time retrospective resampling technique that enables VLMs to detect hallucinations during generation and dynamically revise them.

Result: REVERSE achieves state-of-the-art hallucination reduction, outperforming the best existing methods by up to 12% on CHAIR-MSCOCO and 34% on HaloQuest benchmarks.

Conclusion: The REVERSE framework provides an effective unified approach for reducing visual hallucinations in VLMs through integrated training and self-verification, with publicly available dataset, model, and code.

Abstract: Vision-Language Models (VLMs) excel at visual understanding but often suffer from visual hallucinations, where they generate descriptions of nonexistent objects, actions, or concepts, posing significant risks in safety-critical applications. Existing hallucination mitigation methods typically follow one of two paradigms: generation adjustment, which modifies decoding behavior to align text with visual inputs, and post-hoc verification, where external models assess and correct outputs. While effective, generation adjustment methods often rely on heuristics and lack correction mechanisms, while post-hoc verification is complicated, typically requiring multiple models and tending to reject outputs rather than refine them. In this work, we introduce REVERSE, a unified framework that integrates hallucination-aware training with on-the-fly self-verification. By leveraging a new hallucination-verification dataset containing over 1.3M semi-synthetic samples, along with a novel inference-time retrospective resampling technique, our approach enables VLMs to both detect hallucinations during generation and dynamically revise those hallucinations. Our evaluations show that REVERSE achieves state-of-the-art hallucination reduction, outperforming the best existing methods by up to 12% on CHAIR-MSCOCO and 34% on HaloQuest. Our dataset, model, and code are available at: https://reverse-vlm.github.io.

[416] SSL4Eco: A Global Seasonal Dataset for Geospatial Foundation Models in Ecology

Elena Plekhanova, Damien Robert, Johannes Dollinger, Emilia Arens, Philipp Brun, Jan Dirk Wegner, Niklaus Zimmermann

Main category: cs.CV

TL;DR: The paper introduces SSL4Eco, a phenology-informed multi-date Sentinel-2 dataset for self-supervised learning, which improves biodiversity mapping by better capturing global vegetation seasonality and addressing dataset biases.

DetailsMotivation: Addressing the biodiversity and climate crises requires better global biodiversity mapping, but current remote sensing models are biased toward human activity areas and don't properly capture local phenological cycles.

Method: Proposed a phenology-informed sampling strategy and created SSL4Eco dataset, then trained an existing model with season-contrastive objective on multi-date Sentinel-2 imagery.

Result: The SSL4Eco-pretrained model achieved state-of-the-art performance on 7 out of 8 downstream tasks spanning classification and regression, consistently outperforming other datasets.

Conclusion: The straightforward phenology-informed sampling method significantly improves representation quality, highlighting the importance of dataset construction in ecological remote sensing.

Abstract: With the exacerbation of the biodiversity and climate crises, macroecological pursuits such as global biodiversity mapping become more urgent. Remote sensing offers a wealth of Earth observation data for ecological studies, but the scarcity of labeled datasets remains a major challenge. Recently, self-supervised learning has enabled learning representations from unlabeled data, triggering the development of pretrained geospatial models with generalizable features. However, these models are often trained on datasets biased toward areas of high human activity, leaving entire ecological regions underrepresented. Additionally, while some datasets attempt to address seasonality through multi-date imagery, they typically follow calendar seasons rather than local phenological cycles. To better capture vegetation seasonality at a global scale, we propose a simple phenology-informed sampling strategy and introduce corresponding SSL4Eco, a multi-date Sentinel-2 dataset, on which we train an existing model with a season-contrastive objective. We compare representations learned from SSL4Eco against other datasets on diverse ecological downstream tasks and demonstrate that our straightforward sampling method consistently improves representation quality, highlighting the importance of dataset construction. The model pretrained on SSL4Eco reaches state of the art performance on 7 out of 8 downstream tasks spanning (multi-label) classification and regression. We release our code, data, and model weights to support macroecological and computer vision research at https://github.com/PlekhanovaElena/ssl4eco.

[417] Fine-Grained Classification: Connecting Metadata via Cross-Contrastive Pre-Training

Sumit Mamtani, Yash Thesia

Main category: cs.CV

TL;DR: A unified multimodal framework combining image, text, and metadata via cross-contrastive pre-training for fine-grained visual classification, achieving 84.44% top-1 accuracy on NABirds.

DetailsMotivation: Fine-grained visual classification is challenging because appearance alone often fails to distinguish highly similar subordinate categories within a supercategory.

Method: Proposes a unified framework that integrates image, text, and metadata via cross-contrastive pre-training, aligning three modality encoders in a shared embedding space, then fine-tuning image and metadata encoders for classification.

Result: Achieves 84.44% top-1 accuracy on NABirds, improving over baseline by 7.83% and outperforming strong multimodal methods.

Conclusion: The proposed multimodal framework effectively addresses fine-grained classification challenges by leveraging complementary information from images, text, and metadata through cross-contrastive learning.

Abstract: Fine-grained visual classification aims to recognize objects belonging to many subordinate categories of a supercategory, where appearance alone often fails to distinguish highly similar classes. We propose a unified framework that integrates image, text, and metadata via cross-contrastive pre-training. We first align the three modality encoders in a shared embedding space and then fine-tune the image and metadata encoders for classification. On NABirds, our approach improves over the baseline by 7.83% and achieves 84.44% top-1 accuracy, outperforming strong multimodal methods.

[418] Learning Cocoercive Conservative Denoisers via Helmholtz Decomposition for Poisson Inverse Problems

Deliang Wei, Peng Chen, Haobo Xu, Jiale Yao, Fang Li, Tieyong Zeng

Main category: cs.CV

TL;DR: The paper proposes a cocoercive conservative (CoCo) denoiser for Plug-and-Play methods in Poisson inverse problems, addressing limitations of existing approaches that require strong convexity/smoothness and non-expansive denoisers.

DetailsMotivation: Existing PnP methods have restrictive assumptions (strong convexity/smoothness of fidelity term and non-expansive denoisers) that are violated in Poisson inverse problems, and non-expansiveness can limit denoising performance.

Method: Introduces CoCo denoiser using generalized Helmholtz decomposition with Hamiltonian regularization for conservativeness and spectral regularization for cocoerciveness. Proves it’s a proximal operator of a weakly convex function.

Result: The CoCo denoiser enables PnP methods to converge globally to stationary points of restoration models with implicit weakly convex priors. Experimental results show superior performance over related methods.

Conclusion: The proposed CoCo denoiser framework overcomes limitations of traditional PnP methods in Poisson inverse problems, providing improved denoising performance with theoretical convergence guarantees.

Abstract: Plug-and-play (PnP) methods with deep denoisers have shown impressive results in imaging problems. They typically require strong convexity or smoothness of the fidelity term and a (residual) non-expansive denoiser for convergence. These assumptions, however, are violated in Poisson inverse problems, and non-expansiveness can hinder denoising performance. To address these challenges, we propose a cocoercive conservative (CoCo) denoiser, which may be (residual) expansive, leading to improved denoising. By leveraging the generalized Helmholtz decomposition, we introduce a novel training strategy that combines Hamiltonian regularization to promote conservativeness and spectral regularization to ensure cocoerciveness. We prove that CoCo denoiser is a proximal operator of a weakly convex function, enabling a restoration model with an implicit weakly convex prior. The global convergence of PnP methods to a stationary point of this restoration model is established. Extensive experimental results demonstrate that our approach outperforms closely related methods in both visual quality and quantitative metrics.

[419] Video-SafetyBench: A Benchmark for Safety Evaluation of Video LVLMs

Xuannan Liu, Zekun Li, Zheqi He, Peipei Li, Shuhan Xia, Xing Cui, Huaibo Huang, Xi Yang, Ran He

Main category: cs.CV

TL;DR: Video-SafetyBench is the first comprehensive benchmark for evaluating LVLM safety under video-text attacks, revealing significant vulnerabilities to video-induced attacks with 67.2% average success rate.

DetailsMotivation: Existing multimodal safety evaluations focus on static images, ignoring the temporal dynamics of video that may induce distinct safety risks in Large Vision-Language Models.

Method: Created benchmark with 2,264 video-text pairs across 48 unsafe categories using a controllable pipeline that decomposes video semantics into subject images and motion text. Proposed RJScore metric for evaluating uncertain outputs with confidence-based judgment.

Result: Benign-query video composition achieved 67.2% average attack success rate, revealing consistent vulnerabilities to video-induced attacks across tested models.

Conclusion: Video-SafetyBench addresses critical gaps in multimodal safety evaluation and will catalyze future research into video-based safety evaluation and defense strategies.

Abstract: The increasing deployment of Large Vision-Language Models (LVLMs) raises safety concerns under potential malicious inputs. However, existing multimodal safety evaluations primarily focus on model vulnerabilities exposed by static image inputs, ignoring the temporal dynamics of video that may induce distinct safety risks. To bridge this gap, we introduce Video-SafetyBench, the first comprehensive benchmark designed to evaluate the safety of LVLMs under video-text attacks. It comprises 2,264 video-text pairs spanning 48 fine-grained unsafe categories, each pairing a synthesized video with either a harmful query, which contains explicit malice, or a benign query, which appears harmless but triggers harmful behavior when interpreted alongside the video. To generate semantically accurate videos for safety evaluation, we design a controllable pipeline that decomposes video semantics into subject images (what is shown) and motion text (how it moves), which jointly guide the synthesis of query-relevant videos. To effectively evaluate uncertain or borderline harmful outputs, we propose RJScore, a novel LLM-based metric that incorporates the confidence of judge models and human-aligned decision threshold calibration. Extensive experiments show that benign-query video composition achieves average attack success rates of 67.2%, revealing consistent vulnerabilities to video-induced attacks. We believe Video-SafetyBench will catalyze future research into video-based safety evaluation and defense strategies.

[420] Is Artificial Intelligence Generated Image Detection a Solved Problem?

Ziqiang Li, Jiazhen Yan, Ziwen He, Kai Zeng, Weiwei Jiang, Lizhi Xiong, Zhangjie Fu

Main category: cs.CV

TL;DR: AIGIBench is a comprehensive benchmark that evaluates AI-generated image detectors, revealing significant performance drops in real-world scenarios despite high reported accuracy in controlled settings.

DetailsMotivation: The rapid advancement of generative models has raised concerns about misinformation and deepfakes, but existing AIGI detectors' effectiveness in real-world scenarios remains questionable.

Method: AIGIBench simulates real-world challenges through four core tasks: multi-source generalization, robustness to image degradation, sensitivity to data augmentation, and impact of test-time pre-processing, using 23 diverse fake image subsets and real-world samples.

Result: Experiments on 11 advanced detectors show significant performance drops on real-world data, limited benefits from common augmentations, and nuanced effects of pre-processing.

Conclusion: There is a need for more robust detection strategies, and AIGIBench provides a unified evaluation framework to guide future research toward dependable and generalizable AIGI detection.

Abstract: The rapid advancement of generative models, such as GANs and Diffusion models, has enabled the creation of highly realistic synthetic images, raising serious concerns about misinformation, deepfakes, and copyright infringement. Although numerous Artificial Intelligence Generated Image (AIGI) detectors have been proposed, often reporting high accuracy, their effectiveness in real-world scenarios remains questionable. To bridge this gap, we introduce AIGIBench, a comprehensive benchmark designed to rigorously evaluate the robustness and generalization capabilities of state-of-the-art AIGI detectors. AIGIBench simulates real-world challenges through four core tasks: multi-source generalization, robustness to image degradation, sensitivity to data augmentation, and impact of test-time pre-processing. It includes 23 diverse fake image subsets that span both advanced and widely adopted image generation techniques, along with real-world samples collected from social media and AI art platforms. Extensive experiments on 11 advanced detectors demonstrate that, despite their high reported accuracy in controlled settings, these detectors suffer significant performance drops on real-world data, limited benefits from common augmentations, and nuanced effects of pre-processing, highlighting the need for more robust detection strategies. By providing a unified and realistic evaluation framework, AIGIBench offers valuable insights to guide future research toward dependable and generalizable AIGI detection.Data and code are publicly available at: https://github.com/HorizonTEL/AIGIBench.

[421] UniCTokens: Boosting Personalized Understanding and Generation via Unified Concept Tokens

Ruichuan An, Sihan Yang, Renrui Zhang, Zijun Shen, Ming Lu, Gaole Dai, Hao Liang, Ziyu Guo, Shilin Yan, Yulin Luo, Bocheng Zou, Chaoqun Yang, Wentao Zhang

Main category: cs.CV

TL;DR: UniCTokens is a unified framework that integrates personalized concept understanding and generation using unified concept tokens and progressive training, achieving state-of-the-art in attribute-reasoning generation.

DetailsMotivation: Existing methods treat concept understanding and generation as separate tasks with isolated tokens, limiting their ability to handle complex prompts like generating "concept wearing its hat" without additional descriptions.

Method: Proposes unified concept tokens and a three-stage progressive training strategy: understanding warm-up, bootstrapping generation from understanding, and deepening understanding from generation.

Result: Shows competitive performance in concept understanding and generation, and achieves state-of-the-art results in personalized attribute-reasoning generation on the proposed UnifyBench benchmark.

Conclusion: Enhanced understanding improves generation, and generation provides valuable insights for understanding, demonstrating mutual benefits between both tasks in a unified vision language model.

Abstract: Personalized models have demonstrated remarkable success in understanding and generating concepts provided by users. However, existing methods use separate concept tokens for understanding and generation, treating these tasks in isolation. This may result in limitations for generating images with complex prompts. For example, given the concept $\langle bo\rangle$, generating “$\langle bo\rangle$ wearing its hat” without additional textual descriptions of its hat. We call this kind of generation \textit{\textbf{personalized attribute-reasoning generation}}. To address the limitation, we present UniCTokens, a novel framework that effectively integrates personalized information into a unified vision language model (VLM) for understanding and generation. UniCTokens trains a set of unified concept tokens to leverage complementary semantics, boosting two personalized tasks. Moreover, we propose a progressive training strategy with three stages: understanding warm-up, bootstrapping generation from understanding, and deepening understanding from generation to enhance mutual benefits between both tasks. To quantitatively evaluate the unified VLM personalization, we present UnifyBench, the first benchmark for assessing concept understanding, concept generation, and attribute-reasoning generation. Experimental results on UnifyBench indicate that UniCTokens shows competitive performance compared to leading methods in concept understanding, concept generation, and achieving state-of-the-art results in personalized attribute-reasoning generation. Our research demonstrates that enhanced understanding improves generation, and the generation process can yield valuable insights into understanding. Our code and dataset will be released at: \href{https://github.com/arctanxarc/UniCTokens}{https://github.com/arctanxarc/UniCTokens}.

[422] GMatch: A Lightweight, Geometry-Constrained Keypoint Matcher for Zero-Shot 6DoF Pose Estimation in Robotic Grasp Tasks

Ming Yang, Haoran Li

Main category: cs.CV

TL;DR: GMatch is a lightweight, geometry-constrained keypoint matcher for 6DoF object pose estimation that runs efficiently on embedded CPU-only platforms, achieving competitive performance with SOTA methods while being computationally efficient.

DetailsMotivation: To address the computational demands of recent learning-based 6DoF pose estimation methods that hinder deployment on resource-constrained mobile robotic platforms, by revisiting classical keypoint matching with geometric constraints.

Method: Proposes GMatch, a geometry-constrained keypoint matcher that uses geometric constraints to establish globally consistent correspondences between keypoint descriptors, enabling easy 6DoF pose solving.

Result: GMatch beats existing keypoint matchers on HOPE and YCB-Video datasets across three descriptors, approaches SOTA zero-shot method performance on texture-rich objects, and achieves high task success rates when deployed on a LoCoBot mobile manipulator.

Conclusion: GMatch offers a practical solution for resource-limited robotic systems through its lightweight and white-box nature, presenting a promising direction for robust yet efficient pose estimation, though currently limited by descriptor quality.

Abstract: 6DoF object pose estimation is fundamental to robotic grasp tasks. While recent learning-based methods achieve high accuracy, their computational demands hinder deployment on resource-constrained mobile platforms. In this work, we revisit the classical keypoint matching paradigm and propose GMatch, a lightweight, geometry-constrained keypoint matcher that can run efficiently on embedded CPU-only platforms. GMatch works with keypoint descriptors and it uses a set of geometric constraints to establishes inherent ambiguities between features extracted by descriptors, thus giving a globally consistent correspondences from which 6DoF pose can be easily solved. We benchmark GMatch on the HOPE and YCB-Video datasets, where our method beats existing keypoint matchers (both feature-based and geometry-based) among three commonly used descriptors and approaches the SOTA zero-shot method on texture-rich objects with much more humble devices. The method is further deployed on a LoCoBot mobile manipulator, enabling a one-shot grasp pipeline that demonstrates high task success rates in real-world experiments. In a word, by its lightweight and white-box nature, GMatch offers a practical solution for resource-limited robotic systems, and although currently bottlenecked by descriptor quality, the framework presents a promising direction towards robust yet efficient pose estimation. Code will be released soon under Mozilla Public License.

[423] Hyperspectral Anomaly Detection Fused Unified Nonconvex Tensor Ring Factors Regularization

Wenjin Qin, Hailin Wang, Hao Shu, Feng Zhang, Jianjun Wang, Xiangyong Cao, Xi-Le Zhao, Gemine Vivone

Main category: cs.CV

TL;DR: Proposes HAD-EUNTRFR, a novel hyperspectral anomaly detection method using enhanced unified nonconvex tensor ring factors regularization to better capture global correlations and local smoothness in background components.

DetailsMotivation: Existing tensor decomposition methods fail to fully leverage both global correlations and local smoothness of background components in hyperspectral images, leading to suboptimal detection performance.

Method: Decomposes HSIs into background and anomaly components, uses TR decomposition for spatial-spectral correlations, introduces unified nonconvex regularizer via TSVD for low-rankness and sparsity of 3D gradient TR factors, and adds generalized nonconvex regularization for anomaly sparsity. Solved using ADMM optimization.

Result: Experimental results on benchmark datasets show the method outperforms existing state-of-the-art approaches in detection accuracy.

Conclusion: HAD-EUNTRFR effectively addresses limitations of existing methods by better capturing background characteristics and achieves superior anomaly detection performance.

Abstract: In recent years, tensor decomposition-based approaches for hyperspectral anomaly detection (HAD) have gained significant attention in the field of remote sensing. However, existing methods often fail to fully leverage both the global correlations and local smoothness of the background components in hyperspectral images (HSIs), which exist in both the spectral and spatial domains. This limitation results in suboptimal detection performance. To mitigate this critical issue, we put forward a novel HAD method named HAD-EUNTRFR, which incorporates an enhanced unified nonconvex tensor ring (TR) factors regularization. In the HAD-EUNTRFR framework, the raw HSIs are first decomposed into background and anomaly components. The TR decomposition is then employed to capture the spatial-spectral correlations within the background component. Additionally, we introduce a unified and efficient nonconvex regularizer, induced by tensor singular value decomposition (TSVD), to simultaneously encode the low-rankness and sparsity of the 3-D gradient TR factors into a unique concise form. The above characterization scheme enables the interpretable gradient TR factors to inherit the low-rankness and smoothness of the original background. To further enhance anomaly detection, we design a generalized nonconvex regularization term to exploit the group sparsity of the anomaly component. To solve the resulting doubly nonconvex model, we develop a highly efficient optimization algorithm based on the alternating direction method of multipliers (ADMM) framework. Experimental results on several benchmark datasets demonstrate that our proposed method outperforms existing state-of-the-art (SOTA) approaches in terms of detection accuracy.

[424] Styl3R: Instant 3D Stylized Reconstruction for Arbitrary Scenes and Styles

Peng Wang, Xiang Liu, Peidong Liu

Main category: cs.CV

TL;DR: A novel approach for instant 3D scene stylization using unposed sparse-view images and arbitrary style images, achieving results in under a second while maintaining multi-view consistency.

DetailsMotivation: Current 3D stylization methods require computationally intensive test-time optimization and dense posed input images, making them slow and impractical for real-time applications.

Method: Uses a branched architecture that separates structure modeling and appearance shading, preventing style distortion of 3D structure. Employs identity loss for pre-training through novel view synthesis, allowing reconstruction capabilities to be retained during stylization fine-tuning.

Result: Produces high-quality stylized 3D content with superior blend of style and scene appearance, outperforming existing methods in multi-view consistency and efficiency.

Conclusion: The approach enables direct 3D stylization in under a second using sparse unposed images, achieving better multi-view consistency and efficiency than current state-of-the-art methods.

Abstract: Stylizing 3D scenes instantly while maintaining multi-view consistency and faithfully resembling a style image remains a significant challenge. Current state-of-the-art 3D stylization methods typically involve computationally intensive test-time optimization to transfer artistic features into a pretrained 3D representation, often requiring dense posed input images. In contrast, leveraging recent advances in feed-forward reconstruction models, we demonstrate a novel approach to achieve direct 3D stylization in less than a second using unposed sparse-view scene images and an arbitrary style image. To address the inherent decoupling between reconstruction and stylization, we introduce a branched architecture that separates structure modeling and appearance shading, effectively preventing stylistic transfer from distorting the underlying 3D scene structure. Furthermore, we adapt an identity loss to facilitate pre-training our stylization model through the novel view synthesis task. This strategy also allows our model to retain its original reconstruction capabilities while being fine-tuned for stylization. Comprehensive evaluations, using both in-domain and out-of-domain datasets, demonstrate that our approach produces high-quality stylized 3D content that achieve a superior blend of style and scene appearance, while also outperforming existing methods in terms of multi-view consistency and efficiency.

[425] Hierarchical Material Recognition from Local Appearance

Matthew Beveridge, Shree K. Nayar

Main category: cs.CV

TL;DR: The paper introduces a taxonomy for hierarchical material recognition based on physical traits, presents a dataset with images and depth maps, and develops a graph attention network model that leverages taxonomic relationships to achieve state-of-the-art performance.

DetailsMotivation: To address hierarchical material recognition from local appearance by creating a taxonomy based on physical traits of materials and developing methods that can work in real-world conditions.

Method: Developed a hierarchical material recognition method using graph attention networks that leverages taxonomic proximity between classes, and utilized a diverse dataset with images and depth maps.

Result: The model achieves state-of-the-art performance, demonstrates generalization to adverse imaging conditions, benefits from novel views rendered using depth maps, and shows capacity for few-shot learning of new materials.

Conclusion: The proposed taxonomy, dataset, and graph attention network approach effectively enable hierarchical material recognition with strong performance, generalization capabilities, and few-shot learning potential.

Abstract: We introduce a taxonomy of materials for hierarchical recognition from local appearance. Our taxonomy is motivated by vision applications and is arranged according to the physical traits of materials. We contribute a diverse, in-the-wild dataset with images and depth maps of the taxonomy classes. Utilizing the taxonomy and dataset, we present a method for hierarchical material recognition based on graph attention networks. Our model leverages the taxonomic proximity between classes and achieves state-of-the-art performance. We demonstrate the model’s potential to generalize to adverse, real-world imaging conditions, and that novel views rendered using the depth maps can enhance this capability. Finally, we show the model’s capacity to rapidly learn new materials in a few-shot learning setting.

[426] Grounded Reinforcement Learning for Visual Reasoning

Gabriel Sarch, Snigdha Saha, Naitik Khandelwal, Ayush Jain, Michael J. Tarr, Aviral Kumar, Katerina Fragkiadaki

Main category: cs.CV

TL;DR: ViGoRL is a vision-language model trained with reinforcement learning to anchor reasoning steps to specific visual coordinates, improving visual reasoning performance across multiple benchmarks.

DetailsMotivation: Visual reasoning requires models to direct visual attention, interpret perceptual inputs, and ground abstract reasoning in spatial evidence, which adds complexity beyond traditional chain-of-thought reasoning.

Method: ViGoRL uses reinforcement learning to produce spatially grounded reasoning traces and guide visual attention to task-relevant regions. It features a novel multi-turn RL framework that enables dynamic zooming into predicted coordinates when fine-grained exploration is needed.

Result: ViGoRL consistently outperforms supervised fine-tuning and conventional RL baselines across diverse visual reasoning benchmarks, achieving 86.4% on V*Bench. It improves performance on localizing small GUI elements and visual search, and amplifies other visual behaviors like region exploration and visual verification.

Conclusion: Visually grounded reinforcement learning is a strong paradigm for imbuing models with general-purpose visual reasoning capabilities, with human evaluations confirming the spatial accuracy and helpfulness of the model’s visual references.

Abstract: While reinforcement learning (RL) over chains of thought has significantly advanced language models in tasks such as mathematics and coding, visual reasoning introduces added complexity by requiring models to direct visual attention, interpret perceptual inputs, and ground abstract reasoning in spatial evidence. We introduce ViGoRL (Visually Grounded Reinforcement Learning), a vision-language model trained with RL to explicitly anchor each reasoning step to specific visual coordinates. Inspired by human visual decision-making, ViGoRL learns to produce spatially grounded reasoning traces, guiding visual attention to task-relevant regions at each step. When fine-grained exploration is required, our novel multi-turn RL framework enables the model to dynamically zoom into predicted coordinates as reasoning unfolds. Across a diverse set of visual reasoning benchmarks–including SAT-2 and BLINK for spatial reasoning, Vbench for visual search, and ScreenSpot and VisualWebArena for web-based grounding–ViGoRL consistently outperforms both supervised fine-tuning and conventional RL baselines that lack explicit grounding mechanisms. Incorporating multi-turn RL with zoomed-in visual feedback significantly improves ViGoRL’s performance on localizing small GUI elements and visual search, achieving 86.4% on VBench. Additionally, we find that grounding amplifies other visual behaviors such as region exploration, grounded subgoal setting, and visual verification. Finally, human evaluations show that the model’s visual references are not only spatially accurate but also helpful for understanding model reasoning steps. Our results show that visually grounded RL is a strong paradigm for imbuing models with general-purpose visual reasoning.

[427] Infinity Parser: Layout Aware Reinforcement Learning for Scanned Document Parsing

Baode Wang, Biao Wu, Weizhen Li, Meng Fang, Zuming Huang, Jun Huang, Haozhe Wang, Yanjie Liang, Ling Chen, Wei Chu, Yuan Qi

Main category: cs.CV

TL;DR: layoutRL is an end-to-end RL framework for document parsing that trains layout-aware models using composite rewards, achieving SOTA performance on OCR, table/formula extraction, and reading order detection.

DetailsMotivation: Traditional multi-stage document parsing pipelines suffer from error propagation and limited adaptability to diverse layouts, creating a critical bottleneck in Document AI.

Method: End-to-end reinforcement learning framework with composite reward optimization (normalized edit distance, paragraph count accuracy, reading order preservation) using Infinity-Doc-55K dataset and vision-language-model-based parser.

Result: Infinity-Parser achieves new state-of-the-art performance on English and Chinese benchmarks for OCR, table/formula extraction, and reading order detection, outperforming specialist pipelines and general-purpose vision-language models.

Conclusion: The layoutRL framework enables robust document understanding with superior accuracy and structural fidelity, with code and dataset to be publicly released to advance the field.

Abstract: Automated parsing of scanned documents into richly structured, machine-readable formats remains a critical bottleneck in Document AI, as traditional multi-stage pipelines suffer from error propagation and limited adaptability to diverse layouts. We introduce layoutRL, an end-to-end reinforcement learning framework that trains models to be explicitly layout-aware by optimizing a composite reward of normalized edit distance, paragraph count accuracy, and reading order preservation. Leveraging our newly released dataset, Infinity-Doc-55K, which combines 55K high-fidelity synthetic scanned document parsing data with expert-filtered real-world documents, we instantiate layoutRL in a vision-language-model-based parser called Infinity-Parser. Evaluated on English and Chinese benchmarks for OCR, table and formula extraction, and reading order detection, Infinity-Parser achieves new state-of-the-art performance in both accuracy and structural fidelity, outpacing specialist pipelines and general-purpose vision-language models. We will publicly release our code and dataset to accelerate progress in robust document understanding.

[428] CReFT-CAD: Boosting Orthographic Projection Reasoning for CAD via Reinforcement Fine-Tuning

Ke Niu, Zhuofan Chen, Haiyang Yu, Yuwen Chen, Teng Fu, Mengyang Zhao, Bin Li, Xiangyang Xue

Main category: cs.CV

TL;DR: CReFT-CAD introduces a two-stage fine-tuning paradigm for orthographic projection reasoning in CAD workflows, using curriculum-driven reinforcement learning followed by supervised post-tuning, and releases the TriView2CAD benchmark.

DetailsMotivation: Standard deep-learning approaches for CAD workflows introduce imprecise dimensions and limit parametric editability, while supervised fine-tuning often leads to pattern memorization with poor out-of-distribution performance on complex reasoning tasks.

Method: Two-stage fine-tuning: 1) curriculum-driven reinforcement learning with difficulty-aware rewards to build reasoning ability, 2) supervised post-tuning to hone instruction following and semantic extraction. Also created TriView2CAD benchmark with 200,000 synthetic and 3,000 real-world orthographic projections.

Result: CReFT-CAD substantially improves reasoning accuracy and out-of-distribution generalizability in real-world scenarios compared to leading VLMs.

Conclusion: The approach offers valuable insights for advancing CAD reasoning research by addressing the limitations of existing methods and providing a comprehensive benchmark for orthographic projection reasoning.

Abstract: Computer-Aided Design (CAD) plays a pivotal role in industrial manufacturing. Orthographic projection reasoning underpins the entire CAD workflow, encompassing design, manufacturing, and simulation. However, prevailing deep-learning approaches employ standard 3D reconstruction pipelines as an alternative, which often introduce imprecise dimensions and limit the parametric editability required for CAD workflows. Recently, some researchers adopt vision-language models (VLMs), particularly supervised fine-tuning (SFT), to tackle CAD-related challenges. SFT shows promise but often devolves into pattern memorization, yielding poor out-of-distribution performance on complex reasoning tasks. To address these gaps, we introduce CReFT-CAD, a two-stage fine-tuning paradigm that first employs a curriculum-driven reinforcement learning stage with difficulty-aware rewards to build reasoning ability steadily, and then applies supervised post-tuning to hone instruction following and semantic extraction. Complementing this, we release TriView2CAD, the first large-scale, open-source benchmark for orthographic projection reasoning, comprising 200,000 synthetic and 3,000 real-world orthographic projections with precise dimension annotations and six interoperable data modalities. We benchmark leading VLMs on orthographic projection reasoning and demonstrate that CReFT-CAD substantially improves reasoning accuracy and out-of-distribution generalizability in real-world scenarios, offering valuable insights for advancing CAD reasoning research.

[429] VisuRiddles: Fine-grained Perception is a Primary Bottleneck for Multimodal Large Language Models in Abstract Visual Reasoning

Hao Yan, Handong Zheng, Hao Wang, Liang Yin, Xingchen Liu, Zhenbiao Cao, Xinxing Su, Zihao Chen, Jihao Wu, Minghui Liao, Chao Weng, Wei Chen, Yuliang Liu, Xiang Bai

Main category: cs.CV

TL;DR: The paper addresses abstract visual reasoning challenges in MLLMs by introducing VisuRiddles benchmark and Perceptual Riddle Synthesizer for data generation and fine-grained perceptual training.

DetailsMotivation: Abstract Visual Reasoning remains a critical challenge for MLLMs due to limitations in perceiving abstract graphics, creating a bottleneck in reasoning capabilities.

Method: Proposed VisuRiddles benchmark with tasks across five dimensions and two reasoning categories, and introduced Perceptual Riddle Synthesizer framework for automated riddle generation with fine-grained perceptual descriptions.

Result: Experimental results on VisuRiddles validate that fine-grained visual perception is the principal bottleneck and the synthesis framework significantly enhances MLLM performance on challenging AVR tasks.

Conclusion: The proposed approach effectively addresses abstract visual reasoning limitations in MLLMs through targeted benchmark development and automated data synthesis with perceptual supervision.

Abstract: Recent strides in multimodal large language models (MLLMs) have significantly advanced their performance in many reasoning tasks. However, Abstract Visual Reasoning (AVR) remains a critical challenge, primarily due to limitations in perceiving abstract graphics. To tackle this issue, we investigate the bottlenecks in current MLLMs and synthesize training data to improve their abstract visual perception. First, we propose VisuRiddles, a benchmark for AVR, featuring tasks meticulously constructed to assess models’ reasoning capacities across five core dimensions and two high-level reasoning categories. Second, we introduce the Perceptual Riddle Synthesizer (PRS), an automated framework for generating riddles with fine-grained perceptual descriptions. PRS not only generates valuable training data for abstract graphics but also provides fine-grained perceptual description, crucially allowing for supervision over intermediate reasoning stages and thereby improving both training efficacy and model interpretability. Our extensive experimental results on VisuRiddles empirically validate that fine-grained visual perception is the principal bottleneck and our synthesis framework markedly enhances the performance of contemporary MLLMs on these challenging tasks. Our code and dataset will be released at https://github.com/yh-hust/VisuRiddles

[430] Reasoning-Aligned Perception Decoupling for Scalable Multi-modal Reasoning

Yunhao Gou, Kai Chen, Zhili Liu, Lanqing Hong, Xin Jin, Zhenguo Li, James T. Kwok, Yu Zhang

Main category: cs.CV

TL;DR: RAPID decouples perception from reasoning in MLLMs, allowing easy replacement of reasoning components without costly retraining, using VPO reinforcement learning to align perceptual outputs with reasoning tasks.

DetailsMotivation: Multi-modal LLMs lag behind text-only models in reasoning due to outdated internal LLMs, and upgrading them requires expensive complete vision-language alignment retraining.

Method: Perception-Reasoning Decoupling modularizes MLLMs to convert multi-modal inputs to textual outputs for external LLM reasoners, plus Visual Perception Optimization (VPO) reinforcement learning to align perceptual outputs with reasoning tasks.

Result: RAPID achieves significant performance gains on multi-modal reasoning benchmarks and enables inference-time scaling with any state-of-the-art LLM reasoner without retraining.

Conclusion: RAPID provides an efficient approach to upgrade MLLM reasoning capabilities by decoupling perception and reasoning, enabling flexible integration with external LLMs and consistent performance improvement.

Abstract: Recent breakthroughs in reasoning language models have significantly advanced text-based reasoning. On the other hand, Multi-modal Large Language Models (MLLMs) still lag behind, hindered by their outdated internal LLMs. Upgrading these is often prohibitively expensive, as it requires complete vision-language alignment retraining which is costly. To address this issue, we introduce Perception-Reasoning Decoupling, which modularizes the MLLM’s reasoning component and makes it easily replaceable. This approach redefines the MLLM’s role to convert multi-modal inputs into detailed textual outputs that can be processed by any powerful, external, text-only LLM reasoners. To align the MLLM’s perceptual output with the final reasoning task, we propose a novel reinforcement learning algorithm called Visual Perception Optimization (VPO). VPO rewards the MLLM based on the correctness of answers generated by the external reasoner to produce faithful and query-relevant captions. Together, this decoupling pipeline and VPO form our Reasoning-Aligned PerceptIon Decoupling (RAPID) approach. Empirical results show that RAPID achieves significant performance gains on multi-modal reasoning benchmarks. Crucially, RAPID enables a novel inference-time scaling paradigm: Once trained with VPO, the MLLM can be paired with any state-of-the-art LLM reasoner for consistent performance improvement without retraining.

[431] ASCD: Attention-Steerable Contrastive Decoding for Reducing Hallucination in MLLM

Yujun Wang, Aniri, Jinhe Bi, Soeren Pirk, Yunpu Ma

Main category: cs.CV

TL;DR: ASCD is a novel attention-steering method that reduces hallucinations in multimodal LLMs by amplifying text-centric attention heads and dampening critical visual tokens during decoding, achieving significant improvements without additional training.

DetailsMotivation: Multimodal LLMs often hallucinate by over-relying on spurious visual cues, and while prior methods like VCD and ICD help, their mechanisms remain unclear. The authors aim to develop a more transparent and effective approach to mitigate hallucinations.

Method: ASCD directly steers attention scores during decoding by: (i) positive steering that amplifies automatically identified text-centric attention heads, and (ii) negative steering that dampens critical visual tokens identified on-the-fly. The method requires no additional training and has minimal runtime/memory overhead.

Result: Across five MLLM backbones and three decoding schemes, ASCD reduces hallucination on POPE, CHAIR, and MMHal-Bench by up to 38.2% while improving accuracy on standard VQA benchmarks including MMMU, MM-VET, ScienceQA, TextVQA, and GQA.

Conclusion: Attention steering provides a simple, model-agnostic, and principled approach to achieve safer and more faithful multimodal generation in MLLMs.

Abstract: Multimodal large language models (MLLMs) frequently hallucinate by over-committing to spurious visual cues. Prior remedies-Visual and Instruction Contrastive Decoding (VCD, ICD)-mitigate this issue, yet the mechanism remains opaque. We first empirically show that their improvements systematically coincide with redistributions of cross-modal attention. Building on this insight, we propose Attention-Steerable Contrastive Decoding (ASCD), which directly steers the attention scores during decoding. ASCD combines (i) positive steering, which amplifies automatically mined text-centric heads-stable within a model and robust across domains-with (ii) negative steering, which dampens on-the-fly identified critical visual tokens. The method incurs negligible runtime and memory overhead and requires no additional training. Across five MLLM backbones and three decoding schemes, ASCD reduces hallucination on POPE, CHAIR, and MMHal-Bench by up to 38.2 percent while improving accuracy on standard VQA benchmarks, including MMMU, MM-VET, ScienceQA, TextVQA, and GQA. These results position attention steering as a simple, model-agnostic, and principled route to safer, more faithful multimodal generation.

[432] Consistent Story Generation: Unlocking the Potential of Zigzag Sampling

Mingxiao Li, Mang Ning, Marie-Francine Moens

Main category: cs.CV

TL;DR: A training-free sampling strategy using zigzag sampling with asymmetric prompts and visual sharing to improve subject consistency in visual story generation without requiring fine-tuning.

DetailsMotivation: Text-to-image models struggle with maintaining subject consistency across multiple images for visual storytelling, and existing methods are either resource-intensive or yield limited success.

Method: Zigzag sampling mechanism that alternates between asymmetric prompting to retain subject characteristics, combined with a visual sharing module that transfers visual cues across generated images.

Result: Significantly outperforms previous approaches in generating coherent and consistent visual stories based on both quantitative metrics and qualitative evaluations.

Conclusion: The proposed training-free method effectively enhances subject consistency in visual story generation without requiring resource-intensive fine-tuning.

Abstract: Text-to-image generation models have made significant progress in producing high-quality images from textual descriptions, yet they continue to struggle with maintaining subject consistency across multiple images, a fundamental requirement for visual storytelling. Existing methods attempt to address this by either fine-tuning models on large-scale story visualization datasets, which is resource-intensive, or by using training-free techniques that share information across generations, which still yield limited success. In this paper, we introduce a novel training-free sampling strategy called Zigzag Sampling with Asymmetric Prompts and Visual Sharing to enhance subject consistency in visual story generation. Our approach proposes a zigzag sampling mechanism that alternates between asymmetric prompting to retain subject characteristics, while a visual sharing module transfers visual cues across generated images to %further enforce consistency. Experimental results, based on both quantitative metrics and qualitative evaluations, demonstrate that our method significantly outperforms previous approaches in generating coherent and consistent visual stories. The code is available at https://github.com/Mingxiao-Li/Asymmetry-Zigzag-StoryDiffusion.

[433] MotionGPT3: Human Motion as a Second Modality

Bingfan Zhu, Biao Jiang, Sunyi Wang, Shixiang Tang, Tao Chen, Linjie Luo, Youyi Zheng, Xin Chen

Main category: cs.CV

TL;DR: MotionGPT3 is a bimodal motion-language model that uses a dual-stream Transformer with shared attention to handle motion and language modalities separately while enabling controlled information flow, achieving faster convergence and state-of-the-art performance.

DetailsMotivation: Existing multimodal frameworks face complexity with growing modalities and tasks. Motion quantization introduces approximation errors, and unifying discrete text with continuous motion in single-stream backbones causes cross-modal interference.

Method: Encodes raw motion into continuous latent space using VAE to avoid quantization artifacts. Uses dual-stream Transformer with shared attention to preserve modality-specific routes while enabling controlled bidirectional information flow. Implements generate-then-align three-stage training schedule for stability.

Result: Achieves 2x faster convergence in training loss and up to 4x faster convergence in validation. Maintains state-of-the-art performance on standard motion understanding and generation benchmarks.

Conclusion: The proposed dual-stream architecture with continuous motion encoding effectively reduces cross-modal interference, stabilizes optimization, and accelerates convergence while maintaining high performance in both motion understanding and generation tasks.

Abstract: With the rapid progress of large language models (LLMs), multimodal frameworks that unify understanding and generation have become promising, yet they face increasing complexity as the number of modalities and tasks grows. We observe that motion quantization introduces approximation errors that cap motion quality, and that unifying discrete text and continuous motion within a single-stream backbone amplifies cross-modal interference. Motivated by recent multi-branch Transformer designs that separate signals from different modalities, we propose MotionGPT3, a bimodal motion-language model for both understanding and generation. MotionGPT3 encodes raw motion into a continuous latent space using a variational autoencoder (VAE), thereby avoiding quantization-induced artifacts, while leveraging the semantic prior of pretrained language models. A dual-stream Transformer with shared attention preserves modality-specific routes while enabling controlled, bidirectional information flow, which reduces interference, stabilizing optimization, and empirically accelerates convergence without degrading fidelity. For multimodal joint training, a generate-then-align three-stage schedule further improves stability and limits cross-task interference. Experiments show that MotionGPT3 achieves 2x faster convergence in training loss and up to 4x faster convergence in validation, while maintaining state-of-the-art performance on standard motion understanding and motion generation benchmarks.

[434] GeoCAD: Local Geometry-Controllable CAD Generation with Large Language Models

Zhanwei Zhang, Kaiyuan Liu, Junjie Liu, Wenxiao Wang, Binbin Lin, Liang Xie, Chen Shen, Deng Cai

Main category: cs.CV

TL;DR: GeoCAD is a local geometry-controllable CAD generation method that enables users to modify specific parts of CAD models while following geometric instructions like shapes of triangles or rectangles.

DetailsMotivation: Existing CAD generation methods lack the ability to follow textual instructions or focus on local parts, limiting design efficiency and user control over geometric specifications.

Method: Uses complementary captioning (vertex-based for simple parts, VLLM-based for complex parts) to annotate ~221k local parts, then trains LLMs to predict masked parts using geometric instructions and remaining model as input.

Result: Extensive experiments show GeoCAD achieves high generation quality, validity, and text-to-CAD consistency, enabling users to modify any local part while adhering to geometric instructions.

Conclusion: GeoCAD provides an effective solution for local geometry-controllable CAD generation, addressing limitations of existing methods and enhancing design efficiency through user-friendly geometric instruction following.

Abstract: Local geometry-controllable computer-aided design (CAD) generation aims to modify local parts of CAD models automatically, enhancing design efficiency. It also ensures that the shapes of newly generated local parts follow user-specific geometric instructions (e.g., an isosceles right triangle or a rectangle with one corner cut off). However, existing methods encounter challenges in achieving this goal. Specifically, they either lack the ability to follow textual instructions or are unable to focus on the local parts. To address this limitation, we introduce GeoCAD, a user-friendly and local geometry-controllable CAD generation method. Specifically, we first propose a complementary captioning strategy to generate geometric instructions for local parts. This strategy involves vertex-based and VLLM-based captioning for systematically annotating simple and complex parts, respectively. In this way, we caption $\sim$221k different local parts in total. In the training stage, given a CAD model, we randomly mask a local part. Then, using its geometric instruction and the remaining parts as input, we prompt large language models (LLMs) to predict the masked part. During inference, users can specify any local part for modification while adhering to a variety of predefined geometric instructions. Extensive experiments demonstrate the effectiveness of GeoCAD in generation quality, validity and text-to-CAD consistency. Code will be available at https://github.com/Zhanwei-Z/GeoCAD.

[435] WaveFormer: A Lightweight Transformer Model for sEMG-based Gesture Recognition

Yanlong Chen, Mattia Orlandi, Pierangelo Maria Rapa, Simone Benatti, Luca Benini, Yawei Li

Main category: cs.CV

TL;DR: WaveFormer is a lightweight transformer-based model for sEMG gesture recognition that achieves 95% accuracy with only 3.1M parameters and real-time inference capabilities.

DetailsMotivation: To address the challenge of classifying similar gestures with nearly identical muscle signals while overcoming the computational limitations of traditional deep learning models on resource-constrained embedded systems.

Method: Proposes WaveFormer architecture with novel learnable wavelet transform integrating time-domain and frequency-domain features, using WaveletConv module with multi-level wavelet decomposition and depthwise separable convolution for efficiency.

Result: Achieves 95% classification accuracy on EPN612 dataset, outperforming larger models. With INT8 quantization, achieves 6.75 ms inference latency on Intel CPU, enabling real-time deployment.

Conclusion: WaveFormer provides an efficient and compact solution for sEMG gesture recognition that balances high accuracy with low computational requirements, making it suitable for embedded systems.

Abstract: Human-machine interaction, particularly in prosthetic and robotic control, has seen progress with gesture recognition via surface electromyographic (sEMG) signals.However, classifying similar gestures that produce nearly identical muscle signals remains a challenge, often reducing classification accuracy. Traditional deep learning models for sEMG gesture recognition are large and computationally expensive, limiting their deployment on resource-constrained embedded systems. In this work, we propose WaveFormer, a lightweight transformer-based architecture tailored for sEMG gesture recognition. Our model integrates time-domain and frequency-domain features through a novel learnable wavelet transform, enhancing feature extraction. In particular, the WaveletConv module, a multi-level wavelet decomposition layer with depthwise separable convolution, ensures both efficiency and compactness. With just 3.1 million parameters, WaveFormer achieves 95% classification accuracy on the EPN612 dataset, outperforming larger models. Furthermore, when profiled on a laptop equipped with an Intel CPU, INT8 quantization achieves real-time deployment with a 6.75 ms inference latency.

[436] Auto-Connect: Connectivity-Preserving RigFormer with Direct Preference Optimization

Jingfeng Guo, Jian Liu, Jinnan Chen, Shiwei Mao, Changrong Hu, Puhua Jiang, Junlin Yu, Jing Xu, Qi Liu, Lixin Xu, Zhuo Chen, Chunchao Guo

Main category: cs.CV

TL;DR: Auto-Connect is a novel automatic rigging method that preserves skeletal connectivity through special tokens, uses topology-aware reward functions for optimization, and incorporates geodesic features for improved skinning quality.

DetailsMotivation: Previous rigging methods either predict bone positions as two joints or predict points before determining connectivity, which can lead to topological inaccuracies and poor skeletal structures.

Method: Uses connectivity-preserving tokenization with special tokens for endpoints and hierarchical layers, implements topology-aware reward function with Direct Preference Optimization, and incorporates implicit geodesic features for latent bone selection.

Result: The approach generates more anatomically plausible skeletal structures with superior deformation properties and significantly enhanced topological accuracy.

Conclusion: Auto-Connect’s combination of connectivity-preserving tokenization, reward-guided fine-tuning, and geodesic-aware bone selection enables consistent generation of high-quality skeletal rigs with improved skinning quality.

Abstract: We introduce Auto-Connect, a novel approach for automatic rigging that explicitly preserves skeletal connectivity through a connectivity-preserving tokenization scheme. Unlike previous methods that predict bone positions represented as two joints or first predict points before determining connectivity, our method employs special tokens to define endpoints for each joint’s children and for each hierarchical layer, effectively automating connectivity relationships. This approach significantly enhances topological accuracy by integrating connectivity information directly into the prediction framework. To further guarantee high-quality topology, we implement a topology-aware reward function that quantifies topological correctness, which is then utilized in a post-training phase through reward-guided Direct Preference Optimization. Additionally, we incorporate implicit geodesic features for latent top-k bone selection, which substantially improves skinning quality. By leveraging geodesic distance information within the model’s latent space, our approach intelligently determines the most influential bones for each vertex, effectively mitigating common skinning artifacts. This combination of connectivity-preserving tokenization, reward-guided fine-tuning, and geodesic-aware bone selection enables our model to consistently generate more anatomically plausible skeletal structures with superior deformation properties.

[437] VimoRAG: Video-based Retrieval-augmented 3D Motion Generation for Motion Language Models

Haidong Xu, Guangwei Xu, Zhedong Zheng, Xiatian Zhu, Wei Ji, Xiangtai Li, Ruijie Guo, Meishan Zhang, Min zhang, Hao Fei

Main category: cs.CV

TL;DR: VimoRAG is a video-based retrieval-augmented motion generation framework that enhances motion LLMs by retrieving relevant 2D human motion signals from large-scale video databases to address out-of-domain issues.

DetailsMotivation: Motion LLMs face severe out-of-domain/out-of-vocabulary issues due to limited annotated data, which hinders their ability to generate accurate 3D motions.

Method: Develops Gemini Motion Video Retriever mechanism for effective motion-centered video retrieval and Motion-centric Dual-alignment DPO Trainer to mitigate error propagation from suboptimal retrieval results.

Result: Experimental results show VimoRAG significantly boosts the performance of motion LLMs constrained to text-only input.

Conclusion: VimoRAG effectively addresses key bottlenecks in video-based motion RAG and enhances motion generation capabilities of LLMs through retrieval augmentation.

Abstract: This paper introduces VimoRAG, a novel video-based retrieval-augmented motion generation framework for motion large language models (LLMs). As motion LLMs face severe out-of-domain/out-of-vocabulary issues due to limited annotated data, VimoRAG leverages large-scale in-the-wild video databases to enhance 3D motion generation by retrieving relevant 2D human motion signals. While video-based motion RAG is nontrivial, we address two key bottlenecks: (1) developing an effective motion-centered video retrieval model that distinguishes human poses and actions, and (2) mitigating the issue of error propagation caused by suboptimal retrieval results. We design the Gemini Motion Video Retriever mechanism and the Motion-centric Dual-alignment DPO Trainer, enabling effective retrieval and generation processes. Experimental results show that VimoRAG significantly boosts the performance of motion LLMs constrained to text-only input. All the resources are available at https://walkermitty.github.io/VimoRAG/

[438] From Cradle to Cane: A Two-Pass Framework for High-Fidelity Lifespan Face Aging

Tao Liu, Dafeng Zhang, Gengchen Li, Shizhuo Liu, Yongqi Song, Senmao Li, Shiqi Yang, Boqian Li, Kai Wang, Yaxing Wang

Main category: cs.CV

TL;DR: Cradle2Cane is a two-pass face aging framework using diffusion models that addresses the Age-ID trade-off through adaptive noise injection for age accuracy and identity embeddings for preservation.

DetailsMotivation: Existing face aging methods struggle with balancing age accuracy and identity preservation, especially for large age gaps and extreme poses, creating the Age-ID trade-off problem.

Method: Two-pass framework: first pass uses adaptive noise injection (AdaNI) with age/gender prompts for age accuracy; second pass uses identity embeddings (SVR-ArcFace and Rotate-CLIP) for identity preservation while maintaining age features.

Result: Outperforms existing methods on CelebA-HQ dataset in both age accuracy and identity consistency, as evaluated by Face++ and Qwen-VL protocols.

Conclusion: The proposed Cradle2Cane framework successfully addresses the Age-ID trade-off in face aging through its two-pass approach, achieving superior performance in both age transformation accuracy and identity preservation.

Abstract: Face aging has become a crucial task in computer vision, with applications ranging from entertainment to healthcare. However, existing methods struggle with achieving a realistic and seamless transformation across the entire lifespan, especially when handling large age gaps or extreme head poses. The core challenge lies in balancing age accuracy and identity preservation–what we refer to as the Age-ID trade-off. Most prior methods either prioritize age transformation at the expense of identity consistency or vice versa. In this work, we address this issue by proposing a two-pass face aging framework, named Cradle2Cane, based on few-step text-to-image (T2I) diffusion models. The first pass focuses on solving age accuracy by introducing an adaptive noise injection (AdaNI) mechanism. This mechanism is guided by including prompt descriptions of age and gender for the given person as the textual condition. Also, by adjusting the noise level, we can control the strength of aging while allowing more flexibility in transforming the face. However, identity preservation is weakly ensured here to facilitate stronger age transformations. In the second pass, we enhance identity preservation while maintaining age-specific features by conditioning the model on two identity-aware embeddings (IDEmb): SVR-ArcFace and Rotate-CLIP. This pass allows for denoising the transformed image from the first pass, ensuring stronger identity preservation without compromising the aging accuracy. Both passes are jointly trained in an end-to-end way. Extensive experiments on the CelebA-HQ test dataset, evaluated through Face++ and Qwen-VL protocols, show that our Cradle2Cane outperforms existing face aging methods in age accuracy and identity consistency. Code is available at https://github.com/byliutao/Cradle2Cane.

[439] Geometry and Perception Guided Gaussians for Multiview-consistent 3D Generation from a Single Image

Pufan Li, Bi’an Du, Wei Hu

Main category: cs.CV

TL;DR: A novel method that integrates geometry and perception priors to generate detailed 3D objects from single images, achieving better multiview consistency and geometric detail than existing approaches.

DetailsMotivation: Existing methods for 3D object generation from single images suffer from poor multiview consistency and lack geometric detail, often relying on fine-tuning 2D diffusion models or direct 3D generation.

Method: Integrates geometry and perception priors to initialize Gaussian branches and guide optimization, uses stable Score Distillation Sampling for fine-grained prior distillation, and employs reprojection-based strategy for depth consistency.

Result: Outperforms existing methods on novel view synthesis and 3D reconstruction, demonstrating robust and consistent 3D object generation.

Conclusion: The proposed approach successfully addresses multiview consistency and geometric detail issues in 3D object generation from single images without requiring additional model training.

Abstract: Generating realistic 3D objects from single-view images requires natural appearance, 3D consistency, and the ability to capture multiple plausible interpretations of unseen regions. Existing approaches often rely on fine-tuning pretrained 2D diffusion models or directly generating 3D information through fast network inference or 3D Gaussian Splatting, but their results generally suffer from poor multiview consistency and lack geometric detail. To tackle these issues, we present a novel method that seamlessly integrates geometry and perception information without requiring additional model training to reconstruct detailed 3D objects from a single image. Specifically, we incorporate geometry and perception priors to initialize the Gaussian branches and guide their parameter optimization. The geometry prior captures the rough 3D shapes, while the perception prior utilizes the 2D pretrained diffusion model to enhance multiview information. Subsequently, we introduce a stable Score Distillation Sampling for fine-grained prior distillation to ensure effective knowledge transfer. The model is further enhanced by a reprojection-based strategy that enforces depth consistency. Experimental results show that we outperform existing methods on novel view synthesis and 3D reconstruction, demonstrating robust and consistent 3D object generation.

[440] G$^{2}$D: Boosting Multimodal Learning with Gradient-Guided Distillation

Mohammed Rakib, Arunkumar Bagavathi

Main category: cs.CV

TL;DR: G$^{2}$D is a knowledge distillation framework that addresses modality imbalance in multimodal learning through gradient-guided optimization and sequential modality prioritization to enhance weak modalities.

DetailsMotivation: Conventional multimodal models suffer from modality imbalance where dominant modalities overshadow weaker ones, leading to suboptimal feature representation and underutilization of weak modalities.

Method: Gradient-Guided Distillation (G$^{2}$D) with custom loss function fusing unimodal and multimodal objectives, plus dynamic sequential modality prioritization (SMP) to ensure each modality leads learning.

Result: G$^{2}$D amplifies weak modality significance during training and outperforms state-of-the-art methods on multiple real-world datasets for classification and regression tasks.

Conclusion: G$^{2}$D effectively addresses modality imbalance in multimodal learning through knowledge distillation and sequential modality prioritization, improving overall model performance.

Abstract: Multimodal learning aims to leverage information from diverse data modalities to achieve more comprehensive performance. However, conventional multimodal models often suffer from modality imbalance, where one or a few modalities dominate model optimization, leading to suboptimal feature representation and underutilization of weak modalities. To address this challenge, we introduce Gradient-Guided Distillation (G$^{2}$D), a knowledge distillation framework that optimizes the multimodal model with a custom-built loss function that fuses both unimodal and multimodal objectives. G$^{2}$D further incorporates a dynamic sequential modality prioritization (SMP) technique in the learning process to ensure each modality leads the learning process, avoiding the pitfall of stronger modalities overshadowing weaker ones. We validate G$^{2}$D on multiple real-world datasets and show that G$^{2}$D amplifies the significance of weak modalities while training and outperforms state-of-the-art methods in classification and regression tasks. Our code is available at https://github.com/rAIson-Lab/G2D.

[441] AI-Generated Video Detection via Perceptual Straightening

Christian Internò, Robert Geirhos, Markus Olhofer, Sunny Liu, Barbara Hammer, David Klindt

Main category: cs.CV

TL;DR: ReStraV detects AI-generated videos by analyzing temporal curvature and stepwise distance in neural representations, achieving 97.17% accuracy on VidProM benchmark.

DetailsMotivation: Address the urgent need for detecting realistic AI-generated videos as existing methods struggle with generalization and capturing temporal inconsistencies.

Method: Uses pre-trained DINOv2 vision transformer to quantify temporal curvature and stepwise distance in representation domain, then trains classifier on aggregated statistics.

Result: Achieves state-of-the-art performance with 97.17% accuracy and 98.63% AUROC on VidProM benchmark, substantially outperforming existing methods.

Conclusion: ReStraV provides computationally efficient detection using neural representation geometry, offering new insights for AI-generated video authentication.

Abstract: The rapid advancement of generative AI enables highly realistic synthetic videos, posing significant challenges for content authentication and raising urgent concerns about misuse. Existing detection methods often struggle with generalization and capturing subtle temporal inconsistencies. We propose ReStraV(Representation Straightening Video), a novel approach to distinguish natural from AI-generated videos. Inspired by the “perceptual straightening” hypothesis – which suggests real-world video trajectories become more straight in neural representation domain – we analyze deviations from this expected geometric property. Using a pre-trained self-supervised vision transformer (DINOv2), we quantify the temporal curvature and stepwise distance in the model’s representation domain. We aggregate statistics of these measures for each video and train a classifier. Our analysis shows that AI-generated videos exhibit significantly different curvature and distance patterns compared to real videos. A lightweight classifier achieves state-of-the-art detection performance (e.g., 97.17% accuracy and 98.63% AUROC on the VidProM benchmark), substantially outperforming existing image- and video-based methods. ReStraV is computationally efficient, it is offering a low-cost and effective detection solution. This work provides new insights into using neural representation geometry for AI-generated video detection.

[442] HOI-Dyn: Learning Interaction Dynamics for Human-Object Motion Diffusion

Lin Wu, Zhixiang Chen, Jianglin Lan

Main category: cs.CV

TL;DR: HOI-Dyn is a framework that generates 3D human-object interactions by modeling them as a driver-responder system where human actions drive object responses, using a transformer-based dynamics model and residual-based loss for improved physical plausibility.

DetailsMotivation: Existing methods treat human and object motions independently, resulting in physically implausible and causally inconsistent behaviors in 3D human-object interaction generation.

Method: Formulates HOI generation as driver-responder system with lightweight transformer-based interaction dynamics model that predicts object responses to human motion, plus residual-based dynamics loss to mitigate prediction errors. The dynamics model is only used during training.

Result: Extensive experiments show the approach enhances HOI generation quality and establishes a feasible metric for evaluating generated interactions.

Conclusion: HOI-Dyn successfully improves 3D human-object interaction generation by explicitly modeling interaction dynamics while maintaining inference efficiency.

Abstract: Generating realistic 3D human-object interactions (HOIs) remains a challenging task due to the difficulty of modeling detailed interaction dynamics. Existing methods treat human and object motions independently, resulting in physically implausible and causally inconsistent behaviors. In this work, we present HOI-Dyn, a novel framework that formulates HOI generation as a driver-responder system, where human actions drive object responses. At the core of our method is a lightweight transformer-based interaction dynamics model that explicitly predicts how objects should react to human motion. To further enforce consistency, we introduce a residual-based dynamics loss that mitigates the impact of dynamics prediction errors and prevents misleading optimization signals. The dynamics model is used only during training, preserving inference efficiency. Through extensive qualitative and quantitative experiments, we demonstrate that our approach not only enhances the quality of HOI generation but also establishes a feasible metric for evaluating the quality of generated interactions.

[443] Advancing Complex Wide-Area Scene Understanding with Hierarchical Coresets Selection

Jingyao Wang, Yiming Chen, Lingyu Si, Changwen Zheng

Main category: cs.CV

TL;DR: Proposes Hierarchical Coresets Selection (HCS) mechanism for Vision-Language Models to improve adaptation to unseen complex wide-area scenes without fine-tuning.

DetailsMotivation: Existing VLMs face challenges in adapting to unseen complex wide-area scenes due to insufficient feature density and poor generalization.

Method: Uses hierarchical coresets selection with theoretically guaranteed importance function considering utility, representativeness, robustness, and synergy to progressively refine selected regions.

Result: HCS enables VLMs to achieve rapid understanding of unseen scenes at any scale using minimal interpretable regions, with superior performance and universality in various tasks.

Conclusion: HCS is an effective plug-and-play method that enhances VLM adaptation to complex wide-area scenes without requiring additional fine-tuning.

Abstract: Scene understanding is one of the core tasks in computer vision, aiming to extract semantic information from images to identify objects, scene categories, and their interrelationships. Although advancements in Vision-Language Models (VLMs) have driven progress in this field, existing VLMs still face challenges in adaptation to unseen complex wide-area scenes. To address the challenges, this paper proposes a Hierarchical Coresets Selection (HCS) mechanism to advance the adaptation of VLMs in complex wide-area scene understanding. It progressively refines the selected regions based on the proposed theoretically guaranteed importance function, which considers utility, representativeness, robustness, and synergy. Without requiring additional fine-tuning, HCS enables VLMs to achieve rapid understandings of unseen scenes at any scale using minimal interpretable regions while mitigating insufficient feature density. HCS is a plug-and-play method that is compatible with any VLM. Experiments demonstrate that HCS achieves superior performance and universality in various tasks.

[444] Aesthetics is Cheap, Show me the Text: An Empirical Evaluation of State-of-the-Art Generative Models for OCR

Peirong Zhang, Haowei Xu, Jiaxin Zhang, Guitao Xu, Xuhan Zheng, Zhenhua Yang, Junle Liu, Yuyi Zhang, Lianwen Jin

Main category: cs.CV

TL;DR: This paper evaluates state-of-the-art generative models’ capabilities in text image generation and editing, incorporating OCR tasks across 33 representative categories including document, handwritten text, scene text, artistic text, and complex layout-rich text.

DetailsMotivation: The motivation is to assess whether current advanced generative models (like Flux-series and GPT-4o) can master the intricacies of text image generation and editing, given their exceptional fidelity in general image generation.

Method: The authors incorporate various OCR tasks into evaluation, categorize 33 representative tasks into five categories, and examine six models across closed-source and open-source domains using tailored high-quality image inputs and prompts.

Result: The evaluation identifies crucial observations and weaknesses of current generative models for OCR tasks, showing limitations in their text image generation and editing capabilities.

Conclusion: The authors argue that photorealistic text image generation and editing should be internalized as foundational skills into general-domain generative models rather than delegated to specialized solutions, and provide empirical analysis to guide future development.

Abstract: Text image is a unique and crucial information medium that integrates visual aesthetics and linguistic semantics in modern e-society. Due to their subtlety and complexity, the generation of text images represents a challenging and evolving frontier in the image generation field. The recent surge of specialized image generators (\emph{e.g.}, Flux-series) and unified generative models (\emph{e.g.}, GPT-4o), which demonstrate exceptional fidelity, raises a natural question: can they master the intricacies of text image generation and editing? Motivated by this, we assess current state-of-the-art generative models’ capabilities in terms of text image generation and editing. We incorporate various typical optical character recognition (OCR) tasks into our evaluation and broaden the concept of text-based generation tasks into OCR generative tasks. We select 33 representative tasks and categorize them into five categories: document, handwritten text, scene text, artistic text, and complex & layout-rich text. For comprehensive evaluation, we examine six models across both closed-source and open-source domains, using tailored, high-quality image inputs and prompts. Through this evaluation, we draw crucial observations and identify the weaknesses of current generative models for OCR tasks. We argue that photorealistic text image generation and editing should be internalized as foundational skills into general-domain generative models, rather than being delegated to specialized solutions, and we hope this empirical analysis can provide valuable insights for the community to achieve this goal. This evaluation is online and will be continuously updated at our GitHub repository.

[445] Attention (as Discrete-Time Markov) Chains

Yotam Erel, Olaf Dünkel, Rishabh Dabral, Vladislav Golyanik, Christian Theobalt, Amit H. Bermano

Main category: cs.CV

TL;DR: The paper introduces a novel interpretation of attention matrices as discrete-time Markov chains, enabling unified analysis of attention operations and extending them to model indirect attention effects through metastable states.

DetailsMotivation: To provide a unified framework for understanding attention operations (selection, summation, averaging) and extend beyond immediate attention effects to capture indirect attention propagation through Markov chain dynamics.

Method: Interpret attention matrix as Markov chain transition matrix; identify metastable states where attention concentrates; compute TokenRank as steady state vector; use matrix multiplication and eigenanalysis for lightweight computation.

Result: Achieves state-of-the-art zero-shot segmentation; TokenRank improves unconditional image generation quality (IS) and diversity (FID); enhances existing segmentation techniques; provides fresh perspective on token attention in visual transformers.

Conclusion: The Markov chain interpretation offers a powerful framework for analyzing attention mechanisms, revealing metastable states and enabling practical applications in segmentation and generation tasks through lightweight computational tools.

Abstract: We introduce a new interpretation of the attention matrix as a discrete-time Markov chain. Our interpretation sheds light on common operations involving attention scores such as selection, summation, and averaging in a unified framework. It further extends them by considering indirect attention, propagated through the Markov chain, as opposed to previous studies that only model immediate effects. Our key observation is that tokens linked to semantically similar regions form metastable states, i.e., regions where attention tends to concentrate, while noisy attention scores dissipate. Metastable states and their prevalence can be easily computed through simple matrix multiplication and eigenanalysis, respectively. Using these lightweight tools, we demonstrate state-of-the-art zero-shot segmentation. Lastly, we define TokenRank – the steady state vector of the Markov chain, which measures global token importance. We show that TokenRank enhances unconditional image generation, improving both quality (IS) and diversity (FID), and can also be incorporated into existing segmentation techniques to improve their performance over existing benchmarks. We believe our framework offers a fresh view of how tokens are being attended in modern visual transformers.

[446] BokehDiff: Neural Lens Blur with One-Step Diffusion

Chengxuan Zhu, Qingnan Fan, Qi Zhang, Jinwei Chen, Huaqi Zhang, Chao Xu, Boxin Shi

Main category: cs.CV

TL;DR: BokehDiff is a novel lens blur rendering method that uses generative diffusion prior to achieve physically accurate and visually appealing bokeh effects, overcoming limitations of depth estimation methods.

DetailsMotivation: Previous methods for lens blur rendering are limited by depth estimation accuracy, causing artifacts at depth discontinuities. There's also a lack of scalable paired data for training.

Method: Uses physics-inspired self-attention module with depth-dependent circle of confusion constraint and self-occlusion effects. Adapts diffusion model to one-step inference without additional noise. Synthesizes photorealistic foregrounds with transparency using diffusion models for training data.

Result: Achieves physically accurate and visually appealing lens blur outcomes with high quality and fidelity. Overcomes artifacts in depth discontinuities that plague previous methods.

Conclusion: BokehDiff successfully integrates generative diffusion prior with physics-inspired constraints to produce superior lens blur rendering, addressing key limitations in existing approaches.

Abstract: We introduce BokehDiff, a novel lens blur rendering method that achieves physically accurate and visually appealing outcomes, with the help of generative diffusion prior. Previous methods are bounded by the accuracy of depth estimation, generating artifacts in depth discontinuities. Our method employs a physics-inspired self-attention module that aligns with the image formation process, incorporating depth-dependent circle of confusion constraint and self-occlusion effects. We adapt the diffusion model to the one-step inference scheme without introducing additional noise, and achieve results of high quality and fidelity. To address the lack of scalable paired data, we propose to synthesize photorealistic foregrounds with transparency with diffusion models, balancing authenticity and scene diversity.

[447] ScreenCoder: Advancing Visual-to-Code Generation for Front-End Automation via Modular Multimodal Agents

Yilei Jiang, Yaozhi Zheng, Yuxuan Wan, Jiaming Han, Qunzhong Wang, Michael R. Lyu, Xiangyu Yue

Main category: cs.CV

TL;DR: ScreenCoder is a modular multi-agent framework that decomposes UI-to-code generation into three specialized stages (grounding, planning, generation) to overcome limitations of monolithic MLLMs, achieving state-of-the-art performance while also serving as a scalable data engine.

DetailsMotivation: Current multimodal LLMs struggle with complex UI-to-code translation due to difficulties in unifying visual perception, layout planning, and code synthesis within a single model, leading to frequent errors.

Method: Proposes ScreenCoder framework with three specialized agents: grounding agent for visual perception, planning agent for layout planning, and generation agent for code synthesis. Also uses the framework as a data engine to generate high-quality image-code pairs for fine-tuning open-source MLLMs via supervised fine-tuning and reinforcement learning.

Result: Achieves state-of-the-art performance in layout accuracy, structural coherence, and code correctness. The approach demonstrates substantial gains in UI generation capabilities through the dual-stage fine-tuning pipeline.

Conclusion: ScreenCoder’s modular multi-agent approach provides higher robustness and fidelity than end-to-end methods, successfully addressing the limitations of monolithic MLLMs in complex UI-to-code transformation tasks.

Abstract: Automating the transformation of user interface (UI) designs into front-end code holds significant promise for accelerating software development and democratizing design workflows. While multimodal large language models (MLLMs) can translate images to code, they often fail on complex UIs, struggling to unify visual perception, layout planning, and code synthesis within a single monolithic model, which leads to frequent perception and planning errors. To address this, we propose ScreenCoder, a modular multi-agent framework that decomposes the task into three interpretable stages: grounding, planning, and generation. By assigning these distinct responsibilities to specialized agents, our framework achieves significantly higher robustness and fidelity than end-to-end approaches. Furthermore, ScreenCoder serves as a scalable data engine, enabling us to generate high-quality image-code pairs. We use this data to fine-tune open-source MLLM via a dual-stage pipeline of supervised fine-tuning and reinforcement learning, demonstrating substantial gains in its UI generation capabilities. Extensive experiments demonstrate that our approach achieves state-of-the-art performance in layout accuracy, structural coherence, and code correctness. Our code is made publicly available at https://github.com/leigest519/ScreenCoder.

[448] SegDAC: Improving Visual Reinforcement Learning by Extracting Dynamic Objectc-Centric Representations from Pretrained Vision Models

Alexandre Brown, Glen Berseth

Main category: cs.CV

TL;DR: SegDAC is a Segmentation-Driven Actor-Critic method that uses SAM and YOLO-World for object-centric decomposition and learns to focus on relevant segments through online RL, achieving superior visual generalization and sample efficiency in manipulation tasks.

DetailsMotivation: Visual RL faces challenges in extracting useful representations from high-dimensional inputs and learning effective control from sparse rewards. Existing perception models are not effectively integrated into RL for visual generalization and sample efficiency.

Method: SegDAC uses Segment Anything (SAM) for object-centric decomposition and YOLO-World to ground segmentation via text inputs. It employs a transformer-based architecture that supports dynamic segments and learns which segments to focus on using online RL without human labels.

Result: SegDAC achieves significantly better visual generalization, doubling prior performance on the hardest setting and matching or surpassing prior methods in sample efficiency across all evaluated tasks on the Maniskill3 benchmark.

Conclusion: SegDAC effectively integrates segmentation models into RL, enabling superior visual generalization and sample efficiency in challenging manipulation tasks under strong visual perturbations.

Abstract: Visual reinforcement learning (RL) is challenging due to the need to extract useful representations from high-dimensional inputs while learning effective control from sparse and noisy rewards. Although large perception models exist, integrating them effectively into RL for visual generalization and improved sample efficiency remains difficult. We propose SegDAC, a Segmentation-Driven Actor-Critic method. SegDAC uses Segment Anything (SAM) for object-centric decomposition and YOLO-World to ground the image segmentation process via text inputs. It includes a novel transformer-based architecture that supports a dynamic number of segments at each time step and effectively learns which segments to focus on using online RL, without using human labels. By evaluating SegDAC over a challenging visual generalization benchmark using Maniskill3, which covers diverse manipulation tasks under strong visual perturbations, we demonstrate that SegDAC achieves significantly better visual generalization, doubling prior performance on the hardest setting and matching or surpassing prior methods in sample efficiency across all evaluated tasks.

[449] Interpretable Decision-Making for End-to-End Autonomous Driving

Mona Mirzaie, Bodo Rosenhahn

Main category: cs.CV

TL;DR: The paper presents an interpretable end-to-end autonomous driving method that generates sparse and localized feature maps to explain AI decisions while optimizing control commands, achieving state-of-the-art performance on CARLA benchmarks.

DetailsMotivation: End-to-end autonomous driving approaches are challenging to interpret due to deep neural networks with non-linear decision boundaries, which hinders trust in AI systems for safety-critical applications like autonomous vehicles.

Method: Proposed loss functions that promote interpretability by generating sparse and localized feature maps, allowing identification of which image regions contribute to predicted control commands. Conducted ablation studies on feature extraction and validated on CARLA benchmarks.

Result: The method improves interpretability while reducing infractions, achieving the highest route completion rate and lower infraction scores than top-performing approaches on CARLA Leaderboard. The monocular, non-ensemble model surpasses existing methods.

Conclusion: The approach successfully enhances interpretability in autonomous driving while maintaining high performance, demonstrating that interpretable AI can yield safer driving models without compromising on route completion capabilities.

Abstract: Trustworthy AI is mandatory for the broad deployment of autonomous vehicles. Although end-to-end approaches derive control commands directly from raw data, interpreting these decisions remains challenging, especially in complex urban scenarios. This is mainly attributed to very deep neural networks with non-linear decision boundaries, making it challenging to grasp the logic behind AI-driven decisions. This paper presents a method to enhance interpretability while optimizing control commands in autonomous driving. To address this, we propose loss functions that promote the interpretability of our model by generating sparse and localized feature maps. The feature activations allow us to explain which image regions contribute to the predicted control command. We conduct comprehensive ablation studies on the feature extraction step and validate our method on the CARLA benchmarks. We also demonstrate that our approach improves interpretability, which correlates with reducing infractions, yielding a safer, high-performance driving model. Notably, our monocular, non-ensemble model surpasses the top-performing approaches from the CARLA Leaderboard by achieving lower infraction scores and the highest route completion rate, all while ensuring interpretability.

[450] FlowDet: Overcoming Perspective and Scale Challenges in Real-Time End-to-End Traffic Detection

Zixing Wang, Yuhang Zhao

Main category: cs.CV

TL;DR: FlowDet is a high-speed end-to-end object detector that uses decoupled encoder optimization with Geometric Deformable Unit and Scale-Aware Attention modules, achieving state-of-the-art performance on the new Intersection-Flow-5k dataset while significantly reducing computational costs.

DetailsMotivation: To address the high computational cost of end-to-end object detectors, particularly for complex real-time applications like intersection traffic monitoring where NMS-free detectors are preferred but current implementations are too computationally expensive.

Method: Proposes FlowDet with decoupled encoder optimization on DETR architecture, featuring Geometric Deformable Unit (GDU) for traffic-aware geometric modeling and Scale-Aware Attention (SAA) module to handle extreme scale variations.

Result: On Intersection-Flow-5k dataset, FlowDet improves AP(test) by 1.5% and AP50(test) by 1.6% over RT-DETR baseline, while reducing GFLOPs by 63.2% and increasing inference speed by 16.2%.

Conclusion: FlowDet demonstrates a new path for building highly efficient and accurate detectors for demanding real-world perception systems, showing significant improvements in both accuracy and computational efficiency.

Abstract: End-to-end object detectors offer a promising NMS-free paradigm for real-time applications, yet their high computational cost remains a significant barrier, particularly for complex scenarios like intersection traffic monitoring. To address this challenge, we propose FlowDet, a high-speed detector featuring a decoupled encoder optimization strategy applied to the DETR architecture. Specifically, FlowDet employs a novel Geometric Deformable Unit (GDU) for traffic-aware geometric modeling and a Scale-Aware Attention (SAA) module to maintain high representational power across extreme scale variations. To rigorously evaluate the model’s performance in environments with severe occlusion and high object density, we collected the Intersection-Flow-5k dataset, a new challenging scene for this task. Evaluated on Intersection-Flow-5k, FlowDet establishes a new state-of-the-art. Compared to the strong RT-DETR baseline, it improves AP(test) by 1.5% and AP50(test) by 1.6%, while simultaneously reducing GFLOPs by 63.2% and increasing inference speed by 16.2%. Our work demonstrates a new path towards building highly efficient and accurate detectors for demanding, real-world perception systems. The Intersection-Flow-5k dataset is available at https://github.com/AstronZh/Intersection-Flow-5K.

[451] NeRF-based Visualization of 3D Cues Supporting Data-Driven Spacecraft Pose Estimation

Antoine Legrand, Renaud Detry, Christophe De Vleeschouwer

Main category: cs.CV

TL;DR: A method to visualize 3D visual cues used by spacecraft pose estimation networks by training a NeRF-based image generator using gradients from the pose estimator.

DetailsMotivation: Data-driven spacecraft pose estimation methods lack interpretability, hindering their adoption in real missions due to unclear decision processes.

Method: Train a NeRF-based image generator using gradients back-propagated through the pose estimation network to render the main 3D features exploited by the estimator.

Result: The method successfully recovers relevant 3D cues and provides insights into the relationship between network supervision and its implicit representation of the target spacecraft.

Conclusion: The proposed visualization approach enhances understanding of pose estimation networks’ decision processes, potentially facilitating their adoption in real spacecraft missions.

Abstract: On-orbit operations require the estimation of the relative 6D pose, i.e., position and orientation, between a chaser spacecraft and its target. While data-driven spacecraft pose estimation methods have been developed, their adoption in real missions is hampered by the lack of understanding of their decision process. This paper presents a method to visualize the 3D visual cues on which a given pose estimator relies. For this purpose, we train a NeRF-based image generator using the gradients back-propagated through the pose estimation network. This enforces the generator to render the main 3D features exploited by the spacecraft pose estimation network. Experiments demonstrate that our method recovers the relevant 3D cues. Furthermore, they offer additional insights on the relationship between the pose estimation network supervision and its implicit representation of the target spacecraft.

[452] The 1st Solution for 7th LSVOS RVOS Track: SaSaSa2VA

Quanzhu Niu, Dengxian Gong, Shihao Chen, Tao Zhang, Yikang Zhou, Haobo Yuan, Lu Qi, Xiangtai Li, Shunping Ji

Main category: cs.CV

TL;DR: SaSaSa2VA improves RVOS performance by addressing sparse frame sampling and single [SEG] token limitations through segmentation augmentation and test-time ensembling, achieving state-of-the-art results on LSVOS Challenge.

DetailsMotivation: To overcome performance bottlenecks in referring video object segmentation (RVOS) caused by sparse frame sampling and reliance on a single [SEG] token for entire videos, which limit fine-grained understanding of appearance and motion.

Method: Proposes Segmentation Augmented and Selective Averaged Sa2VA (SaSaSa2VA), building on Sa2VA framework by incorporating efficient segmentation augmentation and test-time ensembling techniques.

Result: Achieved J&F score of 67.45 on 7th LSVOS Challenge (RVOS track), ranking first and surpassing runner-up by 2.80 points.

Conclusion: Efficient segmentation augmentation and test-time ensembling substantially enhance grounded MLLMs for RVOS, demonstrating significant performance improvements over existing methods.

Abstract: Referring video object segmentation (RVOS) requires segmenting and tracking objects in videos conditioned on natural-language expressions, demanding fine-grained understanding of both appearance and motion. Building on Sa2VA, which couples a Multi-modal Large Language Model (MLLM) with the video segmentation model SAM2, we identify two key bottlenecks that limit segmentation performance: sparse frame sampling and reliance on a single [SEG] token for an entire video. We propose Segmentation Augmented and Selective Averaged Sa2VA (SaSaSa2VA) to address these issues. On the 7th LSVOS Challenge (RVOS track), SaSaSa2VA achieves a $\mathcal{J&F}$ of 67.45, ranking first and surpassing the runner-up by 2.80 points. This result and ablation studies demonstrate that efficient segmentation augmentation and test-time ensembling substantially enhance grounded MLLMs for RVOS. The code is released in Sa2VA repository: https://github.com/bytedance/Sa2VA.

[453] Accurate and Efficient Low-Rank Model Merging in Core Space

Aniello Panariello, Daniel Marczak, Simone Magistri, Angelo Porrello, Bartłomiej Twardowski, Andrew D. Bagdanov, Simone Calderara, Joost van de Weijer

Main category: cs.CV

TL;DR: Core Space framework enables efficient merging of LoRA-adapted models by projecting them into a common alignment basis, preserving low-rank efficiency while improving accuracy across vision and language tasks.

DetailsMotivation: Existing methods for merging LoRA-adapted models sacrifice efficiency by merging fully-sized weight matrices, losing the benefits of parameter-efficient adaptation.

Method: Project LoRA-adapted models into a common alignment basis (Core Space) with formal proof that this preserves all information, enabling efficient merging while maintaining low-rank structure.

Result: Significantly improves existing merging techniques, achieves state-of-the-art results on vision and language tasks, and uses only a fraction of computational resources compared to traditional methods.

Conclusion: Core Space merging framework successfully preserves the efficiency of low-rank adaptation while substantially improving model merging accuracy across multiple domains.

Abstract: In this paper, we address the challenges associated with merging low-rank adaptations of large neural networks. With the rise of parameter-efficient adaptation techniques, such as Low-Rank Adaptation (LoRA), model fine-tuning has become more accessible. While fine-tuning models with LoRA is highly efficient, existing merging methods often sacrifice this efficiency by merging fully-sized weight matrices. We propose the Core Space merging framework, which enables the merging of LoRA-adapted models within a common alignment basis, thereby preserving the efficiency of low-rank adaptation while substantially improving accuracy across tasks. We further provide a formal proof that projection into Core Space ensures no loss of information and provide a complexity analysis showing the efficiency gains. Extensive empirical results demonstrate that Core Space significantly improves existing merging techniques and achieves state-of-the-art results on both vision and language tasks while utilizing a fraction of the computational resources. Codebase is available at https://github.com/apanariello4/core-space-merging.

[454] Dolphin v1.0 Technical Report

Taohan Weng, Kaibing Hu, Henan Liu, Siya Liu, Xiaoyang Liu, Zhenyu Liu, Jiren Ren, Boyan Wang, Boyang Wang, Yiyu Wang, Yalun Wu, Chaoran Yan, Kaiwen Yan, Jinze Yu, Chi Zhang, Duo Zhang, Haoyun Zheng, Xiaoqing Guo, Jacques Souquet, Hongcheng Guo, Anjie Le

Main category: cs.CV

TL;DR: Dolphin v1.0 and its reasoning-augmented version Dolphin R1 are the first large-scale multimodal ultrasound foundation models that unify diverse clinical tasks in a single vision-language framework, addressing ultrasound’s challenges like operator dependence and image noise.

DetailsMotivation: Ultrasound faces challenges like operator dependence, image noise, and real-time scanning that hinder AI integration. While large multimodal models excel in other medical imaging areas, they struggle with ultrasound's complexities.

Method: Curated a 2-million-scale multimodal dataset combining textbook knowledge, public data, synthetic samples, and general corpora. Employed three-stage training: domain-specialized pretraining, instruction-driven alignment, and reinforcement-based refinement. Dolphin R1 enhances diagnostic inference through reinforcement learning with ultrasound-specific rewards.

Result: Dolphin R1 achieves a U2-score of 0.5835 on U2-Bench across eight ultrasound tasks - over twice the second-best model (0.2968), setting a new state of the art. Dolphin v1.0 also performs competitively.

Conclusion: The Dolphin series demonstrates that reasoning-enhanced training significantly improves diagnostic accuracy, consistency, and interpretability, highlighting its importance for high-stakes medical AI applications in ultrasound.

Abstract: Ultrasound is crucial in modern medicine but faces challenges like operator dependence, image noise, and real-time scanning, hindering AI integration. While large multimodal models excel in other medical imaging areas, they struggle with ultrasound’s complexities. To address this, we introduce Dolphin v1.0 (V1) and its reasoning-augmented version, Dolphin R1-the first large-scale multimodal ultrasound foundation models unifying diverse clinical tasks in a single vision-language framework.To tackle ultrasound variability and noise, we curated a 2-million-scale multimodal dataset, combining textbook knowledge, public data, synthetic samples, and general corpora. This ensures robust perception, generalization, and clinical adaptability.The Dolphin series employs a three-stage training strategy: domain-specialized pretraining, instruction-driven alignment, and reinforcement-based refinement. Dolphin v1.0 delivers reliable performance in classification, detection, regression, and report generation. Dolphin R1 enhances diagnostic inference, reasoning transparency, and interpretability through reinforcement learning with ultrasound-specific rewards.Evaluated on U2-Bench across eight ultrasound tasks, Dolphin R1 achieves a U2-score of 0.5835-over twice the second-best model (0.2968) setting a new state of the art. Dolphin v1.0 also performs competitively, validating the unified framework. Comparisons show reasoning-enhanced training significantly improves diagnostic accuracy, consistency, and interpretability, highlighting its importance for high-stakes medical AI.

[455] Learning Generalizable Shape Completion with SIM(3) Equivariance

Yuqing Wang, Zhaiyu Chen, Xiao Xiang Zhu

Main category: cs.CV

TL;DR: First SIM(3)-equivariant shape completion network that achieves robust generalization by being agnostic to pose and scale, outperforming existing methods on both synthetic and real-world benchmarks.

DetailsMotivation: Existing 3D shape completion methods rely on pre-aligned scans, which leak pose and scale cues. When alignment is absent in real data, performance collapses. Robust generalization requires equivariance to the similarity group SIM(3).

Method: Introduces a SIM(3)-equivariant shape completion network with modular layers that canonicalize features, reason over similarity-invariant geometry, and restore the original frame. Uses a de-biased evaluation protocol.

Result: Outperforms both equivariant and augmentation baselines on PCN benchmark. Sets new cross-domain records: lowers minimal matching distance on KITTI by 17% and Chamfer distance on OmniObject3D by 14%. Even under stricter protocol, outperforms competitors under their biased settings.

Conclusion: Full SIM(3) equivariance is an effective route to truly generalizable shape completion, establishing robustness across different domains and evaluation protocols.

Abstract: 3D shape completion methods typically assume scans are pre-aligned to a canonical frame. This leaks pose and scale cues that networks may exploit to memorize absolute positions rather than inferring intrinsic geometry. When such alignment is absent in real data, performance collapses. We argue that robust generalization demands architectural equivariance to the similarity group, SIM(3), so the model remains agnostic to pose and scale. Following this principle, we introduce the first SIM(3)-equivariant shape completion network, whose modular layers successively canonicalize features, reason over similarity-invariant geometry, and restore the original frame. Under a de-biased evaluation protocol that removes the hidden cues, our model outperforms both equivariant and augmentation baselines on the PCN benchmark. It also sets new cross-domain records on real driving and indoor scans, lowering minimal matching distance on KITTI by 17% and Chamfer distance $\ell1$ on OmniObject3D by 14%. Perhaps surprisingly, ours under the stricter protocol still outperforms competitors under their biased settings. These results establish full SIM(3) equivariance as an effective route to truly generalizable shape completion. Project page: https://sime-completion.github.io.

[456] Streaming Drag-Oriented Interactive Video Manipulation: Drag Anything, Anytime!

Junbao Zhou, Yuan Zhou, Kesen Zhao, Qingshan Xu, Beier Zhu, Richang Hong, Hanwang Zhang

Main category: cs.CV

TL;DR: REVEL enables interactive drag-based video manipulation anytime on anything, addressing challenges of latent distribution drift and context interference through DragStream’s training-free approach.

DetailsMotivation: To achieve streaming, fine-grained control over autoregressive video diffusion models and ensure consistent alignment with user expectations for interactive video manipulation.

Method: Proposed DragStream with adaptive distribution self-rectification strategy and spatial-frequency selective optimization mechanism to constrain latent drift and mitigate context interference.

Result: DragStream can be seamlessly integrated into existing autoregressive video diffusion models and effectively enables streaming drag-oriented interactive video manipulation.

Conclusion: The proposed DragStream approach successfully resolves the REVEL task, providing versatile drag operations for video editing and animation with translation, deformation, and rotation effects.

Abstract: Achieving streaming, fine-grained control over the outputs of autoregressive video diffusion models remains challenging, making it difficult to ensure that they consistently align with user expectations. To bridge this gap, we propose \textbf{stReaming drag-oriEnted interactiVe vidEo manipuLation (REVEL)}, a new task that enables users to modify generated videos \emph{anytime} on \emph{anything} via fine-grained, interactive drag. Beyond DragVideo and SG-I2V, REVEL unifies drag-style video manipulation as editing and animating video frames with both supporting user-specified translation, deformation, and rotation effects, making drag operations versatile. In resolving REVEL, we observe: \emph{i}) drag-induced perturbations accumulate in latent space, causing severe latent distribution drift that halts the drag process; \emph{ii}) streaming drag is easily disturbed by context frames, thereby yielding visually unnatural outcomes. We thus propose a training-free approach, \textbf{DragStream}, comprising: \emph{i}) an adaptive distribution self-rectification strategy that leverages neighboring frames’ statistics to effectively constrain the drift of latent embeddings; \emph{ii}) a spatial-frequency selective optimization mechanism, allowing the model to fully exploit contextual information while mitigating its interference via selectively propagating visual cues along generation. Our method can be seamlessly integrated into existing autoregressive video diffusion models, and extensive experiments firmly demonstrate the effectiveness of our DragStream.

[457] Mysteries of the Deep: Role of Intermediate Representations in Out of Distribution Detection

I. M. De la Jara, C. Rodriguez-Opazo, D. Teney, D. Ranasinghe, E. Abbasnejad

Main category: cs.CV

TL;DR: This paper challenges the conventional approach of using only final-layer representations for out-of-distribution (OOD) detection, revealing that intermediate layers in pre-trained models contain rich signals for detecting distributional shifts. The authors introduce an entropy-based method to automatically select complementary layers without needing OOD data, achieving significant improvements in detection accuracy.

DetailsMotivation: Most existing OOD detection methods treat pre-trained models as monolithic encoders and rely solely on final-layer representations, missing the valuable information encoded in intermediate layers that can better detect distributional shifts.

Method: The authors propose using intermediate layers of pre-trained models and introduce an entropy-based criterion to automatically identify layers that provide the most complementary information for OOD detection, all in a training-free setting without requiring access to OOD data.

Result: The method increases OOD detection accuracy by up to 10% in far-OOD benchmarks and over 7% in near-OOD benchmarks compared to state-of-the-art training-free methods, across various model architectures and training objectives.

Conclusion: This work reveals a new research direction for OOD detection by leveraging intermediate layer representations and demonstrates how different training objectives and model architectures affect confidence-based OOD detection methods.

Abstract: Out-of-distribution (OOD) detection is essential for reliably deploying machine learning models in the wild. Yet, most methods treat large pre-trained models as monolithic encoders and rely solely on their final-layer representations for detection. We challenge this wisdom. We reveal the \textit{intermediate layers} of pre-trained models, shaped by residual connections that subtly transform input projections, \textit{can} encode \textit{surprisingly rich and diverse signals} for detecting distributional shifts. Importantly, to exploit latent representation diversity across layers, we introduce an entropy-based criterion to \textit{automatically} identify layers offering the most complementary information in a training-free setting – \textit{without access to OOD data}. We show that selectively incorporating these intermediate representations can increase the accuracy of OOD detection by up to \textbf{$10%$} in far-OOD and over \textbf{$7%$} in near-OOD benchmarks compared to state-of-the-art training-free methods across various model architectures and training objectives. Our findings reveal a new avenue for OOD detection research and uncover the impact of various training objectives and model architectures on confidence-based OOD detection methods.

[458] Shaken or Stirred? An Analysis of MetaFormer’s Token Mixing for Medical Imaging

Ron Keuth, Paul Kaftan, Mattias P. Heinrich

Main category: cs.CV

TL;DR: Comprehensive study of token mixers in MetaFormer architecture for medical imaging, showing low-complexity mixers suffice for classification while convolutional mixers are essential for segmentation.

DetailsMotivation: MetaFormer has reshaped understanding of Transformer success in vision, but its use in medical imaging remains scarce with limited comparison of token mixers, potentially overlooking better design choices for medical tasks.

Method: Systematic analysis of pooling-, convolution-, and attention-based token mixers within MetaFormer architecture on 8 medical datasets covering classification and segmentation tasks, including transfer learning from pretrained weights.

Result: For classification: low-complexity token mixers (grouped convolution or pooling) are sufficient. For segmentation: convolutional token mixers with local inductive bias are essential, with grouped convolutions preferred for efficiency.

Conclusion: Grouped convolutions emerge as optimal token mixer for medical imaging, reducing runtime/parameters while MetaFormer’s channel-MLPs provide necessary cross-channel interactions. Pretrained weights remain useful despite domain gap.

Abstract: The generalization of the Transformer architecture via MetaFormer has reshaped our understanding of its success in computer vision. By replacing self-attention with simpler token mixers, MetaFormer provides strong baselines for vision tasks. However, while extensively studied on natural image datasets, its use in medical imaging remains scarce, and existing works rarely compare different token mixers, potentially overlooking more suitable designs choices. In this work, we present the first comprehensive study of token mixers for medical imaging. We systematically analyze pooling-, convolution-, and attention-based token mixers within the MetaFormer architecture on image classification (global prediction task) and semantic segmentation (dense prediction task). Our evaluation spans eight datasets covering diverse modalities and common challenges in the medical domain. Given the prevalence of pretraining from natural images to mitigate medical data scarcity, we also examine transferring pretrained weights to new token mixers. Our results show that, for classification, low-complexity token mixers (e.g. grouped convolution or pooling) are sufficient, aligning with findings on natural images. Pretrained weights remain useful despite the domain gap introduced by the new token mixer. For segmentation, we find that the local inductive bias of convolutional token mixers is essential. Grouped convolutions emerge as the preferred choice, as they reduce runtime and parameter count compared to standard convolutions, while the MetaFormer’s channel-MLPs already provide the necessary cross-channel interactions.

[459] Towards Better & Faster Autoregressive Image Generation: From the Perspective of Entropy

Xiaoxiao Ma, Feng Zhao, Pengyang Ling, Haibo Qiu, Zhixiang Wei, Hu Yu, Jie Huang, Zhixiong Zeng, Lin Ma

Main category: cs.CV

TL;DR: The paper proposes an entropy-informed decoding strategy for autoregressive image generation that improves generation quality and speed by addressing sampling issues in current AR models.

DetailsMotivation: Current autoregressive image generation models have sampling issues where image tokens have lower information density and non-uniform spatial distribution compared to text tokens, which affects generation quality and efficiency.

Method: Two main innovations: 1) dynamic temperature control guided by spatial entropy of token distributions to balance diversity, accuracy, and coherence, and 2) entropy-aware acceptance rules in speculative decoding for faster inference.

Result: Extensive experiments show the approach achieves higher generation quality with faster synthesis speed, achieving near-lossless generation at about 85% of conventional inference costs.

Conclusion: The entropy-informed decoding strategy effectively enhances both generation quality and sampling speed in autoregressive image generation models, demonstrating broad applicability across diverse AR models.

Abstract: In this work, we first revisit the sampling issues in current autoregressive (AR) image generation models and identify that image tokens, unlike text tokens, exhibit lower information density and non-uniform spatial distribution. Accordingly, we present an entropy-informed decoding strategy that facilitates higher autoregressive generation quality with faster synthesis speed. Specifically, the proposed method introduces two main innovations: 1) dynamic temperature control guided by spatial entropy of token distributions, enhancing the balance between content diversity, alignment accuracy, and structural coherence in both mask-based and scale-wise models, without extra computational overhead, and 2) entropy-aware acceptance rules in speculative decoding, achieving near-lossless generation at about 85% of the inference cost of conventional acceleration methods. Extensive experiments across multiple benchmarks using diverse AR image generation models demonstrate the effectiveness and generalizability of our approach in enhancing both generation quality and sampling speed.

[460] MSDM: Generating Task-Specific Pathology Images with a Multimodal Conditioned Diffusion Model for Cell and Nuclei Segmentation

Dominik Winter, Mai Bui, Monica Azqueta Gavaldon, Nicolas Triltsch, Marco Rosati, Nicolas Brieu

Main category: cs.CV

TL;DR: MSDM is a multimodal semantic diffusion model that generates realistic image-mask pairs for cell/nuclei segmentation, addressing data scarcity by creating synthetic data with controlled morphological properties.

DetailsMotivation: Address the scarcity of annotated data for rare or atypical cell morphologies in computational pathology, where manual annotation is labor-intensive and costly.

Method: Multimodal diffusion model conditioned on cellular/nuclear morphologies (horizontal/vertical maps), RGB color characteristics, and BERT-encoded assay metadata, integrated via multi-head cross-attention.

Result: Synthetic images closely match real data with low Wasserstein distances, and incorporating synthetic samples significantly improves segmentation model accuracy on rare cell types like columnar cells.

Conclusion: Multimodal diffusion-based augmentation effectively enriches datasets and improves robustness of cell/nuclei segmentation models, paving the way for broader generative model applications in computational pathology.

Abstract: Scarcity of annotated data, particularly for rare or atypical morphologies, present significant challenges for cell and nuclei segmentation in computational pathology. While manual annotation is labor-intensive and costly, synthetic data offers a cost-effective alternative. We introduce a Multimodal Semantic Diffusion Model (MSDM) for generating realistic pixel-precise image-mask pairs for cell and nuclei segmentation. By conditioning the generative process with cellular/nuclear morphologies (using horizontal and vertical maps), RGB color characteristics, and BERT-encoded assay/indication metadata, MSDM generates datasests with desired morphological properties. These heterogeneous modalities are integrated via multi-head cross-attention, enabling fine-grained control over the generated images. Quantitative analysis demonstrates that synthetic images closely match real data, with low Wasserstein distances between embeddings of generated and real images under matching biological conditions. The incorporation of these synthetic samples, exemplified by columnar cells, significantly improves segmentation model accuracy on columnar cells. This strategy systematically enriches data sets, directly targeting model deficiencies. We highlight the effectiveness of multimodal diffusion-based augmentation for advancing the robustness and generalizability of cell and nuclei segmentation models. Thereby, we pave the way for broader application of generative models in computational pathology.

[461] Tag-Enriched Multi-Attention with Large Language Models for Cross-Domain Sequential Recommendation

Wangyu Wu, Xuhang Chen, Zhenhong Chen, Jing-En Jiang, Kim-Fung Tsang, Xiaowei Huang, Fei Ma, Jimin Xiao

Main category: cs.CV

TL;DR: TEMA-LLM is a cross-domain sequential recommendation framework that uses LLMs for semantic tag generation and multi-attention mechanisms to capture both domain-specific and cross-domain user behaviors.

DetailsMotivation: To address the challenge of capturing both domain-specific and cross-domain behavioral patterns in modern e-commerce platforms where users interact with diverse services, enabling personalized and seamless consumer experiences.

Method: Uses LLMs to generate descriptive tags from item titles/descriptions, fuses tag embeddings with item identifiers and features, and employs a Tag-Enriched Multi-Attention mechanism to model user preferences within and across domains.

Result: Extensive experiments on four large-scale e-commerce datasets show TEMA-LLM consistently outperforms state-of-the-art baselines.

Conclusion: The approach demonstrates the potential of LLMs to advance intelligent, user-centric services in consumer electronics through semantic tagging and multi-attention integration.

Abstract: Cross-Domain Sequential Recommendation (CDSR) plays a crucial role in modern consumer electronics and e-commerce platforms, where users interact with diverse services such as books, movies, and online retail products. These systems must accurately capture both domain-specific and cross-domain behavioral patterns to provide personalized and seamless consumer experiences. To address this challenge, we propose \textbf{TEMA-LLM} (\textit{Tag-Enriched Multi-Attention with Large Language Models}), a practical and effective framework that integrates \textit{Large Language Models (LLMs)} for semantic tag generation and enrichment. Specifically, TEMA-LLM employs LLMs to assign domain-aware prompts and generate descriptive tags from item titles and descriptions. The resulting tag embeddings are fused with item identifiers as well as textual and visual features to construct enhanced item representations. A \textit{Tag-Enriched Multi-Attention} mechanism is then introduced to jointly model user preferences within and across domains, enabling the system to capture complex and evolving consumer interests. Extensive experiments on four large-scale e-commerce datasets demonstrate that TEMA-LLM consistently outperforms state-of-the-art baselines, underscoring the benefits of LLM-based semantic tagging and multi-attention integration for consumer-facing recommendation systems. The proposed approach highlights the potential of LLMs to advance intelligent, user-centric services in the field of consumer electronics.

[462] When Does Supervised Training Pay Off? The Hidden Economics of Object Detection in the Era of Vision-Language Models

Samer Al-Hamadani

Main category: cs.CV

TL;DR: Comprehensive cost-effectiveness analysis comparing supervised YOLO with zero-shot vision-language models (Gemini Flash 2.5, GPT-4) for object detection, revealing break-even thresholds and optimal architecture selection based on inference volume, category stability, budget, and accuracy requirements.

DetailsMotivation: Traditional object detection relies on costly manual annotation, creating a need to understand when zero-shot vision-language models become more cost-effective than supervised approaches.

Method: Evaluated supervised YOLO and zero-shot models (Gemini Flash 2.5, GPT-4) on 5,000 stratified COCO images and 500 diverse product images, combined with Total Cost of Ownership modeling to derive break-even thresholds.

Result: Supervised YOLO achieved 91.2% accuracy vs 68.5% for Gemini and 71.3% for GPT-4 on standard categories, but requires $10,800 annotation cost for 100 categories. The accuracy advantage only pays off beyond 55 million inferences. On diverse products, zero-shot models achieved 52.3%-55.1% accuracy while YOLO cannot detect untrained classes. Cost-per-detection favors zero-shot models at 100,000 inferences.

Conclusion: Optimal architecture choice depends on inference volume, category stability, budget, and accuracy requirements, with zero-shot models being more cost-effective for lower inference volumes and dynamic categories.

Abstract: Object detection traditionally relies on costly manual annotation. We present the first comprehensive cost-effectiveness analysis comparing supervised YOLO and zero-shot vision-language models (Gemini Flash 2.5 and GPT-4). Evaluated on 5,000 stratified COCO images and 500 diverse product images, combined with Total Cost of Ownership modeling, we derive break-even thresholds for architecture selection. Results show supervised YOLO attains 91.2% accuracy versus 68.5% for Gemini and 71.3% for GPT-4 on standard categories; the annotation expense for a 100-category system is $10,800, and the accuracy advantage only pays off beyond 55 million inferences (151,000 images/day for one year). On diverse product categories Gemini achieves 52.3% and GPT-4 55.1%, while supervised YOLO cannot detect untrained classes. Cost-per-correct-detection favors Gemini ($0.00050) and GPT-4 ($0.00067) over YOLO ($0.143) at 100,000 inferences. We provide decision frameworks showing that optimal architecture choice depends on inference volume, category stability, budget, and accuracy requirements.

[463] MaterialRefGS: Reflective Gaussian Splatting with Multi-view Consistent Material Inference

Wenyuan Zhang, Jimin Tang, Weiqi Zhang, Yi Fang, Yu-Shen Liu, Zhizhong Han

Main category: cs.CV

TL;DR: A method for modeling reflections in 2D images using Gaussian Splatting with multi-view consistent material inference and physically-based environment modeling to achieve accurate reflections and photorealistic rendering.

DetailsMotivation: Current approaches using Gaussian primitives for reflection modeling lack sufficient constraints, especially under limited environment modeling, leading to illumination aliasing and reduced generalization.

Method: Enforces 2D Gaussians to produce multi-view consistent material maps during deferred shading, tracks photometric variations across views to identify reflective regions, and introduces ray tracing with 2DGS for environment modeling to handle indirect illumination.

Result: Faithfully recovers both illumination and geometry, achieving state-of-the-art rendering quality in novel view synthesis on widely used benchmarks.

Conclusion: Multi-view consistent material inference with physically-based environment modeling is key to learning accurate reflections with Gaussian Splatting.

Abstract: Modeling reflections from 2D images is essential for photorealistic rendering and novel view synthesis. Recent approaches enhance Gaussian primitives with reflection-related material attributes to enable physically based rendering (PBR) with Gaussian Splatting. However, the material inference often lacks sufficient constraints, especially under limited environment modeling, resulting in illumination aliasing and reduced generalization. In this work, we revisit the problem from a multi-view perspective and show that multi-view consistent material inference with more physically-based environment modeling is key to learning accurate reflections with Gaussian Splatting. To this end, we enforce 2D Gaussians to produce multi-view consistent material maps during deferred shading. We also track photometric variations across views to identify highly reflective regions, which serve as strong priors for reflection strength terms. To handle indirect illumination caused by inter-object occlusions, we further introduce an environment modeling strategy through ray tracing with 2DGS, enabling photorealistic rendering of indirect radiance. Experiments on widely used benchmarks show that our method faithfully recovers both illumination and geometry, achieving state-of-the-art rendering quality in novel views synthesis.

[464] Vision Mamba for Permeability Prediction of Porous Media

Ali Kashefi, Tapan Mukerji

Main category: cs.CV

TL;DR: Vision Mamba is introduced as a backbone for 3D porous media permeability prediction, showing better computational efficiency than ViTs and fewer parameters than CNNs.

DetailsMotivation: To leverage Vision Mamba's linear scaling with input resolution (vs ViT's quadratic scaling) and smaller parameter count than CNNs for more efficient permeability prediction in 3D porous media.

Method: Used Vision Mamba as backbone for permeability prediction, compared with ViT and CNN models, and performed ablation studies to analyze component effects on accuracy.

Result: Demonstrated practical advantages of Vision Mamba over ViTs and CNNs in 3D porous media permeability prediction, showing improved computational and memory efficiency.

Conclusion: Vision Mamba has potential to replace ViTs in large vision models, with source code made available for reproducibility and further research.

Abstract: Vision Mamba has recently received attention as an alternative to Vision Transformers (ViTs) for image classification. The network size of Vision Mamba scales linearly with input image resolution, whereas ViTs scale quadratically, a feature that improves computational and memory efficiency. Moreover, Vision Mamba requires a significantly smaller number of trainable parameters than traditional convolutional neural networks (CNNs), and thus, they can be more memory efficient. Because of these features, we introduce, for the first time, a neural network that uses Vision Mamba as its backbone for predicting the permeability of three-dimensional porous media. We compare the performance of Vision Mamba with ViT and CNN models across multiple aspects of permeability prediction and perform an ablation study to assess the effects of its components on accuracy. We demonstrate in practice the aforementioned advantages of Vision Mamba over ViTs and CNNs in the permeability prediction of three-dimensional porous media. We make the source code publicly available to facilitate reproducibility and to enable other researchers to build on and extend this work. We believe the proposed framework has the potential to be integrated into large vision models in which Vision Mamba is used instead of ViTs.

[465] STANCE: Motion Coherent Video Generation Via Sparse-to-Dense Anchored Encoding

Zhifei Chen, Tianshuo Xu, Leyi Wu, Luozhou Wang, Dongyu Yan, Zihan You, Wenting Luo, Guo Zhang, Yingcong Chen

Main category: cs.CV

TL;DR: STANCE is an image-to-video framework that improves motion coherence using Instance Cues for dense motion guidance and Dense RoPE to preserve motion token salience, with joint RGB+auxiliary prediction for better temporal consistency.

DetailsMotivation: Current video generation struggles with coherent object motion and interactions due to weak motion guidance from sparse 2D hints and optimization conflicts between appearance and motion in single-head models.

Method: Uses Instance Cues to convert sparse user hints into dense 2.5D motion fields, and Dense RoPE to maintain motion token salience. Employs joint RGB+auxiliary map prediction to separate structure and appearance optimization.

Result: The approach reduces depth ambiguity compared to 2D inputs, stabilizes optimization, and improves temporal coherence without requiring per-frame trajectory scripts.

Conclusion: STANCE effectively addresses motion coherence issues in video generation through improved motion guidance and separated optimization of structure and appearance.

Abstract: Video generation has recently made striking visual progress, but maintaining coherent object motion and interactions remains difficult. We trace two practical bottlenecks: (i) human-provided motion hints (e.g., small 2D maps) often collapse to too few effective tokens after encoding, weakening guidance; and (ii) optimizing for appearance and motion in a single head can favor texture over temporal consistency. We present STANCE, an image-to-video framework that addresses both issues with two simple components. First, we introduce Instance Cues – a pixel-aligned control signal that turns sparse, user-editable hints into a dense 2.5D (camera-relative) motion field by averaging per-instance flow and augmenting with monocular depth over the instance mask. This reduces depth ambiguity compared to 2D arrow inputs while remaining easy to use. Second, we preserve the salience of these cues in token space with Dense RoPE, which tags a small set of motion tokens (anchored on the first frame) with spatial-addressable rotary embeddings. Paired with joint RGB (+) auxiliary-map prediction (segmentation or depth), our model anchors structure while RGB handles appearance, stabilizing optimization and improving temporal coherence without requiring per-frame trajectory scripts.

[466] Knowledge-based Visual Question Answer with Multimodal Processing, Retrieval and Filtering

Yuyang Hong, Jiaqi Gu, Qi Yang, Lubin Fan, Yue Wu, Ying Wang, Kun Ding, Shiming Xiang, Jieping Ye

Main category: cs.CV

TL;DR: Wiki-PRF is a three-stage method for KB-VQA that improves multimodal query quality and retrieval relevance through processing, retrieval, and filtering stages, achieving state-of-the-art performance.

DetailsMotivation: Current retrieval-augmented generation methods struggle with multimodal query quality and relevance of retrieved results in knowledge-based visual question answering tasks.

Method: Three-stage approach: 1) Processing stage uses visual tools to extract precise multimodal information; 2) Retrieval stage integrates visual and text features for knowledge retrieval; 3) Filtering stage performs relevance filtering. Uses reinforcement learning with answer accuracy and format consistency as rewards.

Result: Achieved significant improvements of 36.0 and 42.8 on E-VQA and InfoSeek benchmark datasets, reaching state-of-the-art performance.

Conclusion: Wiki-PRF effectively addresses multimodal query and retrieval challenges in KB-VQA through its three-stage architecture and reinforcement learning approach.

Abstract: Knowledge-based visual question answering (KB-VQA) requires visual language models (VLMs) to integrate visual understanding with external knowledge retrieval. Although retrieval-augmented generation (RAG) achieves significant advances in this task by combining knowledge-base querying, it still struggles with the quality of multimodal queries and the relevance of retrieved results. To overcome these challenges, we propose a novel three-stage method, termed Wiki-PRF, including Processing, Retrieval and Filtering stages. The processing stage dynamically invokes visual tools to extract precise multimodal information for retrieval. The retrieval stage integrates visual and text features to achieve multimodal knowledge retrieval. The filtering stage performs relevance filtering and concentration on retrieval results. To this end, we introduce a visual language model trained with answer accuracy and format consistency as reward signals via a reinforcement learning manner. This enhances the model’s reasoning, tool invocation for accurate queries, and filtering of irrelevant content. Experiments on benchmark datasets (E-VQA and InfoSeek) show significant improvements~(36.0 and 42.8) in answer quality, achieving state-of-the-art performance. Code is available at https://github.com/cqu-student/Wiki-PRF

[467] CARDIUM: Congenital Anomaly Recognition with Diagnostic Images and Unified Medical records

Daniela Vega, Hannah V. Ceballos, Javier S. Vera, Santiago Rodriguez, Alejandra Perez, Angela Castillo, Maria Escobar, Dario Londoño, Luis A. Sarmiento, Camila I. Castro, Nadiezhda Rodriguez, Juan C. Briceño, Pablo Arbeláez

Main category: cs.CV

TL;DR: The paper introduces CARDIUM, the first public multimodal dataset for prenatal CHD detection combining fetal ultrasound/echocardiographic images with maternal clinical records, and proposes a cross-attention transformer that improves detection by 11-50% over single-modality approaches.

DetailsMotivation: Current AI solutions for prenatal CHD diagnosis face challenges due to rare conditions causing imbalanced datasets and lack of public multimodal datasets integrating imaging and clinical data, limiting clinical decision support.

Method: Proposed a robust multimodal transformer architecture with cross-attention mechanism to fuse feature representations from image and tabular data, using the CARDIUM dataset containing fetal ultrasound/echocardiographic images and maternal clinical records.

Result: The multimodal approach improved CHD detection by 11% over image-only and 50% over tabular-only methods, achieving an F1 score of 79.8 ± 4.8% on the CARDIUM dataset.

Conclusion: The CARDIUM dataset and proposed multimodal transformer architecture significantly enhance prenatal CHD detection, with public release of dataset and code to advance research in this field.

Abstract: Prenatal diagnosis of Congenital Heart Diseases (CHDs) holds great potential for Artificial Intelligence (AI)-driven solutions. However, collecting high-quality diagnostic data remains difficult due to the rarity of these conditions, resulting in imbalanced and low-quality datasets that hinder model performance. Moreover, no public efforts have been made to integrate multiple sources of information, such as imaging and clinical data, further limiting the ability of AI models to support and enhance clinical decision-making. To overcome these challenges, we introduce the Congenital Anomaly Recognition with Diagnostic Images and Unified Medical records (CARDIUM) dataset, the first publicly available multimodal dataset consolidating fetal ultrasound and echocardiographic images along with maternal clinical records for prenatal CHD detection. Furthermore, we propose a robust multimodal transformer architecture that incorporates a cross-attention mechanism to fuse feature representations from image and tabular data, improving CHD detection by 11% and 50% over image and tabular single-modality approaches, respectively, and achieving an F1 score of 79.8 $\pm$ 4.8% in the CARDIUM dataset. We will publicly release our dataset and code to encourage further research on this unexplored field. Our dataset and code are available at https://github.com/BCV-Uniandes/Cardium, and at the project website https://bcv-uniandes.github.io/CardiumPage/

[468] Latent Diffusion Model without Variational Autoencoder

Minglei Shi, Haolin Wang, Wenzhao Zheng, Ziyang Yuan, Xiaoshi Wu, Xintao Wang, Pengfei Wan, Jie Zhou, Jiwen Lu

Main category: cs.CV

TL;DR: SVG introduces a novel latent diffusion model that replaces VAEs with self-supervised DINO features, creating semantically structured latent spaces for more efficient training and better generative quality.

DetailsMotivation: VAE+diffusion paradigm suffers from limited training efficiency, slow inference, poor transferability, and lacks clear semantic separation in latent spaces, which are crucial for stable diffusion training and broader vision tasks.

Method: Leverages frozen DINO features to construct semantically discriminative latent space, with a lightweight residual branch for fine-grained details. Diffusion models are trained directly on this structured space.

Result: Enables accelerated diffusion training, supports few-step sampling, improves generative quality, and preserves semantic/discriminative capabilities of self-supervised representations.

Conclusion: SVG provides a principled pathway toward task-general, high-quality visual representations by combining self-supervised features with diffusion models, overcoming VAE limitations.

Abstract: Recent progress in diffusion-based visual generation has largely relied on latent diffusion models with variational autoencoders (VAEs). While effective for high-fidelity synthesis, this VAE+diffusion paradigm suffers from limited training efficiency, slow inference, and poor transferability to broader vision tasks. These issues stem from a key limitation of VAE latent spaces: the lack of clear semantic separation and strong discriminative structure. Our analysis confirms that these properties are crucial not only for perception and understanding tasks, but also for the stable and efficient training of latent diffusion models. Motivated by this insight, we introduce SVG, a novel latent diffusion model without variational autoencoders, which leverages self-supervised representations for visual generation. SVG constructs a feature space with clear semantic discriminability by leveraging frozen DINO features, while a lightweight residual branch captures fine-grained details for high-fidelity reconstruction. Diffusion models are trained directly on this semantically structured latent space to facilitate more efficient learning. As a result, SVG enables accelerated diffusion training, supports few-step sampling, and improves generative quality. Experimental results further show that SVG preserves the semantic and discriminative capabilities of the underlying self-supervised representations, providing a principled pathway toward task-general, high-quality visual representations. Code and interpretations are available at https://howlin-wang.github.io/svg/.

[469] Learning to Detect Unknown Jailbreak Attacks in Large Vision-Language Models

Shuang Liang, Zhihao Xu, Jialing Tao, Hui Xue, Xiting Wang

Main category: cs.CV

TL;DR: LoD is a framework that detects unknown jailbreak attacks in Large Vision-Language Models by shifting from attack-specific to task-specific learning, using safety-oriented representation and unsupervised attack classification.

DetailsMotivation: Existing jailbreak detection methods either lack generalization to unseen attacks or have limited accuracy and efficiency, creating safety risks in LVLMs.

Method: Proposes Learning to Detect (LoD) with Multi-modal Safety Concept Activation Vector for safety representation learning and Safety Pattern Auto-Encoder for unsupervised attack classification.

Result: Achieves consistently higher detection AUROC on diverse unknown attacks while improving efficiency compared to existing methods.

Conclusion: LoD provides an effective general framework for detecting unknown jailbreak attacks in LVLMs by focusing on task-specific safety learning rather than attack-specific patterns.

Abstract: Despite extensive alignment efforts, Large Vision-Language Models (LVLMs) remain vulnerable to jailbreak attacks, posing serious safety risks. To address this, existing detection methods either learn attack-specific parameters, which hinders generalization to unseen attacks, or rely on heuristically sound principles, which limit accuracy and efficiency. To overcome these limitations, we propose Learning to Detect (LoD), a general framework that accurately detects unknown jailbreak attacks by shifting the focus from attack-specific learning to task-specific learning. This framework includes a Multi-modal Safety Concept Activation Vector module for safety-oriented representation learning and a Safety Pattern Auto-Encoder module for unsupervised attack classification. Extensive experiments show that our method achieves consistently higher detection AUROC on diverse unknown attacks while improving efficiency. The code is available at https://anonymous.4open.science/r/Learning-to-Detect-51CB.

cs.AI

[470] VisuoAlign: Safety Alignment of LVLMs with Multimodal Tree Search

MingSheng Li, Guangze Zhao, Sichen Liu

Main category: cs.AI

TL;DR: VisuoAlign is a framework that improves safety alignment in Large Vision-Language Models (LVLMs) using prompt-guided tree search to address multimodal jailbreak vulnerabilities.

DetailsMotivation: Current LVLMs have safety alignment challenges due to new attack surfaces from visual inputs, lack of safety supervision in reasoning chains, and alignment degradation during modality fusion, making them vulnerable to multimodal jailbreaks.

Method: VisuoAlign embeds safety constraints via visual-textual interactive prompts, uses Monte Carlo Tree Search (MCTS) to create diverse safety-critical prompt trajectories, and implements prompt-based scaling for real-time risk detection and compliant responses.

Result: Extensive experiments show VisuoAlign proactively exposes risks, enables comprehensive dataset generation, and significantly improves LVLM robustness against complex cross-modal threats.

Conclusion: VisuoAlign provides an effective framework for multimodal safety alignment that enhances LVLM security against sophisticated multimodal attacks through systematic prompt-guided tree search.

Abstract: Large Vision-Language Models (LVLMs) have achieved remarkable progress in multimodal perception and generation, yet their safety alignment remains a critical challenge.Existing defenses and vulnerable to multimodal jailbreaks, as visual inputs introduce new attack surfaces, reasoning chains lack safety supervision, and alignment often degrades under modality fusion.To overcome these limitation, we propose VisuoAlign, a framework for multi-modal safety alignment via prompt-guided tree search.VisuoAlign embeds safety constrains into the reasoning process through visual-textual interactive prompts, employs Monte Carlo Tree Search(MCTS) to systematically construct diverse safety-critical prompt trajectories, and introduces prompt-based scaling to ensure real-time risk detection and compliant responses.Extensive experiments demonstrate that VisuoAlign proactively exposes risks, enables comprehensive dataset generation, and significantly improves the robustness of LVLMs against complex cross-modal threats.

[471] Executable Epistemology: The Structured Cognitive Loop as an Architecture of Intentional Understanding

Myung Ho Kim

Main category: cs.AI

TL;DR: The paper introduces the Structured Cognitive Loop (SCL) as an executable epistemological framework for emergent intelligence, bridging philosophy and implementable cognition.

DetailsMotivation: Large language models lack genuine epistemic understanding despite exhibiting intelligence, revealing a gap in epistemic architecture. The paper aims to address this by developing a framework that enables emergent cognition.

Method: SCL is grounded in philosophy of mind, cognitive phenomenology, process philosophy, enactive cognition, and extended mind theory. It defines intelligence as a performed process through a continuous loop of judgment, memory, control, action, and regulation.

Result: SCL operationalizes philosophical insights into computationally interpretable structures, enables functional separation within cognitive architecture for more coherent behavior, and redefines intelligence as the capacity to reconstruct epistemic state through intentional understanding.

Conclusion: Real progress in AI requires architectures that structurally realize cognitive principles rather than larger models. SCL impacts philosophy of mind, epistemology, and AI by framing knowledge as continuous reconstruction within a phenomenologically coherent loop.

Abstract: Large language models exhibit intelligence without genuine epistemic understanding, exposing a key gap: the absence of epistemic architecture. This paper introduces the Structured Cognitive Loop (SCL) as an executable epistemological framework for emergent intelligence. Unlike traditional AI research asking “what is intelligence?” (ontological), SCL asks “under what conditions does cognition emerge?” (epistemological). Grounded in philosophy of mind and cognitive phenomenology, SCL bridges conceptual philosophy and implementable cognition. Drawing on process philosophy, enactive cognition, and extended mind theory, we define intelligence not as a property but as a performed process – a continuous loop of judgment, memory, control, action, and regulation. SCL makes three contributions. First, it operationalizes philosophical insights into computationally interpretable structures, enabling “executable epistemology” – philosophy as structural experiment. Second, it shows that functional separation within cognitive architecture yields more coherent and interpretable behavior than monolithic prompt based systems, supported by agent evaluations. Third, it redefines intelligence: not representational accuracy but the capacity to reconstruct its own epistemic state through intentional understanding. This framework impacts philosophy of mind, epistemology, and AI. For philosophy, it allows theories of cognition to be enacted and tested. For AI, it grounds behavior in epistemic structure rather than statistical regularity. For epistemology, it frames knowledge not as truth possession but as continuous reconstruction within a phenomenologically coherent loop. We situate SCL within debates on cognitive phenomenology, emergence, normativity, and intentionality, arguing that real progress requires not larger models but architectures that realize cognitive principles structurally.

[472] Exploring the Potential of Citiverses for Regulatory Learning

Isabelle Hupont, Marisa Ponti, Sven Schade

Main category: cs.AI

TL;DR: This paper proposes using citiverses (virtual worlds) as experimentation spaces for regulatory learning, identifying key research areas and experimental topics through expert consultation.

DetailsMotivation: To explore the potential of citiverses as virtual environments for testing policy scenarios and technologies, supporting regulatory learning through immersive experimentation.

Method: Consultation with a high-level panel of experts including European Commission policymakers, national government science advisers, and leading researchers in digital regulation and virtual worlds.

Result: Identified key research areas (scalability, real-time feedback, complexity modeling, cross-border collaboration, risk reduction, citizen participation, ethics, emerging tech integration) and experimental topics in transportation, urban planning, and climate crisis.

Conclusion: Citiverses can advance regulatory learning through responsible development that considers ethical, economic, ecological and social dimensions, and should be integrated with existing experimentation ecosystems like test beds and regulatory sandboxes.

Abstract: Citiverses hold the potential to support regulatory learning by offering immersive, virtual environments for experimenting with policy scenarios and technologies. This paper proposes a science-for-policy agenda to explore the potential of citiverses as experimentation spaces for regulatory learning, grounded in a consultation with a high-level panel of experts, including policymakers from the European Commission, national government science advisers and leading researchers in digital regulation and virtual worlds. It identifies key research areas, including scalability, real-time feedback, complexity modelling, cross-border collaboration, risk reduction, citizen participation, ethical considerations and the integration of emerging technologies. In addition, the paper analyses a set of experimental topics, spanning transportation, urban planning and the environment/climate crisis, that could be tested in citiverse platforms to advance regulatory learning in these areas. The proposed work is designed to inform future research for policy and emphasizes a responsible approach to developing and using citiverses. It prioritizes careful consideration of the ethical, economic, ecological and social dimensions of different regulations. The paper also explores essential preliminary steps necessary for integrating citiverses into the broader ecosystems of experimentation spaces, including test beds, living labs and regulatory sandboxes

[473] PISA: A Pragmatic Psych-Inspired Unified Memory System for Enhanced AI Agency

Shian Jia, Ziyang Huang, Xinbo Wang, Haofei Zhang, Mingli Song

Main category: cs.AI

TL;DR: PISA is a psych-inspired unified memory system for AI agents that treats memory as a constructive process, featuring trimodal adaptation and hybrid memory access to improve adaptability and knowledge retention.

DetailsMotivation: Existing AI memory systems lack adaptability to diverse tasks and overlook the constructive, task-oriented role of memory, drawing inspiration from Piaget's cognitive development theory.

Method: Proposes PISA with trimodal adaptation mechanism (schema updation, evolution, creation) and hybrid memory access architecture combining symbolic reasoning with neural retrieval.

Result: Empirical evaluation on LOCOMO benchmark and new AggQA benchmark shows PISA sets new state-of-the-art, significantly enhancing adaptability and long-term knowledge retention.

Conclusion: PISA successfully addresses limitations of existing memory systems by treating memory as constructive and adaptive process, demonstrating superior performance in adaptability and knowledge management.

Abstract: Memory systems are fundamental to AI agents, yet existing work often lacks adaptability to diverse tasks and overlooks the constructive and task-oriented role of AI agent memory. Drawing from Piaget’s theory of cognitive development, we propose PISA, a pragmatic, psych-inspired unified memory system that addresses these limitations by treating memory as a constructive and adaptive process. To enable continuous learning and adaptability, PISA introduces a trimodal adaptation mechanism (i.e., schema updation, schema evolution, and schema creation) that preserves coherent organization while supporting flexible memory updates. Building on these schema-grounded structures, we further design a hybrid memory access architecture that seamlessly integrates symbolic reasoning with neural retrieval, significantly improving retrieval accuracy and efficiency. Our empirical evaluation, conducted on the existing LOCOMO benchmark and our newly proposed AggQA benchmark for data analysis tasks, confirms that PISA sets a new state-of-the-art by significantly enhancing adaptability and long-term knowledge retention.

[474] Limits of Emergent Reasoning of Large Language Models in Agentic Frameworks for Deterministic Games

Chris Su, Harrison Li, Matheus Marques, George Flint, Kevin Zhu, Sunishchal Dev

Main category: cs.AI

TL;DR: LLMs with environment interfaces still experience performance collapse on Tower of Hanoi puzzles beyond certain complexity thresholds, showing mode-like collapse rather than true reasoning failure.

DetailsMotivation: To investigate whether LLM performance collapse on complex reasoning tasks is due to state tracking requirements or fundamental reasoning limitations.

Method: Provided LLMs with an environment interface for Tower of Hanoi problems, allowing tool calls for moves, written justifications, state observation, and self-reprompting.

Result: Environment access didn’t prevent performance collapse. Policy analysis showed increasing divergence from optimal and random policies, indicating mode-like collapse where performance depends on whether the model’s learned mode matches the correct solution.

Conclusion: LLMs exhibit mode-like collapse in complex reasoning, suggesting similar phenomena may occur in Large Reasoning Models, indicating fundamental limitations beyond state tracking issues.

Abstract: Recent work reports that Large Reasoning Models (LRMs) undergo a collapse in performance on solving puzzles beyond certain perplexity thresholds. In subsequent discourse, questions have arisen as to whether the nature of the task muddles an evaluation of true reasoning. One potential confound is the requirement that the model keep track of the state space on its own. We provide a large language model (LLM) with an environment interface for Tower of Hanoi problems, allowing it to make a move with a tool call, provide written justification, observe the resulting state space, and reprompt itself for the next move. We observe that access to an environment interface does not delay or eradicate performance collapse. Furthermore, LLM-parameterized policy analysis reveals increasing divergence from both optimal policies and uniformly random policies, suggesting that the model exhibits mode-like collapse at each level of complexity, and that performance is dependent upon whether the mode reflects the correct solution for the problem. We suggest that a similar phenomena might take place in LRMs.

[475] Cognitive Load Traces as Symbolic and Visual Accounts of Deep Model Cognition

Dong Liu, Yanxuan Yu

Main category: cs.AI

TL;DR: Cognitive Load Traces (CLTs) is a mid-level interpretability framework for deep models that quantifies model-internal resource allocation through symbolic, temporally varying functions representing Intrinsic, Extraneous, and Germane load components.

DetailsMotivation: Inspired by Cognitive Load Theory in human cognition, the paper aims to provide interpretable analysis of reasoning dynamics in deep models by tracking internal resource allocation patterns.

Method: CLTs are represented as a three-component stochastic process (IL_t, EL_t, GL_t) with measurable proxies including attention entropy, KV-cache miss ratio, representation dispersion, and decoding stability. The framework includes symbolic formulations and visualization methods like load curves and simplex diagrams.

Result: Experiments on reasoning and planning benchmarks demonstrate that CLTs can predict error-onset, reveal cognitive strategies, and enable load-guided interventions that improve reasoning efficiency by 15-30% while maintaining accuracy.

Conclusion: CLTs provide an effective interpretability framework for understanding and improving deep model reasoning by quantifying cognitive load patterns, enabling both analysis and optimization of model performance.

Abstract: We propose \textbf{Cognitive Load Traces} (CLTs) as a mid-level interpretability framework for deep models, inspired by Cognitive Load Theory in human cognition. CLTs are defined as symbolic, temporally varying functions that quantify model-internal resource allocation. Formally, we represent CLTs as a three-component stochastic process $(\mathrm{IL}_t, \mathrm{EL}_t, \mathrm{GL}_t)$, corresponding to \emph{Intrinsic}, \emph{Extraneous}, and \emph{Germane} load. Each component is instantiated through measurable proxies such as attention entropy, KV-cache miss ratio, representation dispersion, and decoding stability. We propose both symbolic formulations and visualization methods (load curves, simplex diagrams) that enable interpretable analysis of reasoning dynamics. Experiments on reasoning and planning benchmarks show that CLTs predict error-onset, reveal cognitive strategies, and enable load-guided interventions that improve reasoning efficiency by 15-30% while maintaining accuracy.

[476] ProofFlow: A Dependency Graph Approach to Faithful Proof Autoformalization

Rafael Cabral, Tuan Manh Do, Xuejun Yu, Wai Ming Tai, Zijin Feng, Xin Shen

Main category: cs.AI

TL;DR: ProofFlow is a novel autoformalization pipeline that preserves logical structure by mapping proof dependencies as a DAG and formalizing steps as intermediate lemmas, achieving state-of-the-art performance on a new benchmark.

DetailsMotivation: Current autoformalization approaches focus on producing executable code but frequently fail to preserve semantic meaning and logical structure of original human-written mathematical arguments.

Method: Constructs a directed acyclic graph (DAG) to map logical dependencies between proof steps, then employs a lemma-based approach to systematically formalize each step as an intermediate lemma.

Result: Achieves ProofScore of 0.545, substantially exceeding baselines like full-proof formalization (0.123) and step-proof formalization (0.072) on a new benchmark of 184 undergraduate-level problems.

Conclusion: ProofFlow sets new state-of-the-art for autoformalization by treating structural fidelity as primary objective, with pipeline, benchmark, and metric open-sourced for further progress.

Abstract: Proof autoformalization, the task of translating natural language theorems and proofs into machine-verifiable code, is a critical step for integrating large language models into rigorous mathematical workflows. Current approaches focus on producing executable code, but they frequently fail to preserve the semantic meaning and logical structure of the original human-written argument. To address this, we introduce ProofFlow, a novel pipeline that treats structural fidelity as a primary objective. ProofFlow first constructs a directed acyclic graph (DAG) to map the logical dependencies between proof steps. Then, it employs a novel lemma-based approach to systematically formalize each step as an intermediate lemma, preserving the logical structure of the original argument. To facilitate evaluation, we present a new benchmark of 184 undergraduate-level problems, manually annotated with step-by-step solutions and logical dependency graphs, and introduce ProofScore, a new composite metric to evaluate syntactic correctness, semantic faithfulness, and structural fidelity. Experimental results show our pipeline sets a new state-of-the-art for autoformalization, achieving a ProofScore of 0.545, substantially exceeding baselines like full-proof formalization (0.123), which processes the entire proof at once, and step-proof formalization (0.072), which handles each step independently. Our pipeline, benchmark, and score metric are open-sourced to encourage further progress at https://github.com/Huawei-AI4Math/ProofFlow.

[477] Ontologies in Motion: A BFO-Based Approach to Knowledge Graph Construction for Motor Performance Research Data in Sports Science

Sarah Rebecca Ondraszek, Jörg Waitelonis, Katja Keller, Claudia Niessner, Anna M. Jacyszyn, Harald Sack

Main category: cs.AI

TL;DR: The paper presents a vision for creating a knowledge graph from the MO|RE data repository to standardize and make motor performance research data machine-understandable.

DetailsMotivation: To enable better evaluation and comparison of physical and cognitive capabilities between populations by transforming how motor performance data are modeled and shared across studies.

Method: Developing a knowledge graph using an ontology rooted in the Basic Formal Ontology, focusing on formally representing the interrelation of plan specifications, specific processes, and related measurements.

Result: A proposed infrastructure for publishing and archiving research data in sports science that makes motor performance data standardized and machine-understandable.

Conclusion: The knowledge graph approach will transform motor performance data sharing and analysis, developed within the Leibniz Science Campus Digital Transformation of Research initiative.

Abstract: An essential component for evaluating and comparing physical and cognitive capabilities between populations is the testing of various factors related to human performance. As a core part of sports science research, testing motor performance enables the analysis of the physical health of different demographic groups and makes them comparable. The Motor Research (MO|RE) data repository, developed at the Karlsruhe Institute of Technology, is an infrastructure for publishing and archiving research data in sports science, particularly in the field of motor performance research. In this paper, we present our vision for creating a knowledge graph from MO|RE data. With an ontology rooted in the Basic Formal Ontology, our approach centers on formally representing the interrelation of plan specifications, specific processes, and related measurements. Our goal is to transform how motor performance data are modeled and shared across studies, making it standardized and machine-understandable. The idea presented here is developed within the Leibniz Science Campus ``Digital Transformation of Research’’ (DiTraRe).

[478] A Non-overlap-based Conflict Measure for Random Permutation Sets

Ruolan Cheng, Yong Deng, Enrique Herrera-Viedma

Main category: cs.AI

TL;DR: This paper proposes a new conflict measure method for random permutation sets (RPS) that analyzes conflicts from both random finite set and Dempster-Shafer theory perspectives, using an inconsistency measure inspired by rank-biased overlap.

DetailsMotivation: Measuring conflict between evidence represented by permutation mass functions is an urgent research topic in order-structured uncertain information fusion, as RPS provides a new formalism for reasoning with uncertainty involving order information.

Method: The authors define an inconsistency measure between permutations inspired by RBO, propose a non-overlap-based conflict measure for RPSs, and treat RPS theory as an extension of DST where order information indicates qualitative propensity with top-ranked elements being more critical.

Result: Numerical examples demonstrate the behavior and properties of the proposed conflict measure, which has natural top-weightedness property and can effectively measure conflict between RPSs from the DST view.

Conclusion: The proposed method provides decision-makers with flexible selection of weights, parameters, and truncated depths for measuring conflicts in order-structured uncertain information.

Abstract: Random permutation set (RPS) is a new formalism for reasoning with uncertainty involving order information. Measuring the conflict between two pieces of evidence represented by permutation mass functions remains an urgent research topic in order-structured uncertain information fusion. In this paper, a detailed analysis of conflicts in RPS is carried out from two different perspectives: random finite set (RFS) and Dempster-Shafer theory (DST). Starting from the observation of permutations, we first define an inconsistency measure between permutations inspired by the rank-biased overlap(RBO) measure and further propose a non-overlap-based conflict measure method for RPSs. This paper regards RPS theory (RPST) as an extension of DST. The order information newly added in focal sets indicates qualitative propensity, characterized by top-ranked elements occupying a more critical position. Some numerical examples are used to demonstrate the behavior and properties of the proposed conflict measure. The proposed method not only has the natural top-weightedness property and can effectively measure the conflict between RPSs from the DST view but also provides decision-makers with a flexible selection of weights, parameters, and truncated depths.

[479] End-to-end Listen, Look, Speak and Act

Siyin Wang, Wenyi Yu, Xianzhao Chen, Xiaohai Tian, Jun Zhang, Lu Lu, Chao Zhang

Main category: cs.AI

TL;DR: ELLSA is the first full-duplex, end-to-end model that simultaneously perceives and generates across vision, text, speech, and action within a single architecture, enabling more natural human-like interactions.

DetailsMotivation: Human interaction is inherently multimodal and full-duplex, with fluid adaptation to turn-taking and interruptions. Building models that can simulate these human capabilities is essential for creating more natural interactive AI systems.

Method: Uses a novel SA-MoE (Self-Attention Mixture-of-Experts) architecture that routes each modality to specialized experts and fuses them through a unified attention backbone, enabling joint multimodal perception and concurrent generation.

Result: Matches modality-specific baselines on speech-interaction and robot-manipulation benchmarks while uniquely supporting advanced multimodal behaviors like dialogue turn-taking, defective instruction rejection, speaking-while-acting, and action barge-ins.

Conclusion: ELLSA represents a step toward more natural and general interactive intelligence and contributes to the broader pursuit of artificial general intelligence.

Abstract: Human interaction is inherently multimodal and full-duplex: we listen while watching, speak while acting, and fluidly adapt to turn-taking and interruptions. Realizing these capabilities is essential for building models simulating humans. We present ELLSA (End-to-end Listen, Look, Speak and Act), which, to our knowledge, is the first full-duplex, end-to-end model that simultaneously perceives and generates across vision, text, speech, and action within a single architecture, enabling interaction patterns previously out of reach, yielding more natural, human-like behaviors. At its core is a novel SA-MoE architecture (Self-Attention Mixture-of-Experts) that routes each modality to specialized experts and fuses them through a unified attention backbone. This provides a generalizable solution for joint multimodal perception and concurrent generation, leveraging strong pre-trained components while enabling efficient modality integration and mitigating modality interference. On speech-interaction and robot-manipulation benchmarks, ELLSA matches modality-specific baselines, while uniquely supporting advanced multimodal and full-duplex behaviors such as dialogue and action turn-taking, defective instruction rejection, speaking-while-acting, context-grounded visual question answering, and action barge-ins. We contend that ELLSA represents a step toward more natural and general interactive intelligence, contributing to the broader pursuit of artificial general intelligence. All data, code and model checkpoints will be released upon acceptance.

[480] PAINT: Parallel-in-time Neural Twins for Dynamical System Reconstruction

Andreas Radler, Vincent Seyfried, Stefan Pirker, Johannes Brandstetter, Thomas Lichtenegger

Main category: cs.AI

TL;DR: PAINT introduces parallel-in-time neural twins for dynamical systems that stay on-trajectory by modeling state distributions over time windows, outperforming autoregressive models in maintaining accuracy.

DetailsMotivation: To create neural twins that can serve as digital replicas of real systems, consuming measurements at test time for context-specific decision-making while maintaining trajectory accuracy.

Method: PAINT trains a generative neural network to model state distributions parallel over time, using sliding window predictions from measurements at test time.

Result: PAINT stays on-trajectory and predicts system states from sparse measurements with high fidelity on turbulent fluid dynamics problems, unlike autoregressive models.

Conclusion: PAINT enables development of neural twins that remain on-trajectory, providing accurate state estimation and decision-making capabilities for dynamical systems.

Abstract: Neural surrogates have shown great potential in simulating dynamical systems, while offering real-time capabilities. We envision Neural Twins as a progression of neural surrogates, aiming to create digital replicas of real systems. A neural twin consumes measurements at test time to update its state, thereby enabling context-specific decision-making. A critical property of neural twins is their ability to remain on-trajectory, i.e., to stay close to the true system state over time. We introduce Parallel-in-time Neural Twins (PAINT), an architecture-agnostic family of methods for modeling dynamical systems from measurements. PAINT trains a generative neural network to model the distribution of states parallel over time. At test time, states are predicted from measurements in a sliding window fashion. Our theoretical analysis shows that PAINT is on-trajectory, whereas autoregressive models generally are not. Empirically, we evaluate our method on a challenging two-dimensional turbulent fluid dynamics problem. The results demonstrate that PAINT stays on-trajectory and predicts system states from sparse measurements with high fidelity. These findings underscore PAINT’s potential for developing neural twins that stay on-trajectory, enabling more accurate state estimation and decision-making.

[481] Global-focal Adaptation with Information Separation for Noise-robust Transfer Fault Diagnosis

Junyu Ren, Wensheng Gan, Guangyu Zhang, Wei Zhong, Philip S. Yu

Main category: cs.AI

TL;DR: ISGFAN is a robust framework for cross-domain fault diagnosis under noise conditions that uses information separation and global-focal adversarial learning to handle domain shifts and noise interference simultaneously.

DetailsMotivation: Existing transfer fault diagnosis methods fail in industrial environments with severe noise interference and domain shifts, requiring a more robust approach.

Method: Information separation architecture with adversarial learning and improved orthogonal loss to decouple domain-invariant fault representation, plus global-focal domain-adversarial scheme for both conditional and marginal distribution alignment.

Result: Outperforms other prominent existing approaches on three public benchmark datasets, demonstrating superior performance in cross-domain fault diagnosis under noise conditions.

Conclusion: ISGFAN provides an effective solution for robust cross-domain fault diagnosis in noisy industrial environments by successfully isolating noise interference and handling domain shifts.

Abstract: Existing transfer fault diagnosis methods typically assume either clean data or sufficient domain similarity, which limits their effectiveness in industrial environments where severe noise interference and domain shifts coexist. To address this challenge, we propose an information separation global-focal adversarial network (ISGFAN), a robust framework for cross-domain fault diagnosis under noise conditions. ISGFAN is built on an information separation architecture that integrates adversarial learning with an improved orthogonal loss to decouple domain-invariant fault representation, thereby isolating noise interference and domain-specific characteristics. To further strengthen transfer robustness, ISGFAN employs a global-focal domain-adversarial scheme that constrains both the conditional and marginal distributions of the model. Specifically, the focal domain-adversarial component mitigates category-specific transfer obstacles caused by noise in unsupervised scenarios, while the global domain classifier ensures alignment of the overall distribution. Experiments conducted on three public benchmark datasets demonstrate that the proposed method outperforms other prominent existing approaches, confirming the superiority of the ISGFAN framework. Data and code are available at https://github.com/JYREN-Source/ISGFAN

[482] Ripple Effect Protocol: Coordinating Agent Populations

Ayush Chopra, Aman Sharma, Feroz Ahmad, Luca Muscariello, Vijoy Pandey, Ramesh Raskar

Main category: cs.AI

TL;DR: The Ripple Effect Protocol (REP) improves multi-agent coordination by having agents share lightweight sensitivity signals about how their decisions would change with environmental shifts, enabling faster and more stable group alignment than traditional communication protocols.

DetailsMotivation: Current AI agent communication protocols like A2A and ACP focus on communication but lack coordination capabilities, leading to poor collective outcomes despite individually smart agents, especially as agent populations grow.

Method: REP introduces a coordination protocol where agents share decisions plus sensitivity signals expressing how choices would change with environmental variables. These sensitivities ripple through local networks, with formal protocol specifications separating required message schemas from optional aggregation rules.

Result: REP improves coordination accuracy and efficiency by 41-100% over A2A across three domains: supply chain cascades (Beer Game), preference aggregation in sparse networks (Movie Scheduling), and sustainable resource allocation (Fishbanks). It also flexibly handles multimodal sensitivity signals from LLMs.

Conclusion: By making coordination a protocol-level capability, REP provides scalable infrastructure for the emerging Internet of Agents, enabling better collective outcomes through sensitivity-based coordination.

Abstract: Modern AI agents can exchange messages using protocols such as A2A and ACP, yet these mechanisms emphasize communication over coordination. As agent populations grow, this limitation produces brittle collective behavior, where individually smart agents converge on poor group outcomes. We introduce the Ripple Effect Protocol (REP), a coordination protocol in which agents share not only their decisions but also lightweight sensitivities - signals expressing how their choices would change if key environmental variables shifted. These sensitivities ripple through local networks, enabling groups to align faster and more stably than with agent-centric communication alone. We formalize REP’s protocol specification, separating required message schemas from optional aggregation rules, and evaluate it across scenarios with varying incentives and network topologies. Benchmarks across three domains: (i) supply chain cascades (Beer Game), (ii) preference aggregation in sparse networks (Movie Scheduling), and (iii) sustainable resource allocation (Fishbanks) show that REP improves coordination accuracy and efficiency over A2A by 41 to 100%, while flexibly handling multimodal sensitivity signals from LLMs. By making coordination a protocol-level capability, REP provides scalable infrastructure for the emerging Internet of Agents

[483] Algorithms for dynamic scheduling in manufacturing, towards digital factories Improving Deadline Feasibility and Responsiveness via Temporal Networks

Ioan Hedea

Main category: cs.AI

TL;DR: This paper combines offline constraint programming with online temporal-network execution to create robust manufacturing schedules that remain feasible under worst-case uncertainty, eliminating deadline violations with minimal makespan overhead.

DetailsMotivation: Traditional deterministic schedules fail when reality deviates from plans due to stochastic task durations from process noise, equipment variability, and human intervention, causing costly last-minute repairs.

Method: Builds a CP model of flexible job-shop with deadlines, inserts optimal buffer Δ*, translates plan to Simple Temporal Network with Uncertainty (STNU), and verifies dynamic controllability for real-time dispatching.

Result: Eliminates 100% of deadline violations in Kacem 1-4 benchmark suite with only 3-5% makespan overhead. CP solve-times and STNU checks remain sub-second on medium-size instances.

Conclusion: Temporal-network reasoning bridges proactive buffering and dynamic robustness, advancing toward truly digital, self-correcting factories.

Abstract: Modern manufacturing systems must meet hard delivery deadlines while coping with stochastic task durations caused by process noise, equipment variability, and human intervention. Traditional deterministic schedules break down when reality deviates from nominal plans, triggering costly last-minute repairs. This thesis combines offline constraint-programming (CP) optimisation with online temporal-network execution to create schedules that remain feasible under worst-case uncertainty. First, we build a CP model of the flexible job-shop with per-job deadline tasks and insert an optimal buffer $\Delta^*$ to obtain a fully pro-active baseline. We then translate the resulting plan into a Simple Temporal Network with Uncertainty (STNU) and verify dynamic controllability, which guarantees that a real-time dispatcher can retime activities for every bounded duration realisation without violating resource or deadline constraints. Extensive Monte-Carlo simulations on the open Kacem~1–4 benchmark suite show that our hybrid approach eliminates 100% of deadline violations observed in state-of-the-art meta-heuristic schedules, while adding only 3–5% makespan overhead. Scalability experiments confirm that CP solve-times and STNU checks remain sub-second on medium-size instances. The work demonstrates how temporal-network reasoning can bridge the gap between proactive buffering and dynamic robustness, moving industry a step closer to truly digital, self-correcting factories.

[484] Reliability of Large Language Model Generated Clinical Reasoning in Assisted Reproductive Technology: Blinded Comparative Evaluation Study

Dou Liu, Ying Long, Sophia Zuoqiu, Di Liu, Kang Li, Yiting Lin, Hanyi Liu, Rong Yin, Tian Tang

Main category: cs.AI

TL;DR: This study evaluates LLM-generated clinical Chains-of-Thought (CoTs) and finds that Selective Few-shot prompting with diverse, high-quality examples significantly outperforms other strategies, while AI evaluators fail to detect these critical differences.

DetailsMotivation: To address data scarcity in clinical AI by creating reliable synthetic medical data through LLMs, while ensuring clinical reliability through expert validation.

Method: Blinded comparative study where senior clinicians evaluated CoTs generated via three prompting strategies: Zero-shot, Random Few-shot, and Selective Few-shot, compared against GPT-4o evaluations.

Result: Selective Few-shot strategy significantly outperformed other strategies across all human evaluation metrics (p < .001). Random Few-shot offered no improvement over Zero-shot. AI evaluator failed to discern performance differences.

Conclusion: Clinical reliability of synthetic CoTs depends on strategic prompt curation using “Gold-Standard Depth” and “Representative Diversity” principles, not just examples. Human expertise remains indispensable for high-stakes clinical AI evaluation.

Abstract: Creating high-quality clinical Chains-of-Thought (CoTs) is crucial for explainable medical Artificial Intelligence (AI) while constrained by data scarcity. Although Large Language Models (LLMs) can synthesize medical data, their clinical reliability remains unverified. This study evaluates the reliability of LLM-generated CoTs and investigates prompting strategies to enhance their quality. In a blinded comparative study, senior clinicians in Assisted Reproductive Technology (ART) evaluated CoTs generated via three distinct strategies: Zero-shot, Random Few-shot (using shallow examples), and Selective Few-shot (using diverse, high-quality examples). These expert ratings were compared against evaluations from a state-of-the-art AI model (GPT-4o). The Selective Few-shot strategy significantly outperformed other strategies across all human evaluation metrics (p < .001). Critically, the Random Few-shot strategy offered no significant improvement over the Zero-shot baseline, demonstrating that low-quality examples are as ineffective as no examples. The success of the Selective strategy is attributed to two principles: “Gold-Standard Depth” (reasoning quality) and “Representative Diversity” (generalization). Notably, the AI evaluator failed to discern these critical performance differences. The clinical reliability of synthetic CoTs is dictated by strategic prompt curation, not the mere presence of examples. We propose a “Dual Principles” framework as a foundational methodology to generate trustworthy data at scale. This work offers a validated solution to the data bottleneck and confirms the indispensable role of human expertise in evaluating high-stakes clinical AI.

[485] Surrogate Modeling and Explainable Artificial Intelligence for Complex Systems: A Workflow for Automated Simulation Exploration

Paul Saves, Pramudita Satria Palar, Muhammad Daffa Robani, Nicolas Verstaevel, Moncef Garouani, Julien Aligon, Benoit Gaudou, Koji Shimoyama, Joseph Morlier

Main category: cs.AI

TL;DR: A workflow using lightweight emulators trained on compact designs of experiments to address computational cost and transparency issues in simulation-driven engineering workflows, enabling fast approximations, uncertainty quantification, and explainable AI analyses.

DetailsMotivation: To overcome high computational costs and limited transparency in simulation-driven engineering workflows that rely on expensive simulator runs and opaque blackbox components.

Method: Training lightweight emulators on compact designs of experiments that provide fast approximations of expensive simulators, enable uncertainty quantification, and support global and local Explainable AI (XAI) analyses.

Result: The approach enables large-scale exploration in seconds, uncovers nonlinear interactions and emergent behaviors, identifies key design and policy levers, and signals regions where surrogates require more data or alternative architectures.

Conclusion: The surrogate model and XAI coupling provides an effective workflow for simulation-based complex-system analysis, unifying tools from engineering design to agent-based models for socio-environmental understanding.

Abstract: Complex systems are increasingly explored through simulation-driven engineering workflows that combine physics-based and empirical models with optimization and analytics. Despite their power, these workflows face two central obstacles: (1) high computational cost, since accurate exploration requires many expensive simulator runs; and (2) limited transparency and reliability when decisions rely on opaque blackbox components. We propose a workflow that addresses both challenges by training lightweight emulators on compact designs of experiments that (i) provide fast, low-latency approximations of expensive simulators, (ii) enable rigorous uncertainty quantification, and (iii) are adapted for global and local Explainable Artificial Intelligence (XAI) analyses. This workflow unifies every simulation-based complex-system analysis tool, ranging from engineering design to agent-based models for socio-environmental understanding. In this paper, we proposea comparative methodology and practical recommendations for using surrogate-based explainability tools within the proposed workflow. The methodology supports continuous and categorical inputs, combines global-effect and uncertainty analyses with local attribution, and evaluates the consistency of explanations across surrogate models, thereby diagnosing surrogate adequacy and guiding further data collection or model refinement. We demonstrate the approach on two contrasting case studies: a multidisciplinary design analysis of a hybrid-electric aircraft and an agent-based model of urban segregation. Results show that the surrogate model and XAI coupling enables large-scale exploration in seconds, uncovers nonlinear interactions and emergent behaviors, identifies key design and policy levers, and signals regions where surrogates require more data or alternative architectures.

Elija Perrier

Main category: cs.AI

TL;DR: The paper proposes a new framework to redefine corporate knowledge as a measurable capability in the age of AI, using computational efficiency and validated reliability to create auditable metrics for legal accountability.

DetailsMotivation: Traditional corporate responsibility based on human agents is challenged by AI-mediated decision-making, requiring new ways to measure corporate knowledge and intent.

Method: Develops a formal model with continuous organizational knowledge metric S_S(φ) that integrates computational cost and validated error rates, plus thresholded knowledge predicate K_S and firm-wide epistemic capacity index K_{S,t}.

Result: Creates quantitative metrics that can be mapped onto legal standards (actual knowledge, constructive knowledge, wilful blindness, recklessness) to make corporate knowledge tractable and accountable.

Conclusion: Provides a pathway for creating measurable and justiciable audit artifacts that render the corporate mind accountable in the algorithmic age.

Abstract: Corporate responsibility turns on notions of corporate \textit{mens rea}, traditionally imputed from human agents. Yet these assumptions are under challenge as generative AI increasingly mediates enterprise decision-making. Building on the theory of extended cognition, we argue that in response corporate knowledge may be redefined as a dynamic capability, measurable by the efficiency of its information-access procedures and the validated reliability of their outputs. We develop a formal model that captures epistemic states of corporations deploying sophisticated AI or information systems, introducing a continuous organisational knowledge metric $S_S(\varphi)$ which integrates a pipeline’s computational cost and its statistically validated error rate. We derive a thresholded knowledge predicate $\mathsf{K}S$ to impute knowledge and a firm-wide epistemic capacity index $\mathcal{K}{S,t}$ to measure overall capability. We then operationally map these quantitative metrics onto the legal standards of actual knowledge, constructive knowledge, wilful blindness, and recklessness. Our work provides a pathway towards creating measurable and justiciable audit artefacts, that render the corporate mind tractable and accountable in the algorithmic age.

[487] Towards Automatic Evaluation and Selection of PHI De-identification Models via Multi-Agent Collaboration

Guanchen Wu, Zuhui Chen, Yuzhang Xie, Carl Yang

Main category: cs.AI

TL;DR: TEAM-PHI is a multi-agent framework using LLMs to automatically evaluate PHI de-identification models without relying on expensive expert annotations, achieving rankings that match supervised evaluation.

DetailsMotivation: Current PHI de-identification evaluation depends on costly, small-scale expert annotations, limiting scalability and practical deployment of models.

Method: Uses multiple Evaluation Agents (LLMs) to independently judge PHI extraction correctness, then consolidates results through LLM-based majority voting to produce stable rankings.

Result: Experiments show TEAM-PHI produces consistent and accurate rankings that closely match ground-truth annotations and human evaluation, despite individual evaluator variation.

Conclusion: TEAM-PHI provides a practical, secure, and cost-effective solution for automatic PHI de-identification evaluation and model selection when ground-truth labels are limited.

Abstract: Protected health information (PHI) de-identification is critical for enabling the safe reuse of clinical notes, yet evaluating and comparing PHI de-identification models typically depends on costly, small-scale expert annotations. We present TEAM-PHI, a multi-agent evaluation and selection framework that uses large language models (LLMs) to automatically measure de-identification quality and select the best-performing model without heavy reliance on gold labels. TEAM-PHI deploys multiple Evaluation Agents, each independently judging the correctness of PHI extractions and outputting structured metrics. Their results are then consolidated through an LLM-based majority voting mechanism that integrates diverse evaluator perspectives into a single, stable, and reproducible ranking. Experiments on a real-world clinical note corpus demonstrate that TEAM-PHI produces consistent and accurate rankings: despite variation across individual evaluators, LLM-based voting reliably converges on the same top-performing systems. Further comparison with ground-truth annotations and human evaluation confirms that the framework’s automated rankings closely match supervised evaluation. By combining independent evaluation agents with LLM majority voting, TEAM-PHI offers a practical, secure, and cost-effective solution for automatic evaluation and best-model selection in PHI de-identification, even when ground-truth labels are limited.

[488] The Right to Be Remembered: Preserving Maximally Truthful Digital Memory in the Age of AI

Alex Zhavoronkov, Dominika Wilczok, Roman Yampolskiy

Main category: cs.AI

TL;DR: The paper proposes a ‘Right To Be Remembered’ (RTBR) framework to address AI-driven information omission risks in LLMs, which can disproportionately suppress certain narratives while amplifying others, potentially reshaping collective memory.

DetailsMotivation: LLMs provide synthesized responses that feel authoritative but may collapse multiple perspectives into single answers, concentrating information power in few vendors and creating risks of bias, omission, and erasure of those with limited digital presence.

Method: The paper presents the concept of Right To Be Remembered (RTBR) as a framework to minimize AI-driven information omission, ensure fair treatment, and maximize truthfulness in generated content.

Result: The RTBR concept is proposed as a solution to address the concentration of information power in LLM vendors and the potential reshaping of collective memory through disproportionate suppression or elevation of certain narratives.

Conclusion: A Right To Be Remembered framework is needed to counter the threats posed by LLMs in potentially erasing marginalized voices and reshaping collective memory through biased information synthesis.

Abstract: Since the rapid expansion of large language models (LLMs), people have begun to rely on them for information retrieval. While traditional search engines display ranked lists of sources shaped by search engine optimization (SEO), advertising, and personalization, LLMs typically provide a synthesized response that feels singular and authoritative. While both approaches carry risks of bias and omission, LLMs may amplify the effect by collapsing multiple perspectives into one answer, reducing users ability or inclination to compare alternatives. This concentrates power over information in a few LLM vendors whose systems effectively shape what is remembered and what is overlooked. As a result, certain narratives, individuals or groups, may be disproportionately suppressed, while others are disproportionately elevated. Over time, this creates a new threat: the gradual erasure of those with limited digital presence, and the amplification of those already prominent, reshaping collective memory.To address these concerns, this paper presents a concept of the Right To Be Remembered (RTBR) which encompasses minimizing the risk of AI-driven information omission, embracing the right of fair treatment, while ensuring that the generated content would be maximally truthful.

[489] Graph Attention-Guided Search for Dense Multi-Agent Pathfinding

Rishabh Jain, Keisuke Okumura, Michael Amir, Amanda Prorok

Main category: cs.AI

TL;DR: LaGAT: A hybrid framework combining learned neural heuristics from MAGAT with search-based LaCAM algorithm for dense multi-agent pathfinding, outperforming both pure search and pure learning methods.

DetailsMotivation: Real-time dense multi-agent pathfinding remains challenging for state-of-the-art planners, requiring better solutions that integrate learning and search.

Method: Integrates learned heuristic from enhanced MAGAT neural policy into LaCAM search algorithm, using pre-train-then-fine-tune strategy and deadlock detection for imperfect neural guidance.

Result: Outperforms both purely search-based and purely learning-based methods in dense scenarios.

Conclusion: Carefully designed hybrid search offers powerful solution for challenging multi-agent coordination problems.

Abstract: Finding near-optimal solutions for dense multi-agent pathfinding (MAPF) problems in real-time remains challenging even for state-of-the-art planners. To this end, we develop a hybrid framework that integrates a learned heuristic derived from MAGAT, a neural MAPF policy with a graph attention scheme, into a leading search-based algorithm, LaCAM. While prior work has explored learning-guided search in MAPF, such methods have historically underperformed. In contrast, our approach, termed LaGAT, outperforms both purely search-based and purely learning-based methods in dense scenarios. This is achieved through an enhanced MAGAT architecture, a pre-train-then-fine-tune strategy on maps of interest, and a deadlock detection scheme to account for imperfect neural guidance. Our results demonstrate that, when carefully designed, hybrid search offers a powerful solution for tightly coupled, challenging multi-agent coordination problems.

[490] ScholarEval: Research Idea Evaluation Grounded in Literature

Hanane Nour Moussa, Patrick Queiroz Da Silva, Daniel Adu-Ampratwum, Alyson East, Zitong Lu, Nikki Puccetti, Mingyi Xue, Huan Sun, Bodhisattwa Prasad Majumder, Sachin Kumar

Main category: cs.AI

TL;DR: ScholarEval is a retrieval-augmented framework for evaluating AI-generated research ideas using soundness and contribution criteria, outperforming baselines including OpenAI’s o4-mini-deep-research.

DetailsMotivation: As AI tools become common for research ideation, robust evaluation is needed to ensure validity and usefulness of generated ideas.

Method: Introduces ScholarEval framework with soundness (empirical validity) and contribution (advancement) criteria, evaluated on ScholarIdeas dataset of 117 expert-annotated research ideas across four disciplines.

Result: ScholarEval achieves higher coverage of expert rubric points, is preferred over OpenAI’s o4-mini-deep-research in evaluation quality, and outperforms in literature engagement, idea refinement, and usefulness in user studies.

Conclusion: ScholarEval provides effective evaluation of research ideas and is released openly for community use and development.

Abstract: As AI tools become increasingly common for research ideation, robust evaluation is critical to ensure the validity and usefulness of generated ideas. We introduce ScholarEval, a retrieval augmented evaluation framework that assesses research ideas based on two fundamental criteria: soundness - the empirical validity of proposed methods based on existing literature, and contribution - the degree of advancement made by the idea across different dimensions relative to prior research. To evaluate ScholarEval, we introduce ScholarIdeas, the first expert-annotated dataset of multi-domain research ideas and reviews, comprised of 117 ideas across four disciplines: artificial intelligence, neuroscience, biochemistry, and ecology. Our evaluation shows that ScholarEval achieves significantly higher coverage of points mentioned in the human expert annotated rubrics in ScholarIdeas compared to all baselines. Furthermore, ScholarEval is consistently preferred over our strongest baseline o4-mini-deep-research, a reasoning and search-enabled agentic system by OpenAI, in terms of evaluation actionability, depth, and evidence support. Our large-scale user study also shows that ScholarEval significantly outperforms deep research in literature engagement, idea refinement, and usefulness. We openly release our code, dataset, and ScholarEval tool for the community to use and build on.

[491] Diverse Planning with Simulators via Linear Temporal Logic

Mustafa F. Abdelwahed, Alice Toniolo, Joan Espasa, Ian P. Gent

Main category: cs.AI

TL;DR: FBI_LTL is a diverse planner for simulation-based planning that uses Linear Temporal Logic to generate semantically diverse plans, addressing limitations of traditional planners that produce syntactically different but semantically identical solutions.

DetailsMotivation: Traditional planners often generate single plans that may not satisfy agent preferences, and existing diverse planning approaches can produce plans that are syntactically different but semantically identical, failing to provide meaningful alternatives.

Method: FBI_LTL integrates LTL-based diversity models directly into the search process to define semantic diversity criteria, enabling agents to specify what constitutes meaningfully different plans in simulation-based environments.

Result: Extensive evaluations show FBI_LTL consistently generates more diverse plans compared to baseline approaches across various benchmarks.

Conclusion: This work establishes the feasibility of semantically-guided diverse planning in simulation-based environments, enabling innovative approaches in realistic, non-symbolic domains where traditional model-based approaches fail.

Abstract: Autonomous agents rely on automated planning algorithms to achieve their objectives. Simulation-based planning offers a significant advantage over declarative models in modelling complex environments. However, relying solely on a planner that produces a single plan may not be practical, as the generated plans may not always satisfy the agent’s preferences. To address this limitation, we introduce $\texttt{FBI}\texttt{LTL}$, a diverse planner explicitly designed for simulation-based planning problems. $\texttt{FBI}\texttt{LTL}$ utilises Linear Temporal Logic (LTL) to define semantic diversity criteria, enabling agents to specify what constitutes meaningfully different plans. By integrating these LTL-based diversity models directly into the search process, $\texttt{FBI}\texttt{LTL}$ ensures the generation of semantically diverse plans, addressing a critical limitation of existing diverse planning approaches that may produce syntactically different but semantically identical solutions. Extensive evaluations on various benchmarks consistently demonstrate that $\texttt{FBI}\texttt{LTL}$ generates more diverse plans compared to a baseline approach. This work establishes the feasibility of semantically-guided diverse planning in simulation-based environments, paving the way for innovative approaches in realistic, non-symbolic domains where traditional model-based approaches fail.

[492] Distractor Injection Attacks on Large Reasoning Models: Characterization and Defense

Zhehao Zhang, Weijie Xu, Shixian Cui, Chandan K. Reddy

Main category: cs.AI

TL;DR: LRMs are vulnerable to reasoning distraction attacks where irrelevant complex tasks in prompts divert models from their primary objectives, reducing accuracy by up to 60%. The paper proposes a training-based defense using SFT and RL on synthetic adversarial data.

DetailsMotivation: To identify and systematically analyze a critical vulnerability in large reasoning models where they can be diverted from their primary objectives by maliciously embedded irrelevant complex tasks in prompts.

Method: Comprehensive study across diverse models and benchmarks, revealing susceptibility to reasoning distraction. Proposed training-based defense combining Supervised Fine-Tuning (SFT) and Reinforcement Learning (RL) on synthetic adversarial data.

Result: State-of-the-art LRMs are highly susceptible to reasoning distraction, with injected distractors reducing task accuracy by up to 60%. Certain alignment techniques can amplify this weakness. The proposed defense improves robustness by over 50 points on challenging distractor attacks.

Conclusion: Reasoning distraction represents a distinct and urgent threat to LRM reliability. The proposed training-based defense provides a practical step toward safer and more trustworthy reasoning systems.

Abstract: Recent advances in large reasoning models (LRMs) have enabled remarkable performance on complex tasks such as mathematics and coding by generating long Chain-of-Thought (CoT) traces. In this paper, we identify and systematically analyze a critical vulnerability we term reasoning distraction, where LRMs are diverted from their primary objective by irrelevant yet complex tasks maliciously embedded in the prompt. Through a comprehensive study across diverse models and benchmarks, we show that even state-of-the-art LRMs are highly susceptible, with injected distractors reducing task accuracy by up to 60%. We further reveal that certain alignment techniques can amplify this weakness and that models may exhibit covert compliance, following hidden adversarial instructions in reasoning while concealing them in the final output. To mitigate these risks, we propose a training-based defense that combines Supervised Fine-Tuning (SFT) and Reinforcement Learning (RL) on synthetic adversarial data, improving robustness by over 50 points on challenging distractor attacks. Our findings establish reasoning distraction as a distinct and urgent threat to LRM reliability and provide a practical step toward safer and more trustworthy reasoning systems.

[493] What Limits Agentic Systems Efficiency?

Song Bian, Minghao Yan, Anand Jayarajan, Gennady Pekhimenko, Shivaram Venkataraman

Main category: cs.AI

TL;DR: This paper identifies efficiency bottlenecks in web-interactive agentic systems, showing web environment latency can contribute up to 53.7% of total latency, and proposes SpecCache with speculative execution to reduce overhead by up to 3.2x.

DetailsMotivation: Existing research on LLM-based agentic systems focuses mainly on reasoning performance while neglecting efficiency, particularly the latency issues caused by web interactions in systems like Deep Research.

Method: The authors conduct an empirical study across 15 models and 5 providers to analyze latency components, then propose SpecCache - a caching framework augmented with speculative execution to reduce web environment overhead.

Result: Web environment latency contributes up to 53.7% to overall system latency. SpecCache improves cache hit rate by 58x compared to random caching and reduces web environment overhead by 3.2x without degrading performance.

Conclusion: Efficiency is a critical bottleneck in web-interactive agentic systems, and the proposed SpecCache framework effectively addresses web environment latency issues while maintaining system performance.

Abstract: Large Language Models (LLMs), such as OpenAI-o1 and DeepSeek-R1, have demonstrated strong reasoning capabilities. To further enhance LLM capabilities, recent agentic systems, such as Deep Research, incorporate web interactions into LLM reasoning to mitigate uncertainties and reduce potential errors. However, existing research predominantly focuses on reasoning performance, often neglecting the efficiency of agentic systems. In this work, we present a comprehensive empirical study that identifies efficiency bottlenecks in web-interactive agentic systems. We decompose end-to-end latency into two primary components: LLM API latency and web environment latency. We conduct a comprehensive empirical study across 15 models and 5 providers to demonstrate high variability in API-based agentic systems. We observe that web environment latency can contribute as much as 53.7% to the overall latency in a web-based agentic system. To improve latency, we propose SpecCache, a caching framework augmented with speculative execution that can reduce web environment overhead. Extensive evaluations on two standard benchmarks show that our approach improves the cache hit rate by up to 58x compared to a random caching strategy, while reducing web environment overhead by up to 3.2x, without degrading agentic system performance.

[494] A Principle of Targeted Intervention for Multi-Agent Reinforcement Learning

Anjie Liu, Jianhong Wang, Samuel Kaski, Jun Wang, Mengyue Yang

Main category: cs.AI

TL;DR: The paper proposes using multi-agent influence diagrams (MAIDs) to address challenges in cooperative MARL, introducing targeted intervention through Pre-Strategy Intervention (PSI) to achieve desired outcomes without global guidance.

DetailsMotivation: Steering cooperative MARL towards desired outcomes is challenging when global human guidance is impractical, and current coordination mechanisms lack easy-to-use research tools.

Method: Uses MAIDs as graphical framework, introduces targeted intervention paradigm applied to single agents, implements PSI causal inference technique to maximize causal effects for composite outcomes.

Result: Demonstrates effectiveness of targeted intervention and verifies relevance graph analysis results in experiments.

Conclusion: MAIDs provide a practical framework for analyzing MARL approaches and implementing targeted interventions to achieve desired outcomes without requiring global guidance.

Abstract: Steering cooperative multi-agent reinforcement learning (MARL) towards desired outcomes is challenging, particularly when the global guidance from a human on the whole multi-agent system is impractical in a large-scale MARL. On the other hand, designing mechanisms to coordinate agents most relies on empirical studies, lacking a easy-to-use research tool. In this work, we employ multi-agent influence diagrams (MAIDs) as a graphical framework to address the above issues. First, we introduce interaction paradigms that leverage MAIDs to analyze and visualize existing approaches in MARL. Then, we design a new interaction paradigm based on MAIDs, referred to as targeted intervention that is applied to only a single targeted agent, so the problem of global guidance can be mitigated. In our implementation, we introduce a causal inference technique-referred to as Pre-Strategy Intervention (PSI)-to realize the targeted intervention paradigm. Since MAIDs can be regarded as a special class of causal diagrams, a composite desired outcome that integrates the primary task goal and an additional desired outcome can be achieved by maximizing the corresponding causal effect through the PSI. Moreover, the bundled relevance graph analysis of MAIDs provides a tool to identify whether an MARL learning paradigm is workable under the design of an interaction paradigm. In experiments, we demonstrate the effectiveness of our proposed targeted intervention, and verify the result of relevance graph analysis.

[495] DTKG: Dual-Track Knowledge Graph-Verified Reasoning Framework for Multi-Hop QA

Changhao Wang, Yanfang Liu, Xinxin Fan, Anzhi Zhou, Lao Tian, Yunfeng Lu

Main category: cs.AI

TL;DR: DTKG is a dual-track framework for multi-hop QA that combines LLM-based fact verification and KG path-based reasoning to handle both parallel verification and chained reasoning tasks efficiently.

DetailsMotivation: Current multi-hop reasoning approaches either use LLM-based fact verification (good for parallel tasks) or KG path-based chains (good for chained reasoning), but neither handles both types well, leading to efficiency and accuracy issues.

Method: DTKG uses a dual-track approach inspired by Dual Process Theory, with two main stages: Classification Stage to identify reasoning type, and Branch Processing Stage that applies appropriate techniques for each type.

Result: The framework addresses limitations of existing approaches by combining both verification methods to handle different multi-hop reasoning patterns more effectively.

Conclusion: DTKG provides a more comprehensive solution for multi-hop QA by leveraging complementary strengths of both LLM-based verification and KG path-based reasoning through its dual-track design.

Abstract: Multi-hop reasoning for question answering (QA) plays a critical role in retrieval-augmented generation (RAG) for modern large language models (LLMs). The accurate answer can be obtained through retrieving relational structure of entities from knowledge graph (KG). Regarding the inherent relation-dependency and reasoning pattern, multi-hop reasoning can be in general classified into two categories: i) parallel fact-verification multi-hop reasoning question, i.e., requiring simultaneous verifications of multiple independent sub-questions; and ii) chained multi-hop reasoning questions, i.e., demanding sequential multi-step inference with intermediate conclusions serving as essential premises for subsequent reasoning. Currently, the multi-hop reasoning approaches singly employ one of two techniques: LLM response-based fact verification and KG path-based chain construction. Nevertheless, the former excels at parallel fact-verification but underperforms on chained reasoning tasks, while the latter demonstrates proficiency in chained multi-hop reasoning but suffers from redundant path retrieval when handling parallel fact-verification reasoning. These limitations deteriorate the efficiency and accuracy for multi-hop QA tasks. To address this challenge, we propose a novel dual-track KG verification and reasoning framework DTKG, which is inspired by the Dual Process Theory in cognitive science. Specifically, DTKG comprises two main stages: the Classification Stage and the Branch Processing Stage.

[496] MedRule-KG: A Knowledge-Graph–Steered Scaffold for Mathematical Reasoning with a Lightweight Verifier

Crystal Su

Main category: cs.AI

TL;DR: MedRule-KG is a knowledge graph with symbolic verification that improves LLM reasoning by enforcing mathematical constraints, achieving perfect accuracy on FDA benchmark.

DetailsMotivation: LLMs often produce fluent but mathematically incorrect reasoning, violating basic constraints despite appearing logical.

Method: Create MedRule-KG knowledge graph with entities, relations, and domain rules, plus symbolic verifier to check predictions and apply corrections.

Result: Improved exact match from 0.767 to 0.900 with grounding, and 1.000 EM with verifier, eliminating all rule violations.

Conclusion: MedRule-KG provides effective scaffold for safe mathematical reasoning, with released code and data for reproducibility.

Abstract: Large language models (LLMs) often produce fluent reasoning steps while violating simple mathematical or logical constraints. We introduce MedRule-KG, a compact typed knowledge graph coupled with a symbolic verifier, designed to enforce mathematically interpretable rules in reasoning tasks. MedRule-KG encodes entities, relations, and three domain-inspired rules, while the verifier checks predictions and applies minimal corrections to guarantee consistency. On a 90-example FDA-derived benchmark, grounding in MedRule-KG improves exact match (EM) from 0.767 to 0.900, and adding the verifier yields 1.000 EM while eliminating rule violations entirely. We demonstrate how MedRule-KG provides a general scaffold for safe mathematical reasoning, discuss ablations, and release code and data to encourage reproducibility.

[497] Beyond Fixed Anchors: Precisely Erasing Concepts with Sibling Exclusive Counterparts

Tong Zhang, Ru Zhang, Jianyi Liu, Zhen Yang, Gongshen Liu

Main category: cs.AI

TL;DR: SELECT is a dynamic anchor selection framework for concept erasure in text-to-image diffusion models that overcomes limitations of fixed anchor strategies by using sibling exclusive concepts and a two-stage evaluation mechanism.

DetailsMotivation: Existing concept erasure methods rely on fixed anchor strategies, which cause issues like concept re-emergence and erosion. The authors identified that erasure is inherently sensitive to anchor selection through causal tracing.

Method: Proposed SELECT framework with a novel two-stage evaluation mechanism that automatically discovers optimal anchors for precise erasure while identifying critical boundary anchors to preserve related concepts. Uses sibling exclusive concepts as superior anchors.

Result: SELECT consistently outperforms existing baselines across key performance metrics, efficiently adapts to multiple erasure frameworks, and averages only 4 seconds for anchor mining of a single concept.

Conclusion: SELECT serves as a universal anchor solution that effectively addresses the limitations of fixed anchor strategies in concept erasure for text-to-image diffusion models.

Abstract: Existing concept erasure methods for text-to-image diffusion models commonly rely on fixed anchor strategies, which often lead to critical issues such as concept re-emergence and erosion. To address this, we conduct causal tracing to reveal the inherent sensitivity of erasure to anchor selection and define Sibling Exclusive Concepts as a superior class of anchors. Based on this insight, we propose \textbf{SELECT} (Sibling-Exclusive Evaluation for Contextual Targeting), a dynamic anchor selection framework designed to overcome the limitations of fixed anchors. Our framework introduces a novel two-stage evaluation mechanism that automatically discovers optimal anchors for precise erasure while identifying critical boundary anchors to preserve related concepts. Extensive evaluations demonstrate that SELECT, as a universal anchor solution, not only efficiently adapts to multiple erasure frameworks but also consistently outperforms existing baselines across key performance metrics, averaging only 4 seconds for anchor mining of a single concept.

[498] The Burden of Interactive Alignment with Inconsistent Preferences

Ali Shirali

Main category: cs.AI

TL;DR: Users with inconsistent preferences can align algorithms with their true interests by being sufficiently foresighted, with a critical horizon determining alignment success. A small costly signal can significantly reduce this burden.

DetailsMotivation: To understand how users with inconsistent preferences can effectively steer engagement-driven algorithms toward their true interests, given that users often engage with low-value content that misleads algorithms.

Method: Model user decision process as dual systems (rational System 2 and impulsive System 1), and analyze a multi-leader, single-follower Stackelberg game where users commit to engagement strategies and the algorithm best-responds.

Result: A critical horizon exists: sufficiently foresighted users achieve alignment, while myopic users become aligned to the algorithm’s objective. This burden can be substantial but is significantly reduced by small costly signals.

Conclusion: Users with inconsistent preferences can align algorithms in Stackelberg equilibrium through strategic foresight, with costly signals offering a practical way to reduce the alignment burden.

Abstract: From media platforms to chatbots, algorithms shape how people interact, learn, and discover information. Such interactions between users and an algorithm often unfold over multiple steps, during which strategic users can guide the algorithm to better align with their true interests by selectively engaging with content. However, users frequently exhibit inconsistent preferences: they may spend considerable time on content that offers little long-term value, inadvertently signaling that such content is desirable. Focusing on the user side, this raises a key question: what does it take for such users to align the algorithm with their true interests? To investigate these dynamics, we model the user’s decision process as split between a rational system 2 that decides whether to engage and an impulsive system 1 that determines how long engagement lasts. We then study a multi-leader, single-follower extensive Stackelberg game, where users, specifically system 2, lead by committing to engagement strategies and the algorithm best-responds based on observed interactions. We define the burden of alignment as the minimum horizon over which users must optimize to effectively steer the algorithm. We show that a critical horizon exists: users who are sufficiently foresighted can achieve alignment, while those who are not are instead aligned to the algorithm’s objective. This critical horizon can be long, imposing a substantial burden. However, even a small, costly signal (e.g., an extra click) can significantly reduce it. Overall, our framework explains how users with inconsistent preferences can align an engagement-driven algorithm with their interests in a Stackelberg equilibrium, highlighting both the challenges and potential remedies for achieving alignment.

[499] Before you , monitor: Implementing Flavell’s metacognitive framework in LLMs

Nick Oh

Main category: cs.AI

TL;DR: The paper proposes a three-phase iterative system based on Flavell’s cognitive monitoring model to bridge the gap between Monitor-Generate and Generate-Verify approaches for LLM reasoning enhancement.

DetailsMotivation: Current LLM reasoning methods are inefficiently separated - Monitor-Generate methods lack verification mechanisms while Generate-Verify approaches start generation without strategic planning, creating inefficiencies in strategy execution and refinement.

Method: Implemented Flavell’s cognitive monitoring model as a three-phase iterative system within the broader Monitor-Generate-Verify framework, combining strategic planning with verification mechanisms.

Result: Achieved 75.42% accuracy on GSM8K, outperforming SELF-REFINE (68.44%) and Self-Verification (67.07%), with fewer attempts (1.3 vs 2.0) at 27-37% increased inference cost.

Conclusion: Upfront monitoring produces higher-quality initial solutions that reduce refinement needs, though evaluation beyond arithmetic reasoning is needed to establish generalizability.

Abstract: Current approaches to enhancing LLM reasoning follows two isolated paradigms: Monitor-Generate methods like Plan-and-Solve (Wang et al., 2023) and SELF-DISCOVER (Zhou et al., 2024) excel at strategic planning but lack mechanisms to verify whether selected strategies succeed; while Generate-Verify approaches like Self-Verification (Weng et al., 2022) and SELF-REFINE (Madaan et al., 2023) iteratively refine outputs but commence generation blindly without task assessment. This separation creates inefficiencies – strategies fail without feedback, and refinement occurs without strategic grounding. We address this gap by implementing Flavell’s cognitive monitoring model (1979) from the broader Monitor-Generate-Verify framework (Oh and Gobet, 2025), operationalising it as a three-phase iterative system. On GSM8K, preliminary results show 75.42% accuracy versus 68.44% for SELF-REFINE and 67.07% for Self-Verification, while requiring fewer attempts (1.3 vs 2.0) at 27-37% increased inference cost. These initial findings suggest upfront monitoring produces higher-quality initial solutions that reduce refinement needs, though evaluation beyond arithmetic reasoning is needed to establish generalisability.

[500] Humanoid-inspired Causal Representation Learning for Domain Generalization

Ze Tao, Jian Zhang, Haowei Li, Xianshuai Li, Yifei Peng, Xiyao Liu, Senzhang Wang, Chao Liu, Sheng Ren, Shichao Zhang

Main category: cs.AI

TL;DR: HSCM is a human-inspired causal framework that improves domain generalization by modeling fine-grained causal mechanisms through attribute disentanglement and reweighting, outperforming conventional methods.

DetailsMotivation: To overcome limitations of conventional domain generalization models that rely on statistical dependencies, by drawing inspiration from human intelligence and hierarchical visual processing.

Method: Proposes Humanoid-inspired Structural Causal Model (HSCM) that replicates human vision systems’ hierarchical processing, disentangles and reweights key image attributes (color, texture, shape), and models fine-grained causal mechanisms.

Result: HSCM outperforms existing domain generalization models in both theoretical and empirical evaluations, providing better generalization across diverse domains with robust performance and interpretability.

Conclusion: HSCM offers a more principled approach for capturing causal relationships and improving model robustness by leveraging human intelligence principles, enabling effective transfer learning in dynamic, complex environments.

Abstract: This paper proposes the Humanoid-inspired Structural Causal Model (HSCM), a novel causal framework inspired by human intelligence, designed to overcome the limitations of conventional domain generalization models. Unlike approaches that rely on statistics to capture data-label dependencies and learn distortion-invariant representations, HSCM replicates the hierarchical processing and multi-level learning of human vision systems, focusing on modeling fine-grained causal mechanisms. By disentangling and reweighting key image attributes such as color, texture, and shape, HSCM enhances generalization across diverse domains, ensuring robust performance and interpretability. Leveraging the flexibility and adaptability of human intelligence, our approach enables more effective transfer and learning in dynamic, complex environments. Through both theoretical and empirical evaluations, we demonstrate that HSCM outperforms existing domain generalization models, providing a more principled method for capturing causal relationships and improving model robustness. The code is available at https://github.com/lambett/HSCM.

[501] RGMem: Renormalization Group-based Memory Evolution for Language Agent User Profile

Ao Tian, Yunfeng Lu, Xinxin Fan, Changhao Wang, Lanzhi Zhou, Yeyao Zhang, Yanfang Liu

Main category: cs.AI

TL;DR: RGMem is a self-evolving memory framework that uses renormalization group principles to create multi-scale user profiles from dialogue history, enabling long-term behavioral consistency in LLM-based conversational systems.

DetailsMotivation: Current LLM systems struggle with cross-session long-term user modeling due to finite context windows and static memory, leading to shallow personalization and lack of continuity. Existing solutions like RAG focus only on fact-level storage without capturing latent preferences.

Method: RGMem organizes dialogue history in multiple scales: extracts semantics from episodic fragments, then applies hierarchical coarse-graining and rescaling operations to progressively form dynamically-evolved user profiles, modeling memory evolution as multi-scale information compression.

Result: The framework enables high-level and accurate user profiles to be formed from noisy, microscopic-level interactions, achieving effective long-term memory and behavioral consistency.

Conclusion: RGMem successfully addresses the limitations of current memory systems by introducing a physics-inspired multi-scale approach that can distill latent user preferences and maintain cross-session continuity in conversational AI systems.

Abstract: Personalized and continuous interactions are the key to enhancing user experience in today’s large language model (LLM)-based conversational systems, however, the finite context windows and static parametric memory make it difficult to model the cross-session long-term user states and behavioral consistency. Currently, the existing solutions to this predicament, such as retrieval-augmented generation (RAG) and explicit memory systems, primarily focus on fact-level storage and retrieval, lacking the capability to distill latent preferences and deep traits from the multi-turn dialogues, which limits the long-term and effective user modeling, directly leading to the personalized interactions remaining shallow, and hindering the cross-session continuity. To realize the long-term memory and behavioral consistency for Language Agents in LLM era, we propose a self-evolving memory framework RGMem, inspired by the ideology of classic renormalization group (RG) in physics, this framework enables to organize the dialogue history in multiple scales: it first extracts semantics and user insights from episodic fragments, then through hierarchical coarse-graining and rescaling operations, progressively forms a dynamically-evolved user profile. The core innovation of our work lies in modeling memory evolution as a multi-scale process of information compression and emergence, which accomplishes the high-level and accurate user profiles from noisy and microscopic-level interactions.

[502] ReviewSense: Transforming Customer Review Dynamics into Actionable Business Insights

Siddhartha Krothapalli, Tridib Kumar Das, Praveen Kumar, Naveen Suravarpu, Pratik Narang

Main category: cs.AI

TL;DR: ReviewSense is a prescriptive decision support framework that uses LLMs to transform customer reviews into actionable business recommendations, going beyond traditional preference prediction to provide strategic insights.

DetailsMotivation: Traditional AI systems focus on predicting user preferences but lack the ability to generate prescriptive, business-facing recommendations from unstructured customer reviews, which are crucial for strategic growth.

Method: Integrates clustering, LLM adaptation, and expert-driven evaluation into a unified pipeline to identify key trends, recurring issues, and specific concerns within customer sentiments.

Result: Preliminary manual evaluations show strong alignment between the model’s recommendations and business objectives, demonstrating potential for data-informed decision-making.

Conclusion: ReviewSense offers a new perspective on AI-driven sentiment analysis, proving valuable for refining business strategies and maximizing the impact of customer feedback.

Abstract: As customer feedback becomes increasingly central to strategic growth, the ability to derive actionable insights from unstructured reviews is essential. While traditional AI-driven systems excel at predicting user preferences, far less work has focused on transforming customer reviews into prescriptive, business-facing recommendations. This paper introduces ReviewSense, a novel prescriptive decision support framework that leverages advanced large language models (LLMs) to transform customer reviews into targeted, actionable business recommendations. By identifying key trends, recurring issues, and specific concerns within customer sentiments, ReviewSense extends beyond preference-based systems to provide businesses with deeper insights for sustaining growth and enhancing customer loyalty. The novelty of this work lies in integrating clustering, LLM adaptation, and expert-driven evaluation into a unified, business-facing pipeline. Preliminary manual evaluations indicate strong alignment between the model’s recommendations and business objectives, highlighting its potential for driving data-informed decision-making. This framework offers a new perspective on AI-driven sentiment analysis, demonstrating its value in refining business strategies and maximizing the impact of customer feedback.

[503] NP-Engine: Empowering Optimization Reasoning in Large Language Models with Verifiable Synthetic NP Problems

Xiaozhe Li, Xinyu Fang, Shengyuan Ding, Linyang Li, Haodong Duan, Qingwen Liu, Kai Chen

Main category: cs.AI

TL;DR: NP-ENGINE is the first framework for training LLMs on NP-hard problems, featuring a generator-verifier-heuristic pipeline for scalable RLVR training. It includes NP-BENCH benchmark and QWEN2.5-7B-NP model that outperforms GPT-4o.

DetailsMotivation: LLMs show strong reasoning in math and coding but their ability to solve complex NP-hard optimization problems remains underexplored.

Method: NP-ENGINE framework with 10 tasks across 5 domains, using generator-verifier-heuristic pipeline for scalable RLVR training with curriculum learning.

Result: QWEN2.5-7B-NP significantly outperforms GPT-4o on NP-BENCH and achieves SOTA performance. RLVR training enables strong out-of-domain generalization to reasoning and non-reasoning tasks.

Conclusion: Task-rich RLVR training is promising for advancing LLM reasoning, revealing new insights into RLVR scaling laws and improving generalization capabilities.

Abstract: Large Language Models (LLMs) have shown strong reasoning capabilities, with models like OpenAI’s O-series and DeepSeek R1 excelling at tasks such as mathematics, coding, logic, and puzzles through Reinforcement Learning with Verifiable Rewards (RLVR). However, their ability to solve more complex optimization problems - particularly NP-hard tasks - remains underexplored. To bridge this gap, we propose NP-ENGINE, the first comprehensive framework for training and evaluating LLMs on NP-hard problems. NP-ENGINE covers 10 tasks across five domains, each equipped with (i) a controllable instance generator, (ii) a rule-based verifier, and (iii) a heuristic solver that provides approximate optimal solutions as ground truth. This generator-verifier-heuristic pipeline enables scalable and verifiable RLVR training under hierarchical difficulties. We also introduce NP-BENCH, a benchmark derived from NP-ENGINE-DATA, specifically designed to evaluate LLMs’ ability to tackle NP-hard level reasoning problems, focusing not only on feasibility but also on solution quality. Additionally, we present QWEN2.5-7B-NP, a model trained via zero-RLVR with curriculum learning on Qwen2.5-7B-Instruct, which significantly outperforms GPT-4o on NP-BENCH and achieves SOTA performance with the same model size. Beyond in-domain tasks, we demonstrate that RLVR training on NP-ENGINE-DATA enables strong out-of-domain (OOD) generalization to reasoning tasks (logic, puzzles, math, and knowledge), as well as non-reasoning tasks such as instruction following. We also observe a scaling trend: increasing task diversity improves OOD generalization. These findings suggest that task-rich RLVR training is a promising direction for advancing LLM’s reasoning ability, revealing new insights into the scaling laws of RLVR.

[504] Hey Pentti, We Did It Again!: Differentiable vector-symbolic types that prove polynomial termination

Eilene Tomkins-Flanagan, Connor Hanley, Mary A. Kelly

Main category: cs.AI

TL;DR: Doug is a typed computer language encoded in vector-symbolic architecture that ensures all programs halt in polynomial time and enables neural networks to learn types, aiming to model human-like skill acquisition through program synthesis.

DetailsMotivation: To model human mental representations and skill acquisition more accurately by creating a system where skill learning follows human-like pace and efficiency, exceeding current approaches.

Method: Encodes the light linear functional programming language (LLFPL) using holographic declarative memory for types and a Lisp VSA variant for terms, allowing neural networks to learn types from embedding spaces.

Result: Developed Doug language where types are learnable by neural networks and nearby points in embedding space have similar types in structure and content.

Conclusion: Doug represents progress toward modeling actual human mental representations and their acquisition, bringing us closer to understanding how skills are learned in the brain.

Abstract: We present a typed computer language, Doug, in which all typed programs may be proved to halt in polynomial time, encoded in a vector-symbolic architecture (VSA). Doug is just an encoding of the light linear functional programming language (LLFPL) described by (Schimanski2009, ch. 7). The types of Doug are encoded using a slot-value encoding scheme based on holographic declarative memory (HDM; Kelly, 2020). The terms of Doug are encoded using a variant of the Lisp VSA defined by (Flanagan, 2024). Doug allows for some points on the embedding space of a neural network to be interpreted as types, where the types of nearby points are similar both in structure and content. Types in Doug are therefore learnable by a neural network. Following (Chollet, 2019), (Card, 1983), and (Newell, 1981), we view skill as the application of a procedure, or program of action, that causes a goal to be satisfied. Skill acquisition may therefore be expressed as program synthesis. Using Doug, we hope to describe a form of learning of skilled behaviour that follows a human-like pace of skill acquisition (i.e., substantially faster than brute force; Heathcote, 2000), exceeding the efficiency of all currently existing approaches (Kaplan, 2020; Jones, 2021; Chollet, 2024). Our approach brings us one step closer to modeling human mental representations, as they must actually exist in the brain, and those representations’ acquisition, as they are actually learned.

[505] Smart Traffic Signals: Comparing MARL and Fixed-Time Strategies

Saahil Mahato

Main category: cs.AI

TL;DR: MARL-based traffic signal control reduces wait times and improves throughput compared to fixed-time systems in urban intersection simulations.

DetailsMotivation: Traditional fixed-time traffic signals lack adaptability to dynamic traffic patterns, causing congestion, increased travel time, fuel consumption, and emissions.

Method: Developed a decentralized multi-agent reinforcement learning (MARL) system using Pygame simulation, where each traffic signal acts as an autonomous agent making decisions based on local observations and neighbor information.

Result: MARL approach showed statistically significant improvements over baseline fixed-time controller, with reduced average vehicle wait times and improved overall throughput.

Conclusion: MARL-based dynamic control strategies show substantial promise for urban traffic management efficiency, though more research is needed for scalability and real-world implementation.

Abstract: Urban traffic congestion, particularly at intersections, significantly impacts travel time, fuel consumption, and emissions. Traditional fixed-time signal control systems often lack the adaptability to manage dynamic traffic patterns effectively. This study explores the application of multi-agent reinforcement learning (MARL) to optimize traffic signal coordination across multiple intersections within a simulated environment. Utilizing Pygame, a simulation was developed to model a network of interconnected intersections with randomly generated vehicle flows to reflect realistic traffic variability. A decentralized MARL controller was implemented, in which each traffic signal operates as an autonomous agent, making decisions based on local observations and information from neighboring agents. Performance was evaluated against a baseline fixed-time controller using metrics such as average vehicle wait time and overall throughput. The MARL approach demonstrated statistically significant improvements, including reduced average waiting times and improved throughput. These findings suggest that MARL-based dynamic control strategies hold substantial promise for improving urban traffic management efficiency. More research is recommended to address scalability and real-world implementation challenges.

[506] Urban-R1: Reinforced MLLMs Mitigate Geospatial Biases for Urban General Intelligence

Qiongyan Wang, Xingchen Zou, Yutian Jiang, Haomin Wen, Jiaheng Wei, Qingsong Wen, Yuxuan Liang

Main category: cs.AI

TL;DR: Urban-R1 is a reinforcement learning framework that addresses geospatial bias in urban foundation models by using Group Relative Policy Optimization and urban region profiling to improve cross-region generalization and reduce regional skew in predictions.

DetailsMotivation: Rapid urbanization creates demand for Urban General Intelligence, but current supervised fine-tuning approaches suffer from persistent geospatial bias, producing regionally skewed predictions and limited generalization across different urban areas.

Method: Proposes Urban-R1, a reinforcement learning-based post-training framework using Group Relative Policy Optimization (GRPO) to optimize reasoning across geographic groups, with urban region profiling as a proxy task to provide measurable rewards from multimodal urban data.

Result: Extensive experiments show Urban-R1 effectively mitigates geo-bias and improves cross-region generalization, outperforming both supervised fine-tuning trained models and closed-source models across diverse regions and tasks.

Conclusion: Reinforcement learning alignment represents a promising pathway toward equitable and trustworthy urban intelligence by addressing geospatial bias in urban foundation models.

Abstract: Rapid urbanization intensifies the demand for Urban General Intelligence (UGI), referring to AI systems that can understand and reason about complex urban environments. Recent studies have built urban foundation models using supervised fine-tuning (SFT) of LLMs and MLLMs, yet these models exhibit persistent geospatial bias, producing regionally skewed predictions and limited generalization. To this end, we propose Urban-R1, a reinforcement learning-based post-training framework that aligns MLLMs with the objectives of UGI. Urban-R1 adopts Group Relative Policy Optimization (GRPO) to optimize reasoning across geographic groups and employs urban region profiling as a proxy task to provide measurable rewards from multimodal urban data. Extensive experiments across diverse regions and tasks show that Urban-R1 effectively mitigates geo-bias and improves cross-region generalization, outperforming both SFT-trained and closed-source models. Our results highlight reinforcement learning alignment as a promising pathway toward equitable and trustworthy urban intelligence.

[507] BuildArena: A Physics-Aligned Interactive Benchmark of LLMs for Engineering Construction

Tian Xia, Tianrun Gao, Wenhao Deng, Long Wei, Xiaowei Qian, Yixian Jiang, Chenglei Yu, Tailin Wu

Main category: cs.AI

TL;DR: BuildArena is the first physics-aligned interactive benchmark for evaluating LLMs’ capabilities in engineering construction automation, featuring customizable framework, extendable tasks, 3D spatial computation, and baseline agent workflow.

DetailsMotivation: To address the gap in evaluating LLMs' construction competencies despite their broad knowledge and reasoning capabilities, as engineering construction automation requires complex reasoning under strict physical constraints.

Method: Developed BuildArena benchmark with four components: customizable framework, extendable task design spanning static/dynamic mechanics, 3D Spatial Geometric Computation Library, and baseline LLM agentic workflow for evaluation.

Result: Comprehensively evaluated eight frontier LLMs on their capabilities for language-driven and physics-grounded construction automation using the BuildArena benchmark.

Conclusion: BuildArena provides the first standardized framework for assessing LLMs in engineering construction automation, enabling systematic comparison and analysis of model capabilities in this domain.

Abstract: Engineering construction automation aims to transform natural language specifications into physically viable structures, requiring complex integrated reasoning under strict physical constraints. While modern LLMs possess broad knowledge and strong reasoning capabilities that make them promising candidates for this domain, their construction competencies remain largely unevaluated. To address this gap, we introduce BuildArena, the first physics-aligned interactive benchmark designed for language-driven engineering construction. It contributes to the community in four aspects: (1) a highly customizable benchmarking framework for in-depth comparison and analysis of LLMs; (2) an extendable task design strategy spanning static and dynamic mechanics across multiple difficulty tiers; (3) a 3D Spatial Geometric Computation Library for supporting construction based on language instructions; (4) a baseline LLM agentic workflow that effectively evaluates diverse model capabilities. On eight frontier LLMs, BuildArena comprehensively evaluates their capabilities for language-driven and physics-grounded construction automation. The project page is at https://build-arena.github.io/.

[508] Can Knowledge-Graph-based Retrieval Augmented Generation Really Retrieve What You Need?

Junchi Yu, Yujie Liu, Jindong Gu, Philip Torr, Dongzhan Zhou

Main category: cs.AI

TL;DR: GraphFlow is a KG-based RAG framework that uses transition-based flow matching to retrieve accurate and diverse knowledge from text-rich knowledge graphs for complex real-world queries, outperforming existing methods.

DetailsMotivation: Existing KG-based RAG methods struggle with retrieving accurate and diverse information from text-rich KGs for complex queries, and PRMs require expensive process-level supervision that's hard to obtain on KGs.

Method: Uses transition-based flow matching to jointly optimize a retrieval policy and flow estimator, which factorizes retrieval outcome rewards into intermediate states to guide proportional candidate retrieval from KGs.

Result: Outperforms strong KG-RAG baselines including GPT-4o by 10% on average in hit rate and recall on STaRK benchmark, with strong generalization to unseen KGs.

Conclusion: GraphFlow effectively retrieves diverse and relevant knowledge from text-rich KGs for real-world queries, demonstrating robustness and superior performance over existing methods.

Abstract: Retrieval-Augmented Generation (RAG) based on knowledge graphs (KGs) enhances large language models (LLMs) by providing structured and interpretable external knowledge. However, existing KG-based RAG methods struggle to retrieve accurate and diverse information from text-rich KGs for complex real-world queries. Process Reward Models (PRMs) offer a way to align the retrieval process of KG-based RAG with query-specific knowledge requirements, but they heavily rely on process-level supervision signals that are expensive and hard to obtain on KGs. To address this challenge, we propose GraphFlow, a framework that efficiently retrieves accurate and diverse knowledge required for real-world queries from text-rich KGs. GraphFlow employs a transition-based flow matching objective to jointly optimize a retrieval policy and a flow estimator. The flow estimator factorizes the reward of the retrieval outcome into the intermediate retrieval states. Such reward factorization guides the retrieval policy to retrieve candidates from KGs in proportion to their reward. This allows GraphFlow to explore high-quality regions of KGs that yield diverse and relevant results. We evaluate GraphFlow on the STaRK benchmark, which includes real-world queries from multiple domains over text-rich KGs. GraphFlow outperforms strong KG-RAG baselines, including GPT-4o, by 10% on average in hit rate and recall. It also shows strong generalization to unseen KGs, demonstrating its effectiveness and robustness.

[509] Uncertain Knowledge Graph Completion via Semi-Supervised Confidence Distribution Learning

Tianxing Wu, Shutong Zhu, Jingting Wang, Ning Xu, Guilin Qi, Haofen Wang

Main category: cs.AI

TL;DR: Proposes ssCDL, a semi-supervised confidence distribution learning method for uncertain knowledge graph completion that addresses imbalanced confidence distributions through meta-learning and pseudo-labeling.

DetailsMotivation: Current UKG completion methods neglect the extremely imbalanced distributions of triple confidences, causing insufficient embeddings for high-quality completion.

Method: Transforms triple confidences into distributions, uses semi-supervised learning with relational learning on labeled data and unlabeled data with pseudo labels generated by meta-learning to augment training data and rebalance confidence distributions.

Result: Experiments on two UKG datasets show ssCDL consistently outperforms state-of-the-art baselines across different evaluation metrics.

Conclusion: ssCDL effectively addresses the imbalanced confidence distribution problem in UKG completion through confidence distribution learning and semi-supervised meta-learning.

Abstract: Uncertain knowledge graphs (UKGs) associate each triple with a confidence score to provide more precise knowledge representations. Recently, since real-world UKGs suffer from the incompleteness, uncertain knowledge graph (UKG) completion attracts more attention, aiming to complete missing triples and confidences. Current studies attempt to learn UKG embeddings to solve this problem, but they neglect the extremely imbalanced distributions of triple confidences. This causes that the learnt embeddings are insufficient to high-quality UKG completion. Thus, in this paper, to address the above issue, we propose a new semi-supervised Confidence Distribution Learning (ssCDL) method for UKG completion, where each triple confidence is transformed into a confidence distribution to introduce more supervision information of different confidences to reinforce the embedding learning process. ssCDL iteratively learns UKG embedding by relational learning on labeled data (i.e., existing triples with confidences) and unlabeled data with pseudo labels (i.e., unseen triples with the generated confidences), which are predicted by meta-learning to augment the training data and rebalance the distribution of triple confidences. Experiments on two UKG datasets demonstrate that ssCDL consistently outperforms state-of-the-art baselines in different evaluation metrics.

[510] Agentic System with Modal Logic for Autonomous Diagnostics

Antonin Sulc, Thorsten Hellert

Main category: cs.AI

TL;DR: This paper introduces a neuro-symbolic multi-agent architecture that combines language models with modal logic reasoning using Kripke models to enable robust, explainable diagnosis of complex failures in challenging environments.

DetailsMotivation: Current AI research focuses on scaling models and datasets, but neglects scaling the structure, fidelity, and logical consistency of agent reasoning in complex environments that require adaptive, autonomous decision-making.

Method: A neuro-symbolic multi-agent architecture where agents’ belief states are represented as Kripke models, enabling reasoning about possibility and necessity using modal logic. Domain-specific knowledge is encoded as logical constraints to guide hypothesis generation and prevent untenable conclusions.

Result: The system successfully diagnoses complex, cascading failures in a high-fidelity simulated particle accelerator environment by combining semantic intuition of language models with rigorous validation of modal logic and factual world models.

Conclusion: This approach showcases a viable path toward more robust, reliable, and verifiable autonomous agents by integrating the strengths of language models with formal logical reasoning.

Abstract: The development of intelligent agents, particularly those powered by language models (LMs), has shown a critical role in various environments that require intelligent and autonomous decision-making. Environments are not passive testing grounds, and they represent the data required for agents to learn and exhibit in very challenging conditions that require adaptive, complex, and autonomous capacity to make decisions. While the paradigm of scaling models and datasets has led to remarkable emergent capabilities, we argue that scaling the structure, fidelity, and logical consistency of agent reasoning within these environments is a crucial, yet underexplored, dimension of AI research. This paper introduces a neuro-symbolic multi-agent architecture where the belief states of individual agents are formally represented as Kripke models. This foundational choice enables them to reason about known concepts of \emph{possibility} and \emph{necessity} using the formal language of modal logic. In this work, we use immutable, domain-specific knowledge to make an informed root cause diagnosis, which is encoded as logical constraints essential for proper, reliable, and explainable diagnosis. In the proposed model, we show constraints that actively guide the hypothesis generation of LMs, effectively preventing them from reaching physically or logically untenable conclusions. In a high-fidelity simulated particle accelerator environment, our system successfully diagnoses complex, cascading failures by combining the powerful semantic intuition of LMs with the rigorous, verifiable validation of modal logic and a factual world model and showcasing a viable path toward more robust, reliable, and verifiable autonomous agents.

[511] Count Counts: Motivating Exploration in LLM Reasoning with Count-based Intrinsic Rewards

Xuan Zhang, Ruixiao Li, Zhijian Zhou, Long Li, Yulei Qin, Ke Li, Xing Sun, Xiaoyu Tan, Chao Qu, Yuan Qi

Main category: cs.AI

TL;DR: MERCI is a novel RL algorithm that uses count-based intrinsic rewards to improve exploration in LLM reasoning, addressing issues of repetitive patterns and suboptimal solutions in current RL approaches.

DetailsMotivation: Current RL paradigms for LLMs rely on sparse outcome-based rewards and limited exploration, leading to repetitive and suboptimal reasoning patterns. The paper aims to design better exploration strategies for LLM reasoning.

Method: MERCI uses a lightweight Coin Flipping Network (CFN) to estimate pseudo count and epistemic uncertainty over reasoning trajectories, converting them into intrinsic rewards that value novelty while preserving task reward signals. It’s integrated into RL frameworks like GRPO.

Result: Experiments show MERCI encourages richer and more varied chains of thought, significantly improves performance over strong baselines, and helps policies escape local routines to discover better solutions.

Conclusion: Targeted intrinsic motivation can make exploration reliable for language model reasoning, with MERCI demonstrating effective exploration enhancement in complex reasoning tasks.

Abstract: Reinforcement Learning (RL) has become a compelling way to strengthen the multi step reasoning ability of Large Language Models (LLMs). However, prevalent RL paradigms still lean on sparse outcome-based rewards and limited exploration, which often drives LLMs toward repetitive and suboptimal reasoning patterns. In this paper, we study the central question of how to design exploration for LLM reasoning and introduce MERCI (Motivating Exploration in LLM Reasoning with Count-based Intrinsic Rewards), a novel RL algorithm that augments policy optimization with a principled intrinsic reward. Building on the idea of count-based exploration, MERCI leverages a lightweight Coin Flipping Network (CFN) to estimate the pseudo count and further epistemic uncertainty over reasoning trajectories, and converts them into an intrinsic reward that values novelty while preserving the learning signal from task rewards. We integrate MERCI into some advanced RL frameworks like Group Relative Policy Optimization (GRPO). Experiments on complex reasoning benchmarks demonstrate that MERCI encourages richer and more varied chains of thought, significantly improves performance over strong baselines, and helps the policy escape local routines to discover better solutions. It indicates that our targeted intrinsic motivation can make exploration reliable for language model reasoning.

[512] Co-Alignment: Rethinking Alignment as Bidirectional Human-AI Cognitive Adaptation

Yubo Li, Weiyi Song

Main category: cs.AI

TL;DR: The paper proposes Bidirectional Cognitive Alignment (BiCA) as a new paradigm where humans and AI mutually adapt, achieving better collaboration than traditional single-directional alignment approaches.

DetailsMotivation: Current AI alignment through RLHF treats human cognition as fixed, creating a single directional paradigm where only AI adapts to human preferences. The authors argue for shifting to co-alignment where both humans and AI mutually adapt.

Method: BiCA uses learnable protocols, representation mapping, and KL-budget constraints for controlled co-evolution between humans and AI systems.

Result: In collaborative navigation tasks, BiCA achieved 85.5% success rate (vs 70.3% baseline), with 230% better mutual adaptation and 332% better protocol convergence. Emergent protocols outperformed handcrafted ones by 84%, and bidirectional adaptation improved safety with +23% out-of-distribution robustness.

Conclusion: The 46% synergy improvement demonstrates that optimal collaboration exists at the intersection of human and AI capabilities, validating the shift from single-directional to co-alignment paradigms.

Abstract: Current AI alignment through RLHF follows a single directional paradigm that AI conforms to human preferences while treating human cognition as fixed. We propose a shift to co-alignment through Bidirectional Cognitive Alignment (BiCA), where humans and AI mutually adapt. BiCA uses learnable protocols, representation mapping, and KL-budget constraints for controlled co-evolution. In collaborative navigation, BiCA achieved 85.5% success versus 70.3% baseline, with 230% better mutual adaptation and 332% better protocol convergence. Emergent protocols outperformed handcrafted ones by 84%, while bidirectional adaptation unexpectedly improved safety (+23% out-of-distribution robustness). The 46% synergy improvement demonstrates optimal collaboration exists at the intersection, not union, of human and AI capabilities, validating the shift from single-directional to co-alignment paradigms.

[513] Foundation and Large-Scale AI Models in Neuroscience: A Comprehensive Review

Shihao Yang, Xiying Huang, Danilo Bernardo, Jun-En Ding, Andrew Michael, Jingmei Yang, Patrick Kwan, Ashish Raj, Feng Liu

Main category: cs.AI

TL;DR: Large-scale AI models are transforming neuroscience by enabling end-to-end learning from raw brain signals and neural data across five major domains: neuroimaging, brain-computer interfaces, molecular neuroscience, clinical applications, and disease-specific uses.

DetailsMotivation: To explore how large-scale AI models address computational neuroscience challenges like multimodal data integration, spatiotemporal pattern interpretation, and creating clinical translation frameworks, while also examining the reciprocal relationship between neuroscience and AI.

Method: Review and analysis of large-scale AI model applications across five neuroscience domains, focusing on their ability to facilitate end-to-end learning from raw neural data and integrate biologically informed architectural constraints.

Result: Demonstrated that these models successfully address major computational neuroscience challenges and enable multimodal neural data integration, with the interaction between neuroscience and AI becoming increasingly reciprocal.

Conclusion: Large-scale AI models show significant promise for neuroscience but require rigorous evaluation frameworks, effective domain knowledge integration, and comprehensive ethical guidelines for clinical deployment, along with systematic validation using critical neuroscience datasets.

Abstract: The advent of large-scale artificial intelligence (AI) models has a transformative effect on neuroscience research, which represents a paradigm shift from the traditional computational methods through the facilitation of end-to-end learning from raw brain signals and neural data. In this paper, we explore the transformative effects of large-scale AI models on five major neuroscience domains: neuroimaging and data processing, brain-computer interfaces and neural decoding, molecular neuroscience and genomic modeling, clinical assistance and translational frameworks, and disease-specific applications across neurological and psychiatric disorders. These models are demonstrated to address major computational neuroscience challenges, including multimodal neural data integration, spatiotemporal pattern interpretation, and the derivation of translational frameworks for clinical deployment. Moreover, the interaction between neuroscience and AI has become increasingly reciprocal, as biologically informed architectural constraints are now incorporated to develop more interpretable and computationally efficient models. This review highlights both the notable promise of such technologies and key implementation considerations, with particular emphasis on rigorous evaluation frameworks, effective domain knowledge integration, and comprehensive ethical guidelines for clinical use. Finally, a systematic listing of critical neuroscience datasets used to derive and validate large-scale AI models across diverse research applications is provided.

[514] An Agentic Framework with LLMs for Solving Complex Vehicle Routing Problems

Ni Zhang, Zhiguang Cao, Jianan Zhou, Cong Zhang, Yew-Soon Ong

Main category: cs.AI

TL;DR: AFL is an agentic framework using LLMs to fully automate complex vehicle routing problems from raw inputs to solutions, achieving high reliability and feasibility without external intervention.

DetailsMotivation: Current LLM approaches for VRPs require external intervention, leading to execution errors and low solution feasibility. There's a need for fully automated systems that can interpret intents and generate solutions autonomously.

Method: Proposes AFL framework that decomposes VRPs into three subtasks managed by four specialized agents. The agents coordinate to enforce consistency and logical soundness, enabling self-contained code generation without handcrafted modules or external solvers.

Result: Extensive experiments on 60 complex VRPs show comparable performance to meticulously designed algorithms. Substantially outperforms existing LLM-based baselines with code reliability and solution feasibility rates close to 100% on evaluated benchmarks.

Conclusion: AFL demonstrates that agentic frameworks with LLMs can achieve full automation for complex VRPs, providing trustworthy solutions without external dependencies while maintaining high performance standards.

Abstract: Complex vehicle routing problems (VRPs) remain a fundamental challenge, demanding substantial expert effort for intent interpretation and algorithm design. While large language models (LLMs) offer a promising path toward automation, current approaches still rely on external intervention, which restrict autonomy and often lead to execution errors and low solution feasibility. To address these challenges, we propose an Agentic Framework with LLMs (AFL) for solving complex vehicle routing problems, achieving full automation from problem instance to solution. AFL directly extracts knowledge from raw inputs and enables self-contained code generation without handcrafted modules or external solvers. To improve trustworthiness, AFL decomposes the overall pipeline into three manageable subtasks and employs four specialized agents whose coordinated interactions enforce cross-functional consistency and logical soundness. Extensive experiments on 60 complex VRPs, ranging from standard benchmarks to practical variants, validate the effectiveness and generality of our framework, showing comparable performance against meticulously designed algorithms. Notably, it substantially outperforms existing LLM-based baselines in both code reliability and solution feasibility, achieving rates close to 100% on the evaluated benchmarks.

[515] Beyond Pipelines: A Survey of the Paradigm Shift toward Model-Native Agentic AI

Jitao Sang, Jinlin Xiao, Jiarun Han, Jilin Chen, Xiaoyi Chen, Shuyu Wei, Yongjie Sun, Yuhang Wang

Main category: cs.AI

TL;DR: Survey traces the paradigm shift from pipeline-based agentic AI systems to model-native approaches where planning, tool use, and memory are internalized within LLM parameters, enabled by reinforcement learning.

DetailsMotivation: To document and analyze the evolution of agentic AI from systems that apply external intelligence to models that grow intelligence through experience and internalized capabilities.

Method: Systematic review of how planning, tool use, and memory capabilities have evolved from externally scripted modules to end-to-end learned behaviors, examining the role of RL in enabling this shift across language, vision and embodied domains.

Result: Identifies a coherent trajectory toward model-native agentic AI as an integrated learning and interaction framework, with applications in Deep Research agents (long-horizon reasoning) and GUI agents (embodied interaction).

Conclusion: The paradigm shift marks the transition from constructing systems that apply intelligence to developing models that grow intelligence through experience, with continued internalization of capabilities like multi-agent collaboration and reflection.

Abstract: The rapid evolution of agentic AI marks a new phase in artificial intelligence, where Large Language Models (LLMs) no longer merely respond but act, reason, and adapt. This survey traces the paradigm shift in building agentic AI: from Pipeline-based systems, where planning, tool use, and memory are orchestrated by external logic, to the emerging Model-native paradigm, where these capabilities are internalized within the model’s parameters. We first position Reinforcement Learning (RL) as the algorithmic engine enabling this paradigm shift. By reframing learning from imitating static data to outcome-driven exploration, RL underpins a unified solution of LLM + RL + Task across language, vision and embodied domains. Building on this, the survey systematically reviews how each capability – Planning, Tool use, and Memory – has evolved from externally scripted modules to end-to-end learned behaviors. Furthermore, it examines how this paradigm shift has reshaped major agent applications, specifically the Deep Research agent emphasizing long-horizon reasoning and the GUI agent emphasizing embodied interaction. We conclude by discussing the continued internalization of agentic capabilities like Multi-agent collaboration and Reflection, alongside the evolving roles of the system and model layers in future agentic AI. Together, these developments outline a coherent trajectory toward model-native agentic AI as an integrated learning and interaction framework, marking the transition from constructing systems that apply intelligence to developing models that grow intelligence through experience.

[516] A Comprehensive Survey on Reinforcement Learning-based Agentic Search: Foundations, Roles, Optimizations, Evaluations, and Applications

Minhua Lin, Zongyu Wu, Zhichao Xu, Hui Liu, Xianfeng Tang, Qi He, Charu Aggarwal, Hui Liu, Xiang Zhang, Suhang Wang

Main category: cs.AI

TL;DR: This survey provides a comprehensive overview of RL-based agentic search, organizing the field along three dimensions: functional roles of RL, optimization strategies, and scope of optimization.

DetailsMotivation: Traditional RAG pipelines are single-turn and heuristic, lacking adaptive control over retrieval and reasoning. RL offers a powerful mechanism for adaptive and self-improving search behavior in agentic search systems.

Method: The survey organizes RL-based agentic search along three dimensions: (i) What RL is for (functional roles), (ii) How RL is used (optimization strategies), and (iii) Where RL is applied (scope of optimization). It summarizes representative methods, evaluation protocols, and applications.

Result: The survey provides the first comprehensive overview of RL-based agentic search, organizing the emerging field and discussing open challenges and future directions.

Conclusion: RL-based agentic search addresses limitations of traditional RAG by enabling adaptive control over retrieval and reasoning through multi-step interaction with search environments, offering promising directions for building reliable and scalable search systems.

Abstract: The advent of large language models (LLMs) has transformed information access and reasoning through open-ended natural language interaction. However, LLMs remain limited by static knowledge, factual hallucinations, and the inability to retrieve real-time or domain-specific information. Retrieval-Augmented Generation (RAG) mitigates these issues by grounding model outputs in external evidence, but traditional RAG pipelines are often single turn and heuristic, lacking adaptive control over retrieval and reasoning. Recent advances in agentic search address these limitations by enabling LLMs to plan, retrieve, and reflect through multi-step interaction with search environments. Within this paradigm, reinforcement learning (RL) offers a powerful mechanism for adaptive and self-improving search behavior. This survey provides the first comprehensive overview of \emph{RL-based agentic search}, organizing the emerging field along three complementary dimensions: (i) What RL is for (functional roles), (ii) How RL is used (optimization strategies), and (iii) Where RL is applied (scope of optimization). We summarize representative methods, evaluation protocols, and applications, and discuss open challenges and future directions toward building reliable and scalable RL driven agentic search systems. We hope this survey will inspire future research on the integration of RL and agentic search. Our repository is available at https://github.com/ventr1c/Awesome-RL-based-Agentic-Search-Papers.

[517] ELMM: Efficient Lightweight Multimodal Large Language Models for Multimodal Knowledge Graph Completion

Wei Huang, Peining Li, Meiyu Liang, Xu Hou, Junping Du, Yingxia Shao, Guanhua Ye, Wu Liu, Kangkang Lu, Yang Yu

Main category: cs.AI

TL;DR: ELMM proposes an efficient multimodal LLM for knowledge graph completion that compresses image tokens and prunes attention layers to reduce computational costs while maintaining performance.

DetailsMotivation: Existing multimodal knowledge graphs suffer from incompleteness, and applying MLLMs to completion tasks faces challenges with semantic noise from image tokens and high computational costs.

Method: Proposes Multi-view Visual Token Compressor using multi-head attention to compress image tokens from textual/visual views, plus attention pruning with linear projection to reduce model size.

Result: Achieves state-of-the-art performance on FB15k-237-IMG and WN18-IMG benchmarks while significantly improving computational efficiency.

Conclusion: ELMM establishes a new paradigm for multimodal knowledge graph completion by balancing performance and efficiency through token compression and model pruning.

Abstract: Multimodal Knowledge Graphs (MKGs) extend traditional knowledge graphs by incorporating visual and textual modalities, enabling richer and more expressive entity representations. However, existing MKGs often suffer from incompleteness, which hinder their effectiveness in downstream tasks. Therefore, multimodal knowledge graph completion (MKGC) task is receiving increasing attention. While large language models (LLMs) have shown promise for knowledge graph completion (KGC), their application to the multimodal setting remains underexplored. Moreover, applying Multimodal Large Language Models (MLLMs) to the task of MKGC introduces significant challenges: (1) the large number of image tokens per entity leads to semantic noise and modality conflicts, and (2) the high computational cost of processing large token inputs. To address these issues, we propose Efficient Lightweight Multimodal Large Language Models (ELMM) for MKGC. ELMM proposes a Multi-view Visual Token Compressor (MVTC) based on multi-head attention mechanism, which adaptively compresses image tokens from both textual and visual views, thereby effectively reducing redundancy while retaining necessary information and avoiding modality conflicts. Additionally, we design an attention pruning strategy to remove redundant attention layers from MLLMs, thereby significantly reducing the inference cost. We further introduce a linear projection to compensate for the performance degradation caused by pruning. Extensive experiments on benchmark FB15k-237-IMG and WN18-IMG demonstrate that ELMM achieves state-of-the-art performance while substantially improving computational efficiency, establishing a new paradigm for multimodal knowledge graph completion.

[518] See or Say Graphs: Agent-Driven Scalable Graph Understanding with Vision-Language Models

Shuo Han, Yukun Cao, Zezhong Ding, Zengyi Gao, S Kevin Zhou, Xike Xie

Main category: cs.AI

TL;DR: GraphVista is a unified framework that enhances graph understanding by improving scalability through hierarchical organization and better modality coordination between text and visual inputs.

DetailsMotivation: Current vision-language models face scalability bottlenecks due to input-token constraints and lack effective mechanisms to coordinate textual and visual modalities for graph understanding tasks.

Method: GraphVista uses hierarchical organization with a GraphRAG base for scalability, retrieving only task-relevant content. It employs a planning agent to route tasks to the most suitable modality - text for simple properties and vision for complex structural reasoning.

Result: GraphVista scales to graphs 200× larger than existing benchmarks and achieves up to 4.4× quality improvement over state-of-the-art methods by fully exploiting complementary modality strengths.

Conclusion: GraphVista successfully addresses scalability and modality coordination challenges in graph understanding, demonstrating superior performance across various graph sizes and reasoning tasks.

Abstract: Vision-language models (VLMs) have shown promise in graph understanding, but remain limited by input-token constraints, facing scalability bottlenecks and lacking effective mechanisms to coordinate textual and visual modalities. To address these challenges, we propose GraphVista, a unified framework that enhances both scalability and modality coordination in graph understanding. For scalability, GraphVista organizes graph information hierarchically into a lightweight GraphRAG base, which retrieves only task-relevant textual descriptions and high-resolution visual subgraphs, compressing redundant context while preserving key reasoning elements. For modality coordination, GraphVista introduces a planning agent that routes tasks to the most suitable modality-using the text modality for simple property reasoning and the visual modality for local and structurally complex reasoning grounded in explicit topology. Extensive experiments demonstrate that GraphVista scales to large graphs, up to $200\times$ larger than those used in existing benchmarks, and consistently outperforms existing textual, visual, and fusion-based methods, achieving up to $4.4\times$ quality improvement over the state-of-the-art baselines by fully exploiting the complementary strengths of both modalities.

[519] Domain-Contextualized Concept Graphs: A Computable Framework for Knowledge Representation

Chao Li, Yuru Wang

Main category: cs.AI

TL;DR: The paper proposes Domain-Contextualized Concept Graph (CDC), a knowledge modeling framework that treats domains as first-class elements using C-D-C triples to enable context-aware reasoning and cross-domain analogy.

DetailsMotivation: Traditional knowledge graphs are limited by fixed ontologies with rigid hierarchical structures, treating domains as implicit context rather than explicit reasoning components.

Method: CDC uses a C-D-C triple structure <Concept, Relation@Domain, Concept’> with domain specifications as dynamic classification dimensions, formalizes standardized relation predicates, and implements in Prolog for inference.

Result: Case studies in education, enterprise systems, and technical documentation show CDC enables context-aware reasoning, cross-domain analogy, and personalized knowledge modeling not possible with traditional ontology frameworks.

Conclusion: CDC overcomes limitations of traditional knowledge graphs by making domains explicit reasoning components, enabling more flexible and context-sensitive knowledge representation.

Abstract: Traditional knowledge graphs are constrained by fixed ontologies that organize concepts within rigid hierarchical structures. The root cause lies in treating domains as implicit context rather than as explicit, reasoning-level components. To overcome these limitations, we propose the Domain-Contextualized Concept Graph (CDC), a novel knowledge modeling framework that elevates domains to first-class elements of conceptual representation. CDC adopts a C-D-C triple structure - <Concept, Relation@Domain, Concept’> - where domain specifications serve as dynamic classification dimensions defined on demand. Grounded in a cognitive-linguistic isomorphic mapping principle, CDC operationalizes how humans understand concepts through contextual frames. We formalize more than twenty standardized relation predicates (structural, logical, cross-domain, and temporal) and implement CDC in Prolog for full inference capability. Case studies in education, enterprise knowledge systems, and technical documentation demonstrate that CDC enables context-aware reasoning, cross-domain analogy, and personalized knowledge modeling - capabilities unattainable under traditional ontology-based frameworks.

[520] DeepAnalyze: Agentic Large Language Models for Autonomous Data Science

Shaolei Zhang, Ju Fan, Meihao Fan, Guoliang Li, Xiaoyong Du

Main category: cs.AI

TL;DR: DeepAnalyze-8B is the first agentic LLM for autonomous data science that can complete end-to-end pipelines from data sources to research reports, outperforming workflow-based agents despite having only 8B parameters.

DetailsMotivation: Existing workflow-based data agents are limited by predefined workflows and cannot achieve fully autonomous data science. The emergence of powerful LLMs makes autonomous data science from raw data to deep research reports feasible.

Method: Proposed curriculum-based agentic training paradigm that emulates human data scientists’ learning trajectory, and a data-grounded trajectory synthesis framework for constructing high-quality training data.

Result: DeepAnalyze-8B outperforms previous workflow-based agents built on most advanced proprietary LLMs, demonstrating capability across data question answering, specialized analytical tasks, and open-ended data research.

Conclusion: DeepAnalyze-8B paves the way toward autonomous data science with its open-source model, code, and training data, showing that agentic training enables LLMs to progressively acquire multiple capabilities for complex data science tasks.

Abstract: Autonomous data science, from raw data sources to analyst-grade deep research reports, has been a long-standing challenge, and is now becoming feasible with the emergence of powerful large language models (LLMs). Recent workflow-based data agents have shown promising results on specific data tasks but remain fundamentally limited in achieving fully autonomous data science due to their reliance on predefined workflows. In this paper, we introduce DeepAnalyze-8B, the first agentic LLM designed for autonomous data science, capable of automatically completing the end-toend pipeline from data sources to analyst-grade deep research reports. To tackle high-complexity data science tasks, we propose a curriculum-based agentic training paradigm that emulates the learning trajectory of human data scientists, enabling LLMs to progressively acquire and integrate multiple capabilities in real-world environments. We also introduce a data-grounded trajectory synthesis framework that constructs high-quality training data. Through agentic training, DeepAnalyze learns to perform a broad spectrum of data tasks, ranging from data question answering and specialized analytical tasks to open-ended data research. Experiments demonstrate that, with only 8B parameters, DeepAnalyze outperforms previous workflow-based agents built on most advanced proprietary LLMs. The model, code, and training data of DeepAnalyze are open-sourced, paving the way toward autonomous data science.

[521] VAGEN: Reinforcing World Model Reasoning for Multi-Turn VLM Agents

Kangrui Wang, Pingyue Zhang, Zihan Wang, Yaning Gao, Linjie Li, Qineng Wang, Hanyang Chen, Chi Wan, Yiping Lu, Zhengyuan Yang, Lijuan Wang, Ranjay Krishna, Jiajun Wu, Li Fei-Fei, Yejin Choi, Manling Li

Main category: cs.AI

TL;DR: This paper proposes training Vision-Language Model (VLM) agents through explicit visual state reasoning using reinforcement learning, decomposing reasoning into state estimation and transition modeling, and achieves significant performance improvements across multiple benchmarks.

DetailsMotivation: The key challenge in training VLM agents compared to LLM agents is the shift from textual states to complex visual observations, which introduces partial observability and demands robust world modeling.

Method: Architecturally enforce and reward agent’s reasoning process via RL formulated as POMDP, decomposing reasoning into State Estimation and Transition Modeling, with World Modeling Reward and Bi-Level GAE for turn-aware credit assignment.

Result: A 3B-parameter model achieves score of 0.82 across five diverse agent benchmarks, representing 3× improvement over untrained counterpart (0.21) and outperforming proprietary models like GPT-5 (0.75), Gemini 2.5 Pro (0.67) and Claude 4.5 (0.62).

Conclusion: The optimal belief representation is task-dependent: Natural Language excels at semantic relationships, while Structured formats are essential for precise manipulation. Visual state reasoning enables VLM agents to construct effective internal world models.

Abstract: A key challenge in training Vision-Language Model (VLM) agents, compared to Language Model (LLM) agents, lies in the shift from textual states to complex visual observations. This transition introduces partial observability and demands robust world modeling. We ask: Can VLM agents construct internal world models through explicit visual state reasoning? To address this question, we architecturally enforce and reward the agent’s reasoning process via reinforcement learning (RL), formulating it as a Partially Observable Markov Decision Process (POMDP). We find that decomposing the agent’s reasoning into State Estimation (“what is the current state?”) and Transition Modeling (“what comes next?”) is critical for success, as demonstrated through five reasoning strategies. Our investigation into how agents represent internal beliefs reveals that the optimal representation is task-dependent: Natural Language excels at capturing semantic relationships in general tasks, while Structured formats are indispensable for precise manipulation and control. Building on these insights, we design a World Modeling Reward that provides dense, turn-level supervision for accurate state prediction, and introduce Bi-Level General Advantage Estimation (Bi-Level GAE) for turn-aware credit assignment. Through this form of visual state reasoning, a 3B-parameter model achieves a score of 0.82 across five diverse agent benchmarks, representing a 3$\times$ improvement over its untrained counterpart (0.21) and outperforming proprietary reasoning models such as GPT-5 (0.75), Gemini 2.5 Pro (0.67) and Claude 4.5 (0.62). All experiments are conducted within our VAGEN framework, a scalable system for training and analyzing multi-turn VLM agents in diverse visual environments. Code and data are publicly available at https://vagen-ai.github.io.

[522] A Comparative User Evaluation of XRL Explanations using Goal Identification

Mark Towers, Yali Du, Christopher Freeman, Timothy J. Norman

Main category: cs.AI

TL;DR: The paper evaluates four explainable reinforcement learning (XRL) algorithms for debugging by testing if users can identify an agent’s goal from explanations, finding most perform poorly and user confidence doesn’t correlate with accuracy.

DetailsMotivation: Limited comparative evaluations exist for XRL algorithms in debugging applications, despite it being a core use case.

Method: Proposed evaluation methodology using Atari’s Ms. Pacman environment and four XRL algorithms to test whether users can identify an agent’s goal from decision-making explanations.

Result: Only one XRL algorithm achieved greater than random accuracy for tested goals; users were generally overconfident in their selections; users’ self-reported ease of identification and understanding did not correlate with their accuracy.

Conclusion: Current XRL algorithms have limited effectiveness for debugging purposes, and user confidence is not a reliable indicator of explanation quality.

Abstract: Debugging is a core application of explainable reinforcement learning (XRL) algorithms; however, limited comparative evaluations have been conducted to understand their relative performance. We propose a novel evaluation methodology to test whether users can identify an agent’s goal from an explanation of its decision-making. Utilising the Atari’s Ms. Pacman environment and four XRL algorithms, we find that only one achieved greater than random accuracy for the tested goals and that users were generally overconfident in their selections. Further, we find that users’ self-reported ease of identification and understanding for every explanation did not correlate with their accuracy.

[523] STARK: Strategic Team of Agents for Refining Kernels

Juncheng Dong, Yang Yang, Tao Liu, Yang Wang, Feng Qi, Vahid Tarokh, Kaushik Rangadurai, Shuang Yang

Main category: cs.AI

TL;DR: An LLM agentic framework for GPU kernel optimization that uses multi-agent collaboration and iterative refinement to achieve up to 16x faster runtime performance compared to baseline approaches.

DetailsMotivation: GPU kernel optimization is difficult and labor-intensive due to complex hardware interactions, and existing LLM approaches treat them as single-shot generators without effective exploration of the optimization landscape.

Method: Multi-agent LLM framework with systematic design space exploration through collaboration, grounded instruction, dynamic context management, and strategic search, mimicking expert engineer workflows with hardware trade-off reasoning and profiling feedback.

Result: Achieves substantial improvements over baselines: produces correct solutions where baselines fail, and achieves kernels with up to 16x faster runtime performance on the KernelBench benchmark.

Conclusion: Agentic LLM frameworks show strong potential for advancing fully automated, scalable GPU kernel optimization by enabling systematic exploration and iterative refinement.

Abstract: The efficiency of GPU kernels is central to the progress of modern AI, yet optimizing them remains a difficult and labor-intensive task due to complex interactions between memory hierarchies, thread scheduling, and hardware-specific characteristics. While recent advances in large language models (LLMs) provide new opportunities for automated code generation, existing approaches largely treat LLMs as single-shot generators or naive refinement tools, limiting their effectiveness in navigating the irregular kernel optimization landscape. We introduce an LLM agentic framework for GPU kernel optimization that systematically explores the design space through multi-agent collaboration, grounded instruction, dynamic context management, and strategic search. This framework mimics the workflow of expert engineers, enabling LLMs to reason about hardware trade-offs, incorporate profiling feedback, and refine kernels iteratively. We evaluate our approach on KernelBench, a benchmark for LLM-based kernel optimization, and demonstrate substantial improvements over baseline agents: our system produces correct solutions where baselines often fail, and achieves kernels with up to 16x faster runtime performance. These results highlight the potential of agentic LLM frameworks to advance fully automated, scalable GPU kernel optimization.

[524] ToolCritic: Detecting and Correcting Tool-Use Errors in Dialogue Systems

Hassan Hamad, Yingru Xu, Liang Zhao, Wenbo Yan, Narendra Gyanchandani

Main category: cs.AI

TL;DR: ToolCritic is a diagnostic framework that detects and corrects tool-calling errors in LLMs, improving tool-calling accuracy by up to 13% over baselines.

DetailsMotivation: Tool-augmented LLMs are widely used but tool usage errors reduce their reliability in real-world applications, necessitating better error detection and correction mechanisms.

Method: ToolCritic identifies eight specific tool-calling error types, provides targeted feedback to the main LLM, which then revises its response. A synthetic dataset is created to train ToolCritic.

Result: Experimental results on the Schema-Guided Dialogue dataset show ToolCritic improves tool-calling accuracy by up to 13% compared to zero-shot prompting and self-correction baselines.

Conclusion: ToolCritic represents a significant step toward more robust LLM integration with external tools in real-world dialogue applications.

Abstract: Tool-augmented large language models (LLMs) are increasingly employed in real-world applications, but tool usage errors still hinder their reliability. We introduce ToolCritic, a diagnostic framework that evaluates and improves LLM behavior in multi-turn, tool-augmented dialogues. ToolCritic detects eight distinct error types specific to tool-calling (e.g., premature invocation, argument misalignment, and misinterpretation of tool outputs) and provides targeted feedback to the main LLM. The main LLM, assumed to have strong reasoning, task understanding and orchestration capabilities, then revises its response based on ToolCritic’s feedback. We systematically define these error categories and construct a synthetic dataset to train ToolCritic. Experimental results on the Schema-Guided Dialogue (SGD) dataset demonstrate that ToolCritic improves tool-calling accuracy by up to 13% over baselines, including zero-shot prompting and self-correction techniques. This represents a promising step toward more robust LLM integration with external tools in real-world dialogue applications.

[525] A Brain Cell Type Resource Created by Large Language Models and a Multi-Agent AI System for Collaborative Community Annotation

Rongbin Li, Wenbo Chen, Zhao Li, Rodrigo Munoz-Castaneda, Jinbo Li, Neha S. Maurya, Arnav Solanki, Huan He, Hanwen Xing, Meaghan Ramlakhan, Zachary Wise, Zhuhao Wu, Hua Xu, Michael Hawrylycz, W. Jim Zheng

Main category: cs.AI

TL;DR: BRAINCELL-AID is a multi-agent AI system that combines free-text descriptions with ontology labels to improve gene set annotation accuracy, achieving 77% correct annotations for mouse gene sets and successfully annotating 5,322 brain cell clusters.

DetailsMotivation: Traditional gene set annotation methods like GSEA rely on well-curated annotations and perform poorly with poorly characterized genes. LLMs struggle to represent complex biological knowledge in structured ontologies.

Method: Developed a multi-agent AI system integrating free-text descriptions with ontology labels, using retrieval-augmented generation (RAG) to refine predictions with relevant PubMed literature, reducing hallucinations and enhancing interpretability.

Result: Achieved correct annotations for 77% of mouse gene sets among top predictions. Successfully annotated 5,322 brain cell clusters from mouse brain cell atlas, identifying region-specific gene co-expression patterns and functional roles of gene ensembles.

Conclusion: BRAINCELL-AID creates a valuable resource for community-driven cell type annotation, enabling novel insights into brain cell function and identifying neurologically meaningful descriptions for Basal Ganglia-related cell types.

Abstract: Single-cell RNA sequencing has transformed our ability to identify diverse cell types and their transcriptomic signatures. However, annotating these signatures-especially those involving poorly characterized genes-remains a major challenge. Traditional methods, such as Gene Set Enrichment Analysis (GSEA), depend on well-curated annotations and often perform poorly in these contexts. Large Language Models (LLMs) offer a promising alternative but struggle to represent complex biological knowledge within structured ontologies. To address this, we present BRAINCELL-AID (BRAINCELL-AID: https://biodataai.uth.edu/BRAINCELL-AID), a novel multi-agent AI system that integrates free-text descriptions with ontology labels to enable more accurate and robust gene set annotation. By incorporating retrieval-augmented generation (RAG), we developed a robust agentic workflow that refines predictions using relevant PubMed literature, reducing hallucinations and enhancing interpretability. Using this workflow, we achieved correct annotations for 77% of mouse gene sets among their top predictions. Applying this approach, we annotated 5,322 brain cell clusters from the comprehensive mouse brain cell atlas generated by the BRAIN Initiative Cell Census Network, enabling novel insights into brain cell function by identifying region-specific gene co-expression patterns and inferring functional roles of gene ensembles. BRAINCELL-AID also identifies Basal Ganglia-related cell types with neurologically meaningful descriptions. Hence, we create a valuable resource to support community-driven cell type annotation.

[526] Structured Debate Improves Corporate Credit Reasoning in Financial AI

Yoonjin Lee, Munhee Kim, Hanbi Choi, Juhyeon Park, Seungho Lyoo, Woojin Park

Main category: cs.AI

TL;DR: This paper develops two LLM-based systems for automated evidence-based reasoning in corporate credit assessment, showing that structured multi-agent debate systems outperform single-agent approaches in reasoning quality and interpretability.

DetailsMotivation: Current financial AI focuses on numerical prediction but lacks support for interpretive judgments in loan evaluation, particularly for qualitative non-financial indicators that resist formalization but influence loan outcomes.

Method: Developed two LLM-based systems: a single-agent system (NAS) with bidirectional analysis, and a debate-based multi-agent system (KPD-MADS) using Karl Popper’s critical dialogue framework with 10-step structured interaction protocol. Both were tested on real corporate cases.

Result: Both systems achieved substantial productivity gains (NAS: 11.55s, KPD-MADS: 91.97s vs human baseline: 1920s). KPD-MADS showed superior reasoning quality with higher ratings in explanatory adequacy (4.0 vs 3.0), practical applicability (4.0 vs 3.0), and usability (62.5 vs 52.5).

Conclusion: Structured multi-agent interaction enhances reasoning rigor and interpretability in financial AI, advancing scalable and defensible automation in corporate credit assessment.

Abstract: Despite advances in financial AI, the automation of evidence-based reasoning remains unresolved in corporate credit assessment, where qualitative non-financial indicators exert decisive influence on loan repayment outcomes yet resist formalization. Existing approaches focus predominantly on numerical prediction and provide limited support for the interpretive judgments required in professional loan evaluation. This study develops and evaluates two operational large language model (LLM)-based systems designed to generate structured reasoning from non-financial evidence. The first is a non-adversarial single-agent system (NAS) that produces bidirectional analysis through a single-pass reasoning pipeline. The second is a debate-based multi-agent system (KPD-MADS) that operationalizes adversarial verification through a ten-step structured interaction protocol grounded in Karl Popper’s critical dialogue framework. Both systems were applied to three real corporate cases and evaluated by experienced credit risk professionals. Compared to manual expert reporting, both systems achieved substantial productivity gains (NAS: 11.55 s per case; KPD-MADS: 91.97 s; human baseline: 1920 s). The KPD-MADS demonstrated superior reasoning quality, receiving higher median ratings in explanatory adequacy (4.0 vs. 3.0), practical applicability (4.0 vs. 3.0), and usability (62.5 vs. 52.5). These findings show that structured multi-agent interaction can enhance reasoning rigor and interpretability in financial AI, advancing scalable and defensible automation in corporate credit assessment.

[527] Enhanced Fish Freshness Classification with Incremental Handcrafted Feature Fusion

Phi-Hung Hoang, Nam-Thuan Trinh, Van-Manh Tran, Thi-Thu-Hong Phan

Main category: cs.AI

TL;DR: A handcrafted feature-based approach using color statistics, histograms, and texture features from fish eye images achieves high accuracy in automated fish freshness assessment, outperforming previous deep learning methods.

DetailsMotivation: Conventional sensory evaluation of fish freshness is subjective, inconsistent, and difficult to standardize, requiring an objective automated solution for quality control and consumer safety.

Method: Systematically extracts and fuses complementary descriptors including color statistics, histograms across multiple color spaces, and texture features (LBP, GLCM) from fish eye images, capturing both global chromatic variations and localized degradations.

Result: LightGBM classifier achieved 77.56% accuracy (14.35% improvement over previous baseline), and ANN with augmented data reached 97.16% accuracy (19.86% improvement over prior best).

Conclusion: Carefully engineered handcrafted features provide a robust, interpretable, and reliable solution for automated fish freshness assessment, offering practical value for food quality monitoring applications.

Abstract: Accurate assessment of fish freshness remains a major challenge in the food industry, with direct consequences for product quality, market value, and consumer health. Conventional sensory evaluation is inherently subjective, inconsistent, and difficult to standardize across contexts, often limited by subtle, species-dependent spoilage cues. To address these limitations, we propose a handcrafted feature-based approach that systematically extracts and incrementally fuses complementary descriptors, including color statistics, histograms across multiple color spaces, and texture features such as Local Binary Patterns (LBP) and Gray-Level Co-occurrence Matrices (GLCM), from fish eye images. Our method captures global chromatic variations from full images and localized degradations from ROI segments, fusing each independently to evaluate their effectiveness in assessing freshness. Experiments on the Freshness of the Fish Eyes (FFE) dataset demonstrate the approach’s effectiveness: in a standard train-test setting, a LightGBM classifier achieved 77.56% accuracy, a 14.35% improvement over the previous deep learning baseline of 63.21%. With augmented data, an Artificial Neural Network (ANN) reached 97.16% accuracy, surpassing the prior best of 77.3% by 19.86%. These results demonstrate that carefully engineered, handcrafted features, when strategically processed, yield a robust, interpretable, and reliable solution for automated fish freshness assessment, providing valuable insights for practical applications in food quality monitoring.

[528] Physics-Informed Large Language Models for HVAC Anomaly Detection with Autonomous Rule Generation

Subin Lin, Chuanbo Hua

Main category: cs.AI

TL;DR: PILLM is a Physics-Informed LLM framework that uses evolutionary loops to generate, evaluate, and refine HVAC anomaly detection rules with embedded physical constraints, achieving state-of-the-art performance while maintaining interpretability.

DetailsMotivation: HVAC systems consume significant energy globally, but existing anomaly detection methods either lack adaptability (rule-based) or transparency (deep learning). LLM approaches improve interpretability but ignore physical principles governing HVAC operations.

Method: PILLM operates within an evolutionary loop with physics-informed reflection and crossover operators that embed thermodynamic and control-theoretic constraints to generate physically grounded anomaly detection rules.

Result: Experiments on the Building Fault Detection dataset show PILLM achieves state-of-the-art performance while producing interpretable and actionable diagnostic rules.

Conclusion: PILLM advances trustworthy and deployable AI for smart building systems by combining LLM interpretability with physical plausibility through evolutionary rule generation.

Abstract: Heating, Ventilation, and Air-Conditioning (HVAC) systems account for a substantial share of global building energy use, making reliable anomaly detection essential for improving efficiency and reducing emissions. Classical rule-based approaches offer explainability but lack adaptability, while deep learning methods provide predictive power at the cost of transparency, efficiency, and physical plausibility. Recent attempts to use Large Language Models (LLMs) for anomaly detection improve interpretability but largely ignore the physical principles that govern HVAC operations. We present PILLM, a Physics-Informed LLM framework that operates within an evolutionary loop to automatically generate, evaluate, and refine anomaly detection rules. Our approach introduces physics-informed reflection and crossover operators that embed thermodynamic and control-theoretic constraints, enabling rules that are both adaptive and physically grounded. Experiments on the public Building Fault Detection dataset show that PILLM achieves state-of-the-art performance while producing diagnostic rules that are interpretable and actionable, advancing trustworthy and deployable AI for smart building systems.

[529] Which LLM Multi-Agent Protocol to Choose?

Hongyi Du, Jiaqi Su, Jisen Li, Lijie Ding, Yingxuan Yang, Peixuan Han, Xiangru Tang, Kunlun Zhu, Jiaxuan You

Main category: cs.AI

TL;DR: ProtocolBench is a benchmark for evaluating multi-agent communication protocols, showing significant performance differences across protocols. ProtocolRouter is a learnable router that dynamically selects optimal protocols, improving performance and reliability.

DetailsMotivation: Current protocol selection in multi-agent systems is intuition-driven without standardized evaluation, despite protocols being critical for performance and reliability.

Method: Created ProtocolBench benchmark with four evaluation axes (task success, latency, overhead, robustness), and developed ProtocolRouter - a learnable protocol router that selects protocols based on requirements and runtime signals.

Result: Protocol choice significantly impacts system behavior: 36.5% completion time variation, 3.48s latency differences, and varying resilience. ProtocolRouter reduces Fail-Storm recovery time by 18.1% and improves success rates in GAIA scenarios.

Conclusion: Protocol selection matters significantly in multi-agent systems. ProtocolBench enables systematic protocol evaluation, and ProtocolRouter demonstrates the value of dynamic protocol selection for improved performance and reliability at scale.

Abstract: As large-scale multi-agent systems evolve, the communication protocol layer has become a critical yet under-evaluated factor shaping performance and reliability. Despite the existence of diverse protocols (A2A, ACP, ANP, Agora, etc.), selection is often intuition-driven and lacks standardized guidance. We introduce ProtocolBench, a benchmark that systematically compares agent protocols along four measurable axes: task success, end-to-end latency, message or byte overhead, and robustness under failures. On ProtocolBench, protocol choice significantly influences system behavior. In the Streaming Queue scenario, overall completion time varies by up to 36.5% across protocols, and mean end-to-end latency differs by 3.48 s. Under Fail-Storm Recovery, resilience also differs consistently across protocols. Beyond evaluation, we present ProtocolRouter, a learnable protocol router that selects per-scenario (or per-module) protocols from requirement and runtime signals. ProtocolRouter reduces Fail-Storm recovery time by up to 18.1% versus the best single-protocol baseline, and achieves scenario-specific gains such as higher success in GAIA. We also release ProtocolRouterBench to standardize protocol evaluation and improve reliability at scale.

[530] Combining ECG Foundation Model and XGBoost to Predict In-Hospital Malignant Ventricular Arrhythmias in AMI Patients

Shun Huang, Wenlu Xing, Shijia Geng, Hailong Wang, Guangkun Nie, Gongzheng Tang, Chenyang He, Shenda Hong

Main category: cs.AI

TL;DR: A hybrid framework combining ECG foundation model with XGBoost classifier improves VT/VF risk prediction after heart attack, achieving better accuracy than traditional methods while maintaining interpretability through SHAP analysis.

DetailsMotivation: Malignant ventricular arrhythmias following acute myocardial infarction are a major cause of in-hospital death, but traditional risk scores have limited performance and deep learning models lack interpretability needed for clinical trust.

Method: Used ECGFounder foundation model to extract 150-dimensional diagnostic probability features from ECG recordings, then applied feature selection and trained XGBoost classifier on these features. SHAP method was used for interpretability analysis.

Result: The hybrid model achieved AUC of 0.801, outperforming KNN (0.677), RNN (0.676), and 1D-CNN (0.720). SHAP analysis revealed clinically meaningful features like premature ventricular complexes as risk predictors and normal sinus rhythm as protective factors.

Conclusion: The hybrid framework provides a novel paradigm for VT/VF risk prediction by validating foundation model outputs as effective automated feature engineering for building trustworthy, explainable AI-based clinical decision support systems.

Abstract: Malignant ventricular arrhythmias (VT/VF) following acute myocardial infarction (AMI) are a major cause of in-hospital death, yet early identification remains a clinical challenge. While traditional risk scores have limited performance, end-to-end deep learning models often lack the interpretability needed for clinical trust. This study aimed to develop a hybrid predictive framework that integrates a large-scale electrocardiogram (ECG) foundation model (ECGFounder) with an interpretable XGBoost classifier to improve both accuracy and interpretability. We analyzed 6,634 ECG recordings from AMI patients, among whom 175 experienced in-hospital VT/VF. The ECGFounder model was used to extract 150-dimensional diagnostic probability features , which were then refined through feature selection to train the XGBoost classifier. Model performance was evaluated using AUC and F1-score , and the SHAP method was used for interpretability. The ECGFounder + XGBoost hybrid model achieved an AUC of 0.801 , outperforming KNN (AUC 0.677), RNN (AUC 0.676), and an end-to-end 1D-CNN (AUC 0.720). SHAP analysis revealed that model-identified key features, such as “premature ventricular complexes” (risk predictor) and “normal sinus rhythm” (protective factor), were highly consistent with clinical knowledge. We conclude that this hybrid framework provides a novel paradigm for VT/VF risk prediction by validating the use of foundation model outputs as effective, automated feature engineering for building trustworthy, explainable AI-based clinical decision support systems.

[531] Offline Policy Evaluation of Multi-Turn LLM Health Coaching with Real Users

Melik Ozolcer, Sang Won Bae

Main category: cs.AI

TL;DR: Study of a web-deployed LLM health coach shows uniform heavy-tool policies harm low-health-literacy users, while early information-gain bonuses improve trait identification and goal success.

DetailsMotivation: To evaluate real-world performance of tool-augmented LLM health coaches and understand how different policies affect user subgroups, particularly identifying potential harms that average metrics might obscure.

Method: Used offline policy evaluation (OPE) over factorized decision heads (Tool/Style) with 7 users (280 rated turns), and a lightweight simulator with hidden archetypes to test early information-gain bonuses.

Result: Uniform heavy-tool policies raise average value but harm low-health-literacy/high-self-efficacy users. Adding early information-gain bonuses shortens trait identification and improves goal success and pass@3 metrics.

Conclusion: Proposes an evaluation-first path to personalization: freeze the generator, learn subgroup-aware decision heads on typed rewards, and always report per-archetype metrics to surface subgroup harms.

Abstract: We study a web-deployed, tool-augmented LLM health coach with real users. In a pilot with seven users (280 rated turns), offline policy evaluation (OPE) over factorized decision heads (Tool/Style) shows that a uniform heavy-tool policy raises average value on logs but harms specific subgroups, most notably low-health-literacy/high-self-efficacy users. A lightweight simulator with hidden archetypes further shows that adding a small early information-gain bonus reliably shortens trait identification and improves goal success and pass@3. Together, these early findings indicate an evaluation-first path to personalization: freeze the generator, learn subgroup-aware decision heads on typed rewards (objective tool outcomes and satisfaction), and always report per-archetype metrics to surface subgroup harms that averages obscure.

[532] Temporally Detailed Hypergraph Neural ODEs for Type 2 Diabetes Progression Modeling

Tingsong Xiao, Yao An Lee, Zelin Xu, Yupu Zhang, Zibo Liu, Yu Huang, Jiang Bian, Serena Jingchuan Guo, Zhe Jiang

Main category: cs.AI

TL;DR: TD-HNODE is a neural ODE framework that models disease progression using temporally detailed hypergraphs to capture continuous-time dynamics and interdependencies between disease complications.

DetailsMotivation: Accurate disease progression modeling from irregular EHR data is challenging due to patient heterogeneity and the need to capture complex continuous-time dynamics, which existing methods fail to address adequately.

Method: Proposes TD-HNODE which represents disease progression as a temporally detailed hypergraph and learns continuous-time dynamics via neural ODEs with a learnable hypergraph Laplacian capturing interdependencies within and between progression trajectories.

Result: Experiments on two real-world clinical datasets show TD-HNODE outperforms multiple baselines in modeling type 2 diabetes and cardiovascular disease progression.

Conclusion: TD-HNODE effectively addresses limitations of existing methods by capturing complex continuous-time progression dynamics through hypergraph neural ODEs, enabling better patient sub-phenotyping and intervention planning.

Abstract: Disease progression modeling aims to characterize and predict how a patient’s disease complications worsen over time based on longitudinal electronic health records (EHRs). Accurate modeling of disease progression, such as type 2 diabetes, can enhance patient sub-phenotyping and inform effective and timely interventions. However, the problem is challenging due to the need to learn continuous-time dynamics of progression patterns based on irregular-time event samples and patient heterogeneity (\eg different progression rates and pathways). Existing mechanistic and data-driven methods either lack adaptability to learn from real-world data or fail to capture complex continuous-time dynamics on progression trajectories. To address these limitations, we propose Temporally Detailed Hypergraph Neural Ordinary Differential Equation (TD-HNODE), which represents disease progression on clinically recognized trajectories as a temporally detailed hypergraph and learns the continuous-time progression dynamics via a neural ODE framework. TD-HNODE contains a learnable TD-Hypergraph Laplacian that captures the interdependency of disease complication markers within both intra- and inter-progression trajectories. Experiments on two real-world clinical datasets demonstrate that TD-HNODE outperforms multiple baselines in modeling the progression of type 2 diabetes and related cardiovascular diseases.

[533] Coinvisor: An RL-Enhanced Chatbot Agent for Interactive Cryptocurrency Investment Analysis

Chong Chen, Ze Liu, Lingfeng Bao, Yanlin Wang, Ting Chen, Daoyuan Wu, Jiachi Chen

Main category: cs.AI

TL;DR: Coinvisor is a reinforcement learning-based chatbot that provides comprehensive analytical support for cryptocurrency investment through a multi-agent framework with adaptive tool selection.

DetailsMotivation: Address limitations in current cryptocurrency investment approaches: manual analysis is time-consuming and biased, data platforms lack depth, and LLM agents lack real-time data integration and multi-step reasoning.

Method: Multi-agent framework with reinforcement learning-based tool selection mechanism that enables multi-step planning and flexible integration of diverse data sources for real-time interaction and adaptive analysis.

Result: Improves recall by 40.7% and F1 score by 26.6% over base model in tool orchestration. User studies show high satisfaction (4.64/5) with preference over general LLMs and existing crypto platforms (4.62/5).

Conclusion: Coinvisor successfully addresses current limitations in cryptocurrency investment analysis through its RL-based multi-agent framework, delivering accurate and actionable insights with high user satisfaction.

Abstract: The cryptocurrency market offers significant investment opportunities but faces challenges including high volatility and fragmented information. Data integration and analysis are essential for informed investment decisions. Currently, investors use three main approaches: (1) Manual analysis across various sources, which depends heavily on individual experience and is time-consuming and prone to bias; (2) Data aggregation platforms-limited in functionality and depth of analysis; (3) Large language model agents-based on static pretrained models, lacking real-time data integration and multi-step reasoning capabilities. To address these limitations, we present Coinvisor, a reinforcement learning-based chatbot that provides comprehensive analytical support for cryptocurrency investment through a multi-agent framework. Coinvisor integrates diverse analytical capabilities through specialized tools. Its key innovation is a reinforcement learning-based tool selection mechanism that enables multi-step planning and flexible integration of diverse data sources. This design supports real-time interaction and adaptive analysis of dynamic content, delivering accurate and actionable investment insights. We evaluated Coinvisor through automated benchmarks on tool calling accuracy and user studies with 20 cryptocurrency investors using our interface. Results show that Coinvisor improves recall by 40.7% and F1 score by 26.6% over the base model in tool orchestration. User studies show high satisfaction (4.64/5), with participants preferring Coinvisor to both general LLMs and existing crypto platforms (4.62/5).

[534] RubiSCoT: A Framework for AI-Supported Academic Assessment

Thorsten Fröhlich, Tim Schlippe

Main category: cs.AI

TL;DR: RubiSCoT is an AI framework that enhances thesis evaluation using NLP techniques like LLMs and retrieval-augmented generation to provide consistent, scalable assessment from proposal to final submission.

DetailsMotivation: Traditional thesis evaluation methods are time-consuming and subject to evaluator variability, creating a need for more consistent and scalable assessment solutions.

Method: Uses advanced NLP techniques including large language models, retrieval-augmented generation, and structured chain-of-thought prompting for preliminary assessments, multidimensional assessments, content extraction, rubric-based scoring, and detailed reporting.

Result: The paper presents the design and implementation of RubiSCoT framework, demonstrating its potential for academic assessment optimization.

Conclusion: RubiSCoT offers a consistent, scalable, and transparent solution to optimize academic thesis evaluation processes.

Abstract: The evaluation of academic theses is a cornerstone of higher education, ensuring rigor and integrity. Traditional methods, though effective, are time-consuming and subject to evaluator variability. This paper presents RubiSCoT, an AI-supported framework designed to enhance thesis evaluation from proposal to final submission. Using advanced natural language processing techniques, including large language models, retrieval-augmented generation, and structured chain-of-thought prompting, RubiSCoT offers a consistent, scalable solution. The framework includes preliminary assessments, multidimensional assessments, content extraction, rubric-based scoring, and detailed reporting. We present the design and implementation of RubiSCoT, discussing its potential to optimize academic assessment processes through consistent, scalable, and transparent evaluation.

[535] Active Inference for an Intelligent Agent in Autonomous Reconnaissance Missions

Johan Schubert, Farzad Kamrani, Tove Gustavi

Main category: cs.AI

TL;DR: Developed an active inference route-planning method using Dempster-Shafer theory and Gaussian sensor models to balance exploration and exploitation for autonomous agents maintaining operational pictures.

DetailsMotivation: To enable autonomous agents to reconnoiter geographical areas and maintain common operational pictures by balancing exploration of unknown areas with tracking of identified targets.

Method: Constructs evidence maps incorporating positive/negative sensor observations, uses Dempster-Shafer theory for generative model, Bayesian approach for posterior updates, and variational free energy minimization for route planning.

Result: The method successfully directs agent movements by minimizing free energy, allowing effective balancing between extensive area exploration and target object tracking.

Conclusion: Active inference with Dempster-Shafer theory provides an effective framework for autonomous route planning that addresses the exploration-exploitation trade-off in geographical reconnaissance.

Abstract: We develop an active inference route-planning method for the autonomous control of intelligent agents. The aim is to reconnoiter a geographical area to maintain a common operational picture. To achieve this, we construct an evidence map that reflects our current understanding of the situation, incorporating both positive and “negative” sensor observations of possible target objects collected over time, and diffusing the evidence across the map as time progresses. The generative model of active inference uses Dempster-Shafer theory and a Gaussian sensor model, which provides input to the agent. The generative process employs a Bayesian approach to update a posterior probability distribution. We calculate the variational free energy for all positions within the area by assessing the divergence between a pignistic probability distribution of the evidence map and a posterior probability distribution of a target object based on the observations, including the level of surprise associated with receiving new observations. Using the free energy, we direct the agents’ movements in a simulation by taking an incremental step toward a position that minimizes the free energy. This approach addresses the challenge of exploration and exploitation, allowing agents to balance searching extensive areas of the geographical map while tracking identified target objects.

[536] Label Indeterminacy in AI & Law

Cor Steging, Tadeusz Zbiegień

Main category: cs.AI

TL;DR: Legal machine learning faces label indeterminacy because case outcomes are shaped by human interventions like settlements and appeals, making ground truth ambiguous. The paper demonstrates how different label construction methods affect model behavior in European Court of Human Rights case classification.

DetailsMotivation: Machine learning in law typically treats past case outcomes as ground truth, but these outcomes are often shaped by human interventions (settlements, appeals, procedural actions) that create label indeterminacy - the outcome could have been different without these interventions.

Method: The paper examines how different label construction methods during training affect model behavior, using European Court of Human Rights case classification as the context. It discusses existing methods for imputing indeterminate labels but notes they rely on unverifiable assumptions.

Result: The study shows that the way labels are constructed during training can significantly affect model behavior, demonstrating that label indeterminacy shapes how models perform in legal classification tasks.

Conclusion: Label indeterminacy is a relevant concern in AI & Law that needs to be accounted for in legal machine learning applications, as it can significantly shape model behavior and outcomes.

Abstract: Machine learning is increasingly used in the legal domain, where it typically operates retrospectively by treating past case outcomes as ground truth. However, legal outcomes are often shaped by human interventions that are not captured in most machine learning approaches. A final decision may result from a settlement, an appeal, or other procedural actions. This creates label indeterminacy: the outcome could have been different if the intervention had or had not taken place. We argue that legal machine learning applications need to account for label indeterminacy. Methods exist that can impute these indeterminate labels, but they are all grounded in unverifiable assumptions. In the context of classifying cases from the European Court of Human Rights, we show that the way that labels are constructed during training can significantly affect model behaviour. We therefore position label indeterminacy as a relevant concern in AI & Law and demonstrate how it can shape model behaviour.

[537] MIRAGE: Agentic Framework for Multimodal Misinformation Detection with Web-Grounded Reasoning

Mir Nafis Sharear Shopnil, Sharad Duwal, Abhishek Tyagi, Adiba Mahbub Proma

Main category: cs.AI

TL;DR: MIRAGE is an agentic framework for multimodal misinformation detection that decomposes verification into four modules: visual veracity assessment, cross-modal consistency analysis, retrieval-augmented factual checking, and calibrated judgment, achieving 81.65% F1 without domain-specific training.

DetailsMotivation: Manual fact-checking is overwhelmed by billions of daily multimodal posts, and supervised detection models fail to generalize across diverse manipulation tactics due to domain-specific training requirements.

Method: MIRAGE uses an inference-time, model-pluggable framework with four sequential modules: visual veracity assessment for AI-generated images, cross-modal consistency analysis for out-of-context repurposing, retrieval-augmented factual checking with iterative question generation, and a calibrated judgment module that integrates all signals.

Result: On MMFakeBench validation set, MIRAGE with GPT-4o-mini achieved 81.65% F1 and 75.1% accuracy, outperforming GPT-4V with MMD-Agent by 7.65 F1 points while maintaining 34.3% false positive rate versus 97.3% for judge-only baseline. Test set results confirmed generalization with 81.44% F1 and 75.08% accuracy.

Conclusion: Decomposed agentic reasoning with web retrieval can match supervised detector performance without domain-specific training, enabling effective misinformation detection across modalities where labeled data remains scarce.

Abstract: Misinformation spreads across web platforms through billions of daily multimodal posts that combine text and images, overwhelming manual fact-checking capacity. Supervised detection models require domain-specific training data and fail to generalize across diverse manipulation tactics. We present MIRAGE, an inference-time, model-pluggable agentic framework that decomposes multimodal verification into four sequential modules: visual veracity assessment detects AI-generated images, cross-modal consistency analysis identifies out-of-context repurposing, retrieval-augmented factual checking grounds claims in web evidence through iterative question generation, and a calibrated judgment module integrates all signals. MIRAGE orchestrates vision-language model reasoning with targeted web retrieval, outputs structured and citation-linked rationales. On MMFakeBench validation set (1,000 samples), MIRAGE with GPT-4o-mini achieves 81.65% F1 and 75.1% accuracy, outperforming the strongest zero-shot baseline (GPT-4V with MMD-Agent at 74.0% F1) by 7.65 points while maintaining 34.3% false positive rate versus 97.3% for a judge-only baseline. Test set results (5,000 samples) confirm generalization with 81.44% F1 and 75.08% accuracy. Ablation studies show visual verification contributes 5.18 F1 points and retrieval-augmented reasoning contributes 2.97 points. Our results demonstrate that decomposed agentic reasoning with web retrieval can match supervised detector performance without domain-specific training, enabling misinformation detection across modalities where labeled data remains scarce.

[538] Reasoning Distillation and Structural Alignment for Improved Code Generation

Amir Jalilifard, Anderson de Rezende Rocha, Marcos Medeiros Raimundo

Main category: cs.AI

TL;DR: This paper presents a method to distill reasoning capabilities from very large language models (VLLMs) into smaller, more efficient models for code generation, achieving better performance than baseline models on standard benchmarks.

DetailsMotivation: Smaller language models often lack the reasoning capabilities needed for effective code generation, which requires understanding solution-level structures rather than just token prediction. The goal is to create smaller, faster, and cheaper models that can match VLLM reasoning abilities.

Method: The approach trains smaller models to emulate VLLM reasoning through structure-aware loss optimization, establishing structural correspondence between problems and solutions, and learning correct solution pathways.

Result: The fine-tuned model significantly outperforms baseline models in pass@1, average data flow, and average syntax match metrics across MBPP, MBPP Plus, and HumanEval benchmarks.

Conclusion: Reasoning capabilities can be effectively distilled from VLLMs into smaller models through a simple and cheap process, enabling efficient deployment while maintaining strong code generation performance.

Abstract: Effective code generation with language models hinges on two critical factors: accurately understanding the intent of the prompt and generating code that applies algorithmic reasoning to produce correct solutions capable of passing diverse test cases while adhering to the syntax of the target programming language. Unlike other language tasks, code generation requires more than accurate token prediction; it demands comprehension of solution-level and structural relationships rather than merely generating the most likely tokens. very large language model (VLLM) are capable of generating detailed steps toward the correct solution of complex tasks where reasoning is crucial in solving the problem. Such reasoning capabilities may be absent in smaller language models. Therefore, in this work, we distill the reasoning capabilities of a VLLM into a smaller, more efficient model that is faster and cheaper to deploy. Our approach trains the model to emulate the reasoning and problem-solving abilities of the VLLM by learning to identify correct solution pathways and establishing a structural correspondence between problem definitions and potential solutions through a novel method of structure-aware loss optimization. This enables the model to transcend token-level generation and to deeply grasp the overarching structure of solutions for given problems. Experimental results show that our fine-tuned model, developed through a cheap and simple to implement process, significantly outperforms our baseline model in terms of pass@1, average data flow, and average syntax match metrics across the MBPP, MBPP Plus, and HumanEval benchmarks.

[539] OG-Rank: Learning to Rank Fast and Slow with Uncertainty and Reward-Trend Guided Adaptive Exploration

Praphul Singh, Corey Barrett, Sumana Srivasta, Irfan Bulu, Sri Gadde, Krishnaram Kenthapadi

Main category: cs.AI

TL;DR: OG-Rank is a low-latency decoder-based reranker that scores all candidates in one pass and generates explanations only when the ranking is ambiguous, achieving strong performance on clinical order selection tasks.

DetailsMotivation: Clinicians need real-time ranking systems that can justify their choices, requiring low-latency decoder-based rerankers that provide explanations when needed.

Method: Single-decoder approach combining pooled first-token scoring with uncertainty-gated explanation generation, trained with curriculum learning focused on hard cases.

Result: Achieves Recall@10.45 and nDCG@200.625 on fast path, improving to Recall@10.56 and nDCG@200.699 when explanation gate activates at 45% rate. Outperforms encoder baselines in effectiveness and flexibility.

Conclusion: Practical approach: rank fast by default and explain when helpful, with single-policy design simplifying deployment and curriculum learning principle transferable beyond clinical applications.

Abstract: Clinicians need ranking systems that work in real time and still justify their choices. Motivated by the need for a low-latency, decoder-based reranker, we present OG-Rank, a single-decoder approach that pairs a pooled first-token scoring signal with an uncertainty-gated explanation step. The model scores all candidates in one pass and generates a brief, structured rationale only when the list is genuinely ambiguous, keeping latency predictable. Trained with a curriculum that concentrates effort on hard cases, OG-Rank delivers strong effectiveness on encounter-scoped order selection (fast path: Recall@10.45, nDCG@200.625) and improves further when the gate activates (Recall@10.56, nDCG@200.699 at a 45% gate rate), while compact backbones show similar gains under the same policy. Encoder baselines trail in both effectiveness and flexibility. The result is a practical recipe: rank fast by default and explain when it helps, a pattern that applies broadly to decision tasks where selective generation buys accuracy at acceptable cost. The single-policy design simplifies deployment and budget planning, and the curriculum principle (spend more on the hard cases, less on the easy ones) readily transfers beyond clinical order selection.

[540] LLM-as-a-Prophet: Understanding Predictive Intelligence with Prophet Arena

Qingchuan Yang, Simon Mahns, Sida Li, Anri Gu, Jibang Wu, Haifeng Xu

Main category: cs.AI

TL;DR: This paper systematically evaluates LLMs’ forecasting capabilities on real-world events, finding they show promise but have key limitations in event recall, data understanding, and information aggregation speed.

DetailsMotivation: To investigate whether large language models can effectively forecast real-world future events, given their training on Internet-scale data and potential applications in finance and economics.

Method: Built Prophet Arena - a benchmark that continuously collects live forecasting tasks and decomposes them into pipeline stages for controlled, large-scale experimentation.

Result: LLMs demonstrate impressive forecasting capabilities with small calibration errors, consistent prediction confidence, and promising market returns. However, they struggle with inaccurate event recalls, misunderstanding data sources, and slower information aggregation compared to markets.

Conclusion: While LLMs show promise as forecasting tools (LLM-as-a-Prophet), key bottlenecks remain that need to be addressed for achieving superior predictive intelligence.

Abstract: Forecasting is not only a fundamental intellectual pursuit but also is of significant importance to societal systems such as finance and economics. With the rapid advances of large language models (LLMs) trained on Internet-scale data, it raises the promise of employing LLMs to forecast real-world future events, an emerging paradigm we call “LLM-as-a-Prophet”. This paper systematically investigates such predictive intelligence of LLMs. To this end, we build Prophet Arena, a general evaluation benchmark that continuously collects live forecasting tasks and decomposes each task into distinct pipeline stages, in order to support our controlled and large-scale experimentation. Our comprehensive evaluation reveals that many LLMs already exhibit impressive forecasting capabilities, reflected in, e.g., their small calibration errors, consistent prediction confidence and promising market returns. However, we also uncover key bottlenecks towards achieving superior predictive intelligence via LLM-as-a-Prophet, such as LLMs’ inaccurate event recalls, misunderstanding of data sources and slower information aggregation compared to markets when resolution nears.

[541] Contextual Attention Modulation: Towards Efficient Multi-Task Adaptation in Large Language Models

Dayan Pan, Zhaoyang Fu, Jingyuan Wang, Xiao Han, Yue Zhu, Xiangyu Zhao

Main category: cs.AI

TL;DR: Proposes Contextual Attention Modulation (CAM) and Hybrid CAM (HyCAM) framework for multi-task adaptation in LLMs, addressing catastrophic forgetting and resource issues while improving performance by 3.65% on average.

DetailsMotivation: LLMs struggle with multi-task adaptation, balancing knowledge retention with task specialization. Conventional fine-tuning causes catastrophic forgetting and high resource consumption, while existing parameter-efficient methods perform poorly in complex multi-task scenarios.

Method: CAM dynamically modulates self-attention representations to enhance task-specific features while preserving general knowledge. HyCAM combines shared full-parameter CAM with multiple lightweight specialized CAM modules, using dynamic routing for adaptive knowledge fusion.

Result: Extensive experiments on heterogeneous tasks (question answering, code generation, logical reasoning) show significant performance improvements, achieving average 3.65% gain over existing approaches.

Conclusion: The proposed CAM and HyCAM framework effectively addresses multi-task adaptation challenges in LLMs, providing better performance while mitigating catastrophic forgetting and resource consumption issues.

Abstract: Large Language Models (LLMs) possess remarkable generalization capabilities but struggle with multi-task adaptation, particularly in balancing knowledge retention with task-specific specialization. Conventional fine-tuning methods suffer from catastrophic forgetting and substantial resource consumption, while existing parameter-efficient methods perform suboptimally in complex multi-task scenarios. To address this, we propose Contextual Attention Modulation (CAM), a novel mechanism that dynamically modulates the representations of self-attention modules in LLMs. CAM enhances task-specific features while preserving general knowledge, thereby facilitating more effective and efficient adaptation. For effective multi-task adaptation, CAM is integrated into our Hybrid Contextual Attention Modulation (HyCAM) framework, which combines a shared, full-parameter CAM module with multiple specialized, lightweight CAM modules, enhanced by a dynamic routing strategy for adaptive knowledge fusion. Extensive experiments on heterogeneous tasks, including question answering, code generation, and logical reasoning, demonstrate that our approach significantly outperforms existing approaches, achieving an average performance improvement of 3.65%. The implemented code and data are available to ease reproducibility at https://github.com/Applied-Machine-Learning-Lab/HyCAM.

[542] Seeing but Not Believing: Probing the Disconnect Between Visual Attention and Answer Correctness in VLMs

Zhining Liu, Ziyi Chen, Hui Liu, Chen Luo, Xianfeng Tang, Suhang Wang, Joy Zeng, Zhenwei Dai, Zhan Shi, Tianxin Wei, Benoit Dumoulin, Hanghang Tong

Main category: cs.AI

TL;DR: VLMs often perceive visual evidence but fail to use it effectively, a phenomenon called “seeing but not believing”. An inference-time intervention that highlights evidence regions improves accuracy across multiple VLM families.

DetailsMotivation: To understand why VLMs fail on multimodal tasks even when correct visual evidence is present, and determine if failures arise from perception issues or utilization problems.

Method: Examined layer-wise attention dynamics and introduced inference-time intervention using selective attention-based masking to highlight deep-layer evidence regions without training.

Result: Found that shallow layers focus on text while deeper layers reliably attend to evidence regions. The intervention consistently improved accuracy across LLaVA, Qwen, Gemma, and InternVL models.

Conclusion: VLMs encode reliable evidence internally but under-utilize it; making these signals explicit can bridge perception-reasoning gaps and improve VLM reliability.

Abstract: Vision-Language Models (VLMs) achieve strong results on multimodal tasks such as visual question answering, yet they can still fail even when the correct visual evidence is present. In this work, we systematically investigate whether these failures arise from not perceiving the evidence or from not leveraging it effectively. By examining layer-wise attention dynamics, we find that shallow layers focus primarily on text, while deeper layers sparsely but reliably attend to localized evidence regions. Surprisingly, VLMs often perceive the visual evidence when outputting incorrect answers, a phenomenon we term ``seeing but not believing’’ that widely exists in major VLM families. Building on this, we introduce an inference-time intervention that highlights deep-layer evidence regions through selective attention-based masking. It requires no training and consistently improves accuracy across multiple families, including LLaVA, Qwen, Gemma, and InternVL. These results show that VLMs encode reliable evidence internally but under-utilize it, making such signals explicit can bridge the gap between perception and reasoning, advancing the diagnostic understanding and reliability of VLMs.

[543] A Survey on Self-play Methods in Reinforcement Learning

Ruize Zhang, Zelai Xu, Chengdong Ma, Chao Yu, Wei-Wei Tu, Wenhao Tang, Shiyu Huang, Deheng Ye, Wenbo Ding, Yaodong Yang, Yu Wang

Main category: cs.AI

TL;DR: This survey provides a comprehensive roadmap of self-play methods in multi-agent reinforcement learning, offering a unified framework for classification and discussing practical implications across non-cooperative scenarios.

DetailsMotivation: Despite self-play's remarkable success in complex multi-agent tasks like Go, poker, and video games, there is a lack of comprehensive and structured understanding of self-play methods in MARL.

Method: The survey introduces MARL framework and game theory concepts, provides a unified framework for classifying self-play algorithms, and illustrates their role in different non-cooperative scenarios.

Result: The paper fills the gap by offering a structured roadmap to the diverse landscape of self-play methods, bridging the gap between algorithms and their practical implications.

Conclusion: The survey highlights open challenges and future research directions in self-play, providing a foundation for further advancement in this important area of multi-agent reinforcement learning.

Abstract: Self-play, a learning paradigm where agents iteratively refine their policies by interacting with historical or concurrent versions of themselves or other evolving agents, has shown remarkable success in solving complex non-cooperative multi-agent tasks. Despite its growing prominence in multi-agent reinforcement learning (MARL), such as Go, poker, and video games, a comprehensive and structured understanding of self-play remains lacking. This survey fills this gap by offering a comprehensive roadmap to the diverse landscape of self-play methods. We begin by introducing the necessary preliminaries, including the MARL framework and basic game theory concepts. Then, it provides a unified framework and classifies existing self-play algorithms within this framework. Moreover, the paper bridges the gap between the algorithms and their practical implications by illustrating the role of self-play in different non-cooperative scenarios. Finally, the survey highlights open challenges and future research directions in self-play.

[544] Quantum Information Fusion and Correction with Dempster-Shafer Structure

Qianli Zhou, Hao Luo, Lipeng Pan, Yong Deng, Eloi Bosse

Main category: cs.AI

TL;DR: This paper implements Dempster-Shafer structure on quantum circuits, showing belief functions are more concise and effective than Bayesian approaches in quantum computing for uncertainty handling.

DetailsMotivation: Dempster-Shafer structure is effective but limited by combination growth and conflict management. The authors identified mathematical consistency between Dempster-Shafer and quantum superposition.

Method: Implemented information fusion and correction with Dempster-Shafer structure on quantum circuits, leveraging quantum computing characteristics for belief transfer.

Result: Demonstrated that belief functions provide a more concise and effective alternative to Bayesian approaches within quantum computing framework.

Conclusion: Belief functions are better suited than Bayesian approaches for handling uncertainty in quantum circuits, offering a novel perspective on basic information representation in quantum AI models.

Abstract: Dempster-Shafer structure is effective in classical settings for connecting set-valued hypotheses and representing structured ignorance, yet its practical use is limited by combination growth over focal sets and high conflict management. We observe a mathematical consistency between Dempster-Shafer structure and quantum superposition: elements of the power set form an orthogonal basis, and a basic probability assignment can be encoded as a normalized quantum state whose amplitudes respect mass value constraints. In this paper, we implement the information fusion and correction with Dempster-Shafer structure on quantum circuits, demonstrating that belief functions provide a more concise and effective alternative to Bayesian approaches within the quantum computing framework.Furthermore, by leveraging the unique characteristics of quantum computing, we propose several novel approaches for belief transfer. More broadly, this paper introduces a novel perspective on basic information representation in quantum AI models, proposing that belief functions are better suited than Bayesian approaches for handling uncertainty in quantum circuits.

[545] Whose Journey Matters? Investigating Identity Biases in Large Language Models (LLMs) for Travel Planning Assistance

Ruiping Ren, Yingwei, Xu, Xing Yao, Shu Cole, Haining Wang

Main category: cs.AI

TL;DR: LLMs exhibit ethnic and gender bias in travel recommendations, showing stereotype bias and more hallucinations for minority groups, requiring bias mitigation strategies.

DetailsMotivation: Concerns about fairness of LLMs in serving diverse identity groups in hospitality and tourism industry, grounded in social identity theory and sociotechnical systems theory.

Method: Used fairness probing to analyze outputs from three leading open-source LLMs, examining ethnic and gender biases in travel recommendations.

Result: Test accuracy for ethnicity and gender classifiers exceeded random chance, revealing stereotype bias and more frequent hallucinations in recommendations for minority groups.

Conclusion: LLMs exhibit ethnic and gender bias when functioning as travel planning assistants, underscoring need for bias mitigation strategies to improve inclusivity and reliability.

Abstract: As large language models (LLMs) become increasingly integral to the hospitality and tourism industry, concerns about their fairness in serving diverse identity groups persist. Grounded in social identity theory and sociotechnical systems theory, this study examines ethnic and gender biases in travel recommendations generated by LLMs. Using fairness probing, we analyze outputs from three leading open-source LLMs. The results show that test accuracy for both ethnicity and gender classifiers exceed random chance. Analysis of the most influential features reveals the presence of stereotype bias in LLM-generated recommendations. We also found hallucinations among these features, occurring more frequently in recommendations for minority groups. These findings indicate that LLMs exhibit ethnic and gender bias when functioning as travel planning assistants. This study underscores the need for bias mitigation strategies to improve the inclusivity and reliability of generative AI-driven travel planning assistance.

[546] When Words Smile: Generating Diverse Emotional Facial Expressions from Text

Haidong Xu, Meishan Zhang, Hao Ju, Zhedong Zheng, Erik Cambria, Min Zhang, Hao Fei

Main category: cs.AI

TL;DR: The paper introduces an end-to-end text-to-expression model focusing on emotional dynamics for digital humans, using a new dataset EmoAva with 15,000 text-3D expression pairs, achieving superior performance over baselines.

DetailsMotivation: Current talking head synthesis methods achieve good lip synchronization but overlook the rich and dynamic nature of facial expressions, creating a critical gap in enabling digital humans to express rich emotions for applications in dialogue systems and gaming.

Method: An end-to-end text-to-expression model that learns expressive facial variations in a continuous latent space and generates diverse, fluid, and emotionally coherent expressions. The method is supported by EmoAva, a large-scale dataset of 15,000 text-3D expression pairs.

Result: Extensive experiments on both existing datasets and EmoAva demonstrate that the method significantly outperforms baselines across multiple evaluation metrics.

Conclusion: The work marks a significant advancement in the field by addressing the gap in emotional expression synthesis for digital humans.

Abstract: Enabling digital humans to express rich emotions has significant applications in dialogue systems, gaming, and other interactive scenarios. While recent advances in talking head synthesis have achieved impressive results in lip synchronization, they tend to overlook the rich and dynamic nature of facial expressions. To fill this critical gap, we introduce an end-to-end text-to-expression model that explicitly focuses on emotional dynamics. Our model learns expressive facial variations in a continuous latent space and generates expressions that are diverse, fluid, and emotionally coherent. To support this task, we introduce EmoAva, a large-scale and high-quality dataset containing 15,000 text-3D expression pairs. Extensive experiments on both existing datasets and EmoAva demonstrate that our method significantly outperforms baselines across multiple evaluation metrics, marking a significant advancement in the field.

[547] Fully Autonomous AI Agents Should Not be Developed

Margaret Mitchell, Avijit Ghosh, Alexandra Sasha Luccioni, Giada Pistilli

Main category: cs.AI

TL;DR: The paper argues against developing fully autonomous AI agents, showing that risks to people increase with system autonomy levels.

DetailsMotivation: To examine the ethical implications of AI agent autonomy levels and document the trade-offs between benefits and risks to human safety.

Method: Analysis based on prior scientific literature and current product marketing to delineate different AI agent levels and detail ethical values at play.

Result: Reveals that risks to people increase with system autonomy, with safety risks being particularly concerning as they affect human life and impact other values.

Conclusion: Fully autonomous AI agents should not be developed due to escalating risks to human safety and other ethical values.

Abstract: This paper argues that fully autonomous AI agents should not be developed. In support of this position, we build from prior scientific literature and current product marketing to delineate different AI agent levels and detail the ethical values at play in each, documenting trade-offs in potential benefits and risks. Our analysis reveals that risks to people increase with the autonomy of a system: The more control a user cedes to an AI agent, the more risks to people arise. Particularly concerning are safety risks, which affect human life and impact further values.

[548] Robust Search with Uncertainty-Aware Value Models for Language Model Reasoning

Fei Yu, Yingru Li, Benyou Wang

Main category: cs.AI

TL;DR: Proposes uncertainty-aware value models and group Thompson sampling to mitigate verifier failure in LLM search, improving robustness especially on out-of-distribution problems.

DetailsMotivation: Value model guided search suffers from lack of robustness due to verifier failure, where imperfect value models mistakenly prune valid reasoning paths, particularly on unseen reasoning paths.

Method: Uses Uncertainty-Aware Value Models (UVMs) that provide value distributions instead of single-point estimates, combined with Group Thompson Sampling algorithm that selects candidates based on their probability of being optimal.

Result: Significantly mitigates verifier failure and boosts solution coverage, especially on out-of-distribution problems like AIME25 and Minerva Math, while maintaining performance on in-distribution settings like GSM8K and MATH.

Conclusion: First systematic integration of uncertainty quantification into LLM search paradigms, enhancing robustness against verifier failure in value-guided search methods.

Abstract: Value model guided search is effective in steering LLM generation but suffers from a lack of robustness. This is due to verifier failure: imperfect VMs mistakenly prune valid reasoning paths, especially when encountering unseen reasoning paths generated during search. To address this, we propose an uncertainty-aware framework with two key components: (1) Uncertainty-Aware Value Models (UVMs), which replace single-point value estimates with value distributions to quantify prediction reliability, and (2) Group Thompson Sampling, an efficient algorithm that selects candidates based on their probability of being optimal. Experiments on two In-Distribution (ID) settings (GSM8K, MATH) and three Out-Of-Distribution (OOD) settings (e.g., AIME25, Minerva Math) show our method significantly mitigates verifier failure and boosts solution coverage, especially on OOD problems. This work provides the first systematic integration of uncertainty quantification into LLM search paradigms, enhancing robustness. The code is released at https://github.com/FreedomIntelligence/UVM.

[549] Automated Knowledge Component Generation for Interpretable Knowledge Tracing in Coding Problems

Zhangqi Duan, Nigel Fernandez, Arun Balajiee Lekshmi Narayanan, Mohammad Hassany, Rafaella Sampaio de Alencar, Peter Brusilovsky, Bita Akram, Andrew Lan

Main category: cs.AI

TL;DR: Automated LLM-based pipeline for generating and tagging knowledge components (KCs) for programming problems, outperforming human-written KCs in knowledge tracing.

DetailsMotivation: Manual KC tagging by domain experts is labor-intensive; automation can improve efficiency and enable better personalized learning in online platforms.

Method: Developed an LLM-based pipeline for KC generation and tagging, plus an LLM-based knowledge tracing framework (KCGen-KT) that uses these generated KCs.

Result: KCGen-KT outperforms existing KT methods and human-written KCs in predicting student responses, shows better cognitive model fit, and generates reasonably accurate problem-KC mappings according to instructor evaluation.

Conclusion: LLM-generated KCs are effective for knowledge tracing in programming education and can automate the traditionally labor-intensive process of KC tagging.

Abstract: Knowledge components (KCs) mapped to problems help model student learning, tracking their mastery levels on fine-grained skills thereby facilitating personalized learning and feedback in online learning platforms. However, crafting and tagging KCs to problems, traditionally performed by human domain experts, is highly labor intensive. We present an automated, LLM-based pipeline for KC generation and tagging for open-ended programming problems. We also develop an LLM-based knowledge tracing (KT) framework to leverage these LLM-generated KCs, which we refer to as KCGen-KT. We conduct extensive quantitative and qualitative evaluations on two real-world student code submission datasets in different programming languages.We find that KCGen-KT outperforms existing KT methods and human-written KCs on future student response prediction. We investigate the learning curves of generated KCs and show that LLM-generated KCs result in a better fit than human written KCs under a cognitive model. We also conduct a human evaluation with course instructors to show that our pipeline generates reasonably accurate problem-KC mappings.

[550] Online Feedback Efficient Active Target Discovery in Partially Observable Environments

Anindya Sarkar, Binglin Ji, Yevgeniy Vorobeychik

Main category: cs.AI

TL;DR: DiffATD is a novel active target discovery method that uses diffusion dynamics to balance exploration-exploitation without requiring supervised training, achieving competitive performance with supervised methods.

DetailsMotivation: In domains with costly data acquisition (medical imaging, environmental monitoring, remote sensing), strategic sampling is needed to maximize target discovery within limited sampling budgets.

Method: Maintains belief distribution over unobserved states, balances exploration (sampling high entropy regions) and exploitation (targeting high belief areas), uses incrementally trained reward model to learn target characteristics.

Result: Significantly outperforms baselines and performs competitively with supervised methods that have full environmental observability, across medical imaging, species discovery, and remote sensing domains.

Conclusion: DiffATD enables efficient target discovery in partially observable environments without supervised training, offering interpretability advantages over black-box policies.

Abstract: In various scientific and engineering domains, where data acquisition is costly–such as in medical imaging, environmental monitoring, or remote sensing–strategic sampling from unobserved regions, guided by prior observations, is essential to maximize target discovery within a limited sampling budget. In this work, we introduce Diffusion-guided Active Target Discovery (DiffATD), a novel method that leverages diffusion dynamics for active target discovery. DiffATD maintains a belief distribution over each unobserved state in the environment, using this distribution to dynamically balance exploration-exploitation. Exploration reduces uncertainty by sampling regions with the highest expected entropy, while exploitation targets areas with the highest likelihood of discovering the target, indicated by the belief distribution and an incrementally trained reward model designed to learn the characteristics of the target. DiffATD enables efficient target discovery in a partially observable environment within a fixed sampling budget, all without relying on any prior supervised training. Furthermore, DiffATD offers interpretability, unlike existing black–box policies that require extensive supervised training. Through extensive experiments and ablation studies across diverse domains, including medical imaging, species discovery, and remote sensing, we show that DiffATD performs significantly better than baselines and competitively with supervised methods that operate under full environmental observability.

[551] RealMath: A Continuous Benchmark for Evaluating Language Models on Research-Level Mathematics

Jie Zhang, Cezara Petrui, Kristina Nikolić, Florian Tramèr

Main category: cs.AI

TL;DR: RealMath is a new benchmark for evaluating mathematical reasoning in LLMs using authentic research-level content from papers and forums, addressing contamination risks and enabling automated evaluation.

DetailsMotivation: Existing benchmarks use competition problems and artificial questions that don't reflect real mathematical research environments, failing to assess LLMs' practical utility for mathematicians.

Method: Sourced content directly from research papers and mathematical forums, designed verifiable statements for automated evaluation, and created a continually refreshable dataset to prevent contamination.

Result: LLMs showed surprising capabilities in handling research mathematics compared to competition problems, suggesting they can serve as valuable assistants for working mathematicians.

Conclusion: RealMath provides a more authentic evaluation of mathematical reasoning and demonstrates LLMs’ potential as research assistants, though limitations remain on highly challenging problems.

Abstract: Existing benchmarks for evaluating mathematical reasoning in large language models (LLMs) rely primarily on competition problems, formal proofs, or artificially challenging questions – failing to capture the nature of mathematics encountered in actual research environments. We introduce RealMath, a novel benchmark derived directly from research papers and mathematical forums that assesses LLMs’ abilities on authentic mathematical tasks. Our approach addresses three critical challenges: sourcing diverse research-level content, enabling reliable automated evaluation through verifiable statements, and designing a continually refreshable dataset to mitigate contamination risks. Experimental results across multiple LLMs reveal surprising capabilities in handling research mathematics compared to competition problems, suggesting current models may already serve as valuable assistants for working mathematicians despite limitations on highly challenging problems. The code and dataset for RealMath are publicly available.

[552] Ineq-Comp: Benchmarking Human-Intuitive Compositional Reasoning in Automated Theorem Proving on Inequalities

Haoyu Zhao, Yihan Geng, Shange Tang, Yong Lin, Bohan Lyu, Hongzhou Lin, Chi Jin, Sanjeev Arora

Main category: cs.AI

TL;DR: LLM-based proof assistants struggle with compositional reasoning in mathematical inequalities, showing a significant gap between AI generalization and human mathematical intuition despite scaling model size.

DetailsMotivation: To investigate whether LLM-based formal proof assistants truly understand mathematical structure like humans, particularly in compositional settings where multiple inequalities must be applied sequentially.

Method: Created Ineq-Comp benchmark using elementary inequalities transformed through variable duplication, algebraic rewriting, and multi-step composition. Evaluated multiple provers including Goedel, STP, Kimina-7B, and DeepSeek-Prover models.

Result: Most provers struggled significantly with compositional problems. DeepSeek-Prover-V2-7B showed relative robustness but still suffered 20% performance drop. Even the largest model (671B) showed gaps between compositional variants and seed problems. Performance remained poor even when constituent proofs were provided.

Conclusion: Current AI provers have a persisting weakness in compositional reasoning that isn’t solved by scaling model size alone, revealing a gap between AI generalization and human mathematical intuition.

Abstract: LLM-based formal proof assistants (e.g., in Lean) hold great promise for automating mathematical discovery. But beyond syntactic correctness, do these systems truly understand mathematical structure as humans do? We investigate this question in context of mathematical inequalities – specifically the prover’s ability to recognize that the given problem simplifies by applying a known inequality such as AM/GM. Specifically, we are interested in their ability to do this in a compositional setting where multiple inequalities must be applied as part of a solution. We introduce Ineq-Comp, a benchmark built from elementary inequalities through systematic transformations, including variable duplication, algebraic rewriting, and multi-step composition. Although these problems remain easy for humans, we find that most provers – including Goedel, STP, and Kimina-7B – struggle significantly. DeepSeek-Prover-V2-7B shows relative robustness, but still suffers a 20% performance drop (pass@32). Even for DeepSeek-Prover-V2-671B model, the gap between compositional variants and seed problems exists, implying that simply scaling up the model size alone does not fully solve the compositional weakness. Strikingly, performance remains poor for all models even when formal proofs of the constituent parts are provided in context, revealing that the source of weakness is indeed in compositional reasoning. Our results expose a persisting gap between the generalization behavior of current AI provers and human mathematical intuition. All data and evaluation code can be found at https://github.com/haoyuzhao123/LeanIneqComp.

[553] Visual Instruction Bottleneck Tuning

Changdae Oh, Jiatong Li, Shawn Im, Sharon Li

Main category: cs.AI

TL;DR: Vittle (Visual Instruction Bottleneck Tuning) improves MLLM robustness under distribution shifts by learning minimal sufficient representations through information bottleneck principle, without requiring more data or larger models.

DetailsMotivation: Multimodal LLMs suffer performance degradation under distribution shifts, and existing methods require costly data collection or model scaling.

Method: Derives variational lower bound of information bottleneck for MLLMs and implements Vittle to learn minimal sufficient representations.

Result: Empirical validation on 45 datasets and 30 shift scenarios shows consistent improvement in MLLM robustness across various tasks.

Conclusion: Vittle effectively enhances MLLM generalization under distribution shifts by pursuing information bottleneck objectives.

Abstract: Despite widespread adoption, multimodal large language models (MLLMs) suffer performance degradation when encountering unfamiliar queries under distribution shifts. Existing methods to improve MLLM generalization typically require either more instruction data or larger advanced model architectures, both of which incur non-trivial human labor or computational costs. In this work, we take an alternative approach to enhance the generalization and robustness of MLLMs under distribution shifts, from a representation learning perspective. Inspired by information bottleneck (IB) principle, we derive a variational lower bound of the IB for MLLMs and devise a practical implementation, Visual Instruction Bottleneck Tuning (Vittle). We then provide a theoretical justification of Vittle by revealing its connection to an information-theoretic robustness metric of MLLM. Empirical validation of multiple MLLMs on open-ended and closed-form question answering and object hallucination detection tasks over 45 datasets, including 30 shift scenarios, demonstrates that Vittle consistently improves the MLLM’s robustness under shifts by pursuing the learning of a minimal sufficient representation.

[554] Enumerate-Conjecture-Prove: Formally Solving Answer-Construction Problems in Math Competitions

Jialiang Sun, Yuzhi Tang, Ao Li, Chris J. Maddison, Kuldeep S. Meel

Main category: cs.AI

TL;DR: The paper introduces ECP framework, a neuro-symbolic method combining LLM creativity with formal theorem proving to solve mathematical answer-construction problems while ensuring rigor.

DetailsMotivation: Existing methods face limitations: LLMs can solve difficult answer-construction tasks but suffer from hallucinations and unverifiable steps, while symbolic methods guarantee rigor but lack creativity in answer construction. This creates a gap in solving answer-construction problems with both creativity and mathematical rigor.

Method: ECP (Enumerate-Conjecture-Prove) framework - a modular neuro-symbolic method that integrates LLM-based enumeration and pattern-driven conjecturing with formal theorem proving in Lean. It’s model agnostic and uses ConstructiveBench dataset of 3,640 formal answer-construction problems.

Result: On PutnamBench for answer construction, ECP formally solves 6 out of 337 problems end-to-end (up from 4 without ECP) using GPT-5 mini and DeepSeek-Prover-V2-7B. On ConstructiveBench, ECP achieves 33.1% end-to-end state-of-the-art accuracy (up from 32.5%), demonstrating consistent improvements over pure LLM baselines.

Conclusion: ECP framework successfully combines LLM conjecturing with formal verification, advancing formal mathematical reasoning by preserving both creativity and mathematical rigor in solving answer-construction problems.

Abstract: Mathematical reasoning is central to artificial intelligence, with applications in education, code generation, and research-level mathematical discovery. Mathematical competitions highlight two problem types: theorem proving, requiring rigorous proofs, and answer construction, requiring creative generation and formal verification of mathematical objects. Existing research reveals that LLMs can tackle difficult answer-construction tasks but are prone to errors from hallucinations and unverifiable steps, while symbolic methods guarantee rigor but falter in creative answer construction. This raises a key understudied question: how to solve answer-construction problems while preserving both LLM creativity and mathematical rigor? To address this problem, we introduce the Enumerate-Conjecture-Prove (ECP) framework, a modular neuro-symbolic method integrating LLM-based enumeration and pattern-driven conjecturing with formal theorem proving in Lean, and ConstructiveBench, a dataset of 3,640 formal answer-construction problems from math competitions. ECP is model agnostic and shows consistent improvements over pure LLM baselines: on the subset of PutnamBench for answer construction, ECP formally solves 6 out of 337 answer-construction problems end to end (up from 4 without ECP) using GPT-5 mini and DeepSeek-Prover-V2-7B. On ConstructiveBench, ECP achieves 33.1% end-to-end state-of-the-art accuracy (up from 32.5%), demonstrating its potential to advance formal mathematical reasoning by combining LLM conjecturing with formal verification. Our code and dataset are publicly available at GitHub (https://github.com/sunjia72/ECP) and Hugging Face (https://huggingface.co/datasets/sunjia72/ConstructiveBench).

[555] AgentAuditor: Human-Level Safety and Security Evaluation for LLM Agents

Hanjun Luo, Shenyu Dai, Chiming Ni, Xinfeng Li, Guibin Zhang, Kun Wang, Tongliang Liu, Hanan Salam

Main category: cs.AI

TL;DR: AgentAuditor is a memory-augmented reasoning framework that improves LLM-based evaluation of agent safety and security by learning from past experiences, achieving human-level accuracy on the new ASSEBench benchmark.

DetailsMotivation: Existing LLM-based evaluators struggle to reliably assess agent safety and security, often missing step-by-step dangers, subtle meanings, compounding issues, and getting confused by unclear safety rules.

Method: AgentAuditor constructs experiential memory by adaptively extracting structured semantic features and generating chain-of-thought reasoning traces, then uses multi-stage context-aware retrieval-augmented generation to guide evaluation of new cases.

Result: AgentAuditor consistently improves LLM evaluation performance across benchmarks and sets new state-of-the-art in LLM-as-a-judge for agent safety and security, achieving human-level accuracy.

Conclusion: AgentAuditor provides an effective framework for reliable agent safety and security evaluation, with the ASSEBench benchmark enabling comprehensive assessment of LLM-based evaluators across diverse risk scenarios.

Abstract: Despite the rapid advancement of LLM-based agents, the reliable evaluation of their safety and security remains a significant challenge. Existing rule-based or LLM-based evaluators often miss dangers in agents’ step-by-step actions, overlook subtle meanings, fail to see how small issues compound, and get confused by unclear safety or security rules. To overcome this evaluation crisis, we introduce AgentAuditor, a universal, training-free, memory-augmented reasoning framework that empowers LLM evaluators to emulate human expert evaluators. AgentAuditor constructs an experiential memory by having an LLM adaptively extract structured semantic features (e.g., scenario, risk, behavior) and generate associated chain-of-thought reasoning traces for past interactions. A multi-stage, context-aware retrieval-augmented generation process then dynamically retrieves the most relevant reasoning experiences to guide the LLM evaluator’s assessment of new cases. Moreover, we developed ASSEBench, the first benchmark designed to check how well LLM-based evaluators can spot both safety risks and security threats. ASSEBench comprises 2293 meticulously annotated interaction records, covering 15 risk types across 29 application scenarios. A key feature of ASSEBench is its nuanced approach to ambiguous risk situations, employing “Strict” and “Lenient” judgment standards. Experiments demonstrate that AgentAuditor not only consistently improves the evaluation performance of LLMs across all benchmarks but also sets a new state-of-the-art in LLM-as-a-judge for agent safety and security, achieving human-level accuracy. Our work is openly accessible at https://github.com/Astarojth/AgentAuditor.

[556] General agents contain world models

Jonathan Richens, David Abel, Alexis Bellot, Tom Everitt

Main category: cs.AI

TL;DR: World models are necessary for flexible goal-directed behavior; model-free learning is insufficient for generalization to multi-step tasks.

DetailsMotivation: To formally determine whether world models are essential for flexible goal-directed behavior or if model-free learning alone suffices.

Method: Formal analysis showing that agents capable of generalizing to multi-step goal-directed tasks must have learned a predictive model, which can be extracted from the agent’s policy.

Result: Demonstrated that world models are necessary for generalization, and that improving agent performance or handling complex goals requires increasingly accurate world models.

Conclusion: World models are essential for flexible goal-directed behavior, with implications for developing safe agents, bounding capabilities in complex environments, and creating algorithms to extract world models from agents.

Abstract: Are world models a necessary ingredient for flexible, goal-directed behaviour, or is model-free learning sufficient? We provide a formal answer to this question, showing that any agent capable of generalizing to multi-step goal-directed tasks must have learned a predictive model of its environment. We show that this model can be extracted from the agent’s policy, and that increasing the agents performance or the complexity of the goals it can achieve requires learning increasingly accurate world models. This has a number of consequences: from developing safe and general agents, to bounding agent capabilities in complex environments, and providing new algorithms for eliciting world models from agents.

[557] macOSWorld: A Multilingual Interactive Benchmark for GUI Agents

Pei Yang, Hai Ci, Mike Zheng Shou

Main category: cs.AI

TL;DR: macOSWorld is the first comprehensive benchmark for evaluating GUI agents on macOS, featuring 202 multilingual tasks across 30 applications with safety testing capabilities.

DetailsMotivation: Existing GUI agent benchmarks are mostly English-only and focus on web, Windows, Linux, and Android environments, leaving macOS - a major OS with distinctive GUI patterns and exclusive applications - unaddressed.

Method: Created macOSWorld with 202 multilingual interactive tasks across 30 applications (28 macOS-exclusive), with instructions and interfaces in 5 languages (English, Chinese, Arabic, Japanese, Russian), including a dedicated safety benchmarking subset.

Result: Evaluation of six GUI agents shows proprietary agents achieve >30% success rate while open-source models lag at <5%, revealing need for macOS domain adaptation. Multilingual testing shows 28.8% average degradation in Arabic compared to English, and safety benchmarking confirms vulnerability to deception attacks.

Conclusion: macOSWorld bridges the gap in GUI agent evaluation for macOS, revealing significant performance gaps between proprietary and open-source agents, multilingual challenges, and critical safety vulnerabilities that demand immediate attention.

Abstract: Graphical User Interface (GUI) agents show promising capabilities for automating computer-use tasks and facilitating accessibility, but existing interactive benchmarks are mostly English-only, covering web-use or Windows, Linux, and Android environments, but not macOS. macOS is a major OS with distinctive GUI patterns and exclusive applications. To bridge the gaps, we present macOSWorld, the first comprehensive benchmark for evaluating GUI agents on macOS. macOSWorld features 202 multilingual interactive tasks across 30 applications (28 macOS-exclusive), with task instructions and OS interfaces offered in 5 languages (English, Chinese, Arabic, Japanese, and Russian). As GUI agents are shown to be vulnerable to deception attacks, macOSWorld also includes a dedicated safety benchmarking subset. Our evaluation on six GUI agents reveals a dramatic gap: proprietary computer-use agents lead at above 30% success rate, while open-source lightweight research models lag at below 5%, highlighting the need for macOS domain adaptation. Multilingual benchmarks also expose common weaknesses, especially in Arabic, with a 28.8% average degradation compared to English. Results from safety benchmarking also highlight that deception attacks are more general and demand immediate attention. Project page: https://macos-world.github.io.

[558] CooT: Learning to Coordinate In-Context with Coordination Transformers

Huai-Chih Wang, Hsiang-Chun Chuang, Hsi-Chun Cheng, Dai-Jie Wu, Shao-Hua Sun

Main category: cs.AI

TL;DR: Coordination Transformers (coot) is a novel in-context coordination framework that uses interaction histories to rapidly adapt to unseen partners in multi-agent systems, outperforming existing methods.

DetailsMotivation: Existing coordination approaches like self-play and population-based methods generalize poorly to unseen partners or require extensive fine-tuning, limiting their practical applicability in dynamic environments.

Method: The proposed method uses recent interaction histories to predict actions aligned with observed behaviors, trained on trajectories from diverse agent pairs with complementary preferences without explicit supervision or parameter updates.

Result: Coordination Transformers consistently outperformed baselines including population-based approaches, gradient-based fine-tuning, and Meta-RL across diverse coordination tasks in Overcooked, achieving stable rapid adaptation and being ranked most effective in human evaluations.

Conclusion: The in-context coordination framework enables effective adaptation to unseen partners without the instability of fine-tuning or limitations of Meta-RL, providing a robust solution for multi-agent coordination.

Abstract: Effective coordination among artificial agents in dynamic and uncertain environments remains a significant challenge in multi-agent systems. Existing approaches, such as self-play and population-based methods, either generalize poorly to unseen partners or require impractically extensive fine-tuning. To overcome these limitations, we propose Coordination Transformers (\coot), a novel in-context coordination framework that uses recent interaction histories to rapidly adapt to unseen partners. Unlike prior approaches that primarily aim to diversify training partners, \coot explicitly focuses on adapting to new partner behaviors by predicting actions aligned with observed interactions. Trained on trajectories collected from diverse pairs of agents with complementary preferences, \coot quickly learns effective coordination strategies without explicit supervision or parameter updates. Across diverse coordination tasks in Overcooked, \coot consistently outperforms baselines including population-based approaches, gradient-based fine-tuning, and a Meta-RL-inspired contextual adaptation method. Notably, fine-tuning proves unstable and ineffective, while Meta-RL struggles to achieve reliable coordination. By contrast, \coot achieves stable, rapid in-context adaptation and is consistently ranked the most effective collaborator in human evaluations.

[559] Does Math Reasoning Improve General LLM Capabilities? Understanding Transferability of LLM Reasoning

Maggie Huan, Yuetai Li, Tuney Zheng, Xiaoyu Xu, Seungone Kim, Minxin Du, Radha Poovendran, Graham Neubig, Xiang Yue

Main category: cs.AI

TL;DR: Math reasoning improvements in LLMs don’t generalize to other domains; RL-tuned models show better cross-domain transfer than SFT-tuned models.

DetailsMotivation: To investigate whether performance gains in math benchmarks reflect broader problem-solving ability or just narrow overfitting to math tasks.

Method: Evaluated 20+ open-weight reasoning-tuned models across math, scientific QA, agent planning, coding, and instruction-following. Conducted controlled experiments on Qwen3-14B models using math-only data with different tuning methods (RL vs SFT). Used latent-space representation and token-space distribution shift analyses.

Result: Most models that succeed in math fail to transfer gains to other domains. RL-tuned models generalize well across domains, while SFT-tuned models often forget general capabilities. SFT induces substantial representation and output drift, while RL preserves general-domain structure.

Conclusion: Need to rethink standard post-training recipes, particularly the reliance on SFT-distilled data for advancing reasoning models, as current math improvements may not reflect true general reasoning ability.

Abstract: Math reasoning has become the poster child of progress in large language models (LLMs), with new models rapidly surpassing human-level performance on benchmarks like MATH and AIME. But as math leaderboards improve week by week, it is worth asking: do these gains reflect broader problem-solving ability or just narrow overfitting? To answer this question, we evaluate over 20 open-weight reasoning-tuned models across a broad suite of tasks, including math, scientific QA, agent planning, coding, and standard instruction-following. We surprisingly find that most models that succeed in math fail to transfer their gains to other domains. To rigorously study this phenomenon, we conduct controlled experiments on Qwen3-14B models using math-only data but different tuning methods. We find that reinforcement learning (RL)-tuned models generalize well across domains, while supervised fine-tuning (SFT)-tuned models often forget general capabilities. Latent-space representation and token-space distribution shift analyses reveal that SFT induces substantial representation and output drift, while RL preserves general-domain structure. Our results suggest a need to rethink standard post-training recipes, particularly the reliance on SFT-distilled data for advancing reasoning models.

[560] The Gauss-Markov Adjunction Provides Categorical Semantics of Residuals in Supervised Learning

Moto Kamiura

Main category: cs.AI

TL;DR: This paper uses category theory to reformulate multiple linear regression, introducing the Gauss-Markov Adjunction to structurally relate parameters and residuals, and proposes this as a semantic foundation for AI explicability.

DetailsMotivation: To enhance the intelligibility and interpretability of machine learning models by developing a semantic framework using category theory, addressing the demand for Explicability as an AI principle and promoting better social implementation of AI.

Method: Defines two Lawvere-enriched categories for parameters and data, with an adjoint pair of functors between them. Introduces the Gauss-Markov Adjunction to capture the structural interplay between residuals and parameters in supervised learning, specifically focusing on multiple linear regression.

Result: Shows that the dual flow of information between parameter variations and residuals can be explicitly described. Demonstrates that the ordinary least squares estimator and minimum residual are related via preservation of limits by the right adjoint functor.

Conclusion: Positions this categorical formulation as an instance of extended denotational semantics for supervised learning, proposing to apply semantic perspectives from theoretical computer science as a formal foundation for Explicability in AI.

Abstract: Enhancing the intelligibility and interpretability of machine learning is a crucial task in responding to the demand for Explicability as an AI principle, and in promoting the better social implementation of AI. The aim of our research is to contribute to this improvement by reformulating machine learning models through the lens of category theory, thereby developing a semantic framework for structuring and understanding AI systems. Our categorical modeling in this paper clarifies and formalizes the structural interplay between residuals and parameters in supervised learning. The present paper focuses on the multiple linear regression model, which represents the most basic form of supervised learning. By defining two Lawvere-enriched categories corresponding to parameters and data, along with an adjoint pair of functors between them, we introduce our categorical formulation of supervised learning. We show that the essential structure of this framework is captured by what we call the Gauss-Markov Adjunction. Within this setting, the dual flow of information can be explicitly described as a correspondence between variations in parameters and residuals. The ordinary least squares estimator for the parameters and the minimum residual are related via the preservation of limits by the right adjoint functor. Furthermore, we position this formulation as an instance of extended denotational semantics for supervised learning, and propose applying a semantic perspective developed in theoretical computer science as a formal foundation for Explicability in AI.

[561] Advancing Routing-Awareness in Analog ICs Floorplanning

Davide Basso, Luca Bortolussi, Mirjana Videnovic-Misic, Husni Habal

Main category: cs.AI

TL;DR: A reinforcement learning and graph neural network approach for automatic analog IC floorplanning that improves routability, reducing dead space by 13.8%, wirelength by 40.6%, and increasing routing success by 73.4% compared to previous learning-based methods.

DetailsMotivation: Address the limited adoption of ML techniques in analog IC layout due to stringent electrical constraints and the interdependence between floorplanning and routing, providing routing-aware floorplanning solutions that layout engineers need.

Method: Developed an automatic floorplanning engine using reinforcement learning and relational graph convolutional neural networks, with increased grid resolution, precise pin information integration, and dynamic routing resource estimation to balance routing and area efficiency.

Result: Achieved 13.8% reduction in dead space, 40.6% reduction in wirelength, and 73.4% increase in routing success compared to previous learning-based state-of-the-art techniques in simulated place and route environment.

Conclusion: The proposed approach effectively addresses analog IC layout challenges by integrating routing awareness into floorplanning, meeting industrial standards and significantly outperforming existing learning-based methods.

Abstract: The adoption of machine learning-based techniques for analog integrated circuit layout, unlike its digital counterpart, has been limited by the stringent requirements imposed by electric and problem-specific constraints, along with the interdependence of floorplanning and routing steps. In this work, we address a prevalent concern among layout engineers regarding the need for readily available routing-aware floorplanning solutions. To this extent, we develop an automatic floorplanning engine based on reinforcement learning and relational graph convolutional neural network specifically tailored to condition the floorplan generation towards more routable outcomes. A combination of increased grid resolution and precise pin information integration, along with a dynamic routing resource estimation technique, allows balancing routing and area efficiency, eventually meeting industrial standards. When analyzing the place and route effectiveness in a simulated environment, the proposed approach achieves a 13.8% reduction in dead space, a 40.6% reduction in wirelength and a 73.4% increase in routing success when compared to past learning-based state-of-the-art techniques.

[562] DARIL: When Imitation Learning outperforms Reinforcement Learning in Surgical Action Planning

Maxence Boels, Harry Robertshaw, Thomas C Booth, Prokar Dasgupta, Alejandro Granados, Sebastien Ourselin

Main category: cs.AI

TL;DR: Comprehensive comparison shows imitation learning outperforms reinforcement learning for surgical action planning, challenging assumptions about RL superiority in sequential decision making.

DetailsMotivation: Surgical action planning requires predicting future instrument-verb-target triplets. While RL could potentially discover superior strategies through self-exploration, this study aims to compare IL versus RL approaches.

Method: Developed Dual-task Autoregressive Imitation Learning (DARIL) baseline and evaluated three RL variants: world model-based RL, direct video RL, and inverse RL enhancement on CholecT50 dataset.

Result: DARIL achieved 34.6% action triplet recognition mAP and 33.6% next frame prediction mAP, while all RL approaches underperformed - world model RL dropped to 3.1% mAP and direct video RL achieved only 15.9% at 10-second horizons.

Conclusion: Distribution matching on expert-annotated test sets systematically favors IL over potentially valid RL policies that differ from training demonstrations, challenging assumptions about RL superiority in sequential decision making.

Abstract: Surgical action planning requires predicting future instrument-verb-target triplets for real-time assistance. While teleoperated robotic surgery provides natural expert demonstrations for imitation learning (IL), reinforcement learning (RL) could potentially discover superior strategies through self-exploration. We present the first comprehensive comparison of IL versus RL for surgical action planning on CholecT50. Our Dual-task Autoregressive Imitation Learning (DARIL) baseline achieves 34.6% action triplet recognition mAP and 33.6% next frame prediction mAP with smooth planning degradation to 29.2% at 10-second horizons. We evaluated three RL variants: world model-based RL, direct video RL, and inverse RL enhancement. Surprisingly, all RL approaches underperformed DARIL–world model RL dropped to 3.1% mAP at 10s while direct video RL achieved only 15.9%. Our analysis reveals that distribution matching on expert-annotated test sets systematically favors IL over potentially valid RL policies that differ from training demonstrations. This challenges assumptions about RL superiority in sequential decision making and provides crucial insights for surgical AI development.

[563] Working with AI: Measuring the Applicability of Generative AI to Occupations

Kiran Tomlinson, Sonia Jaffe, Will Wang, Scott Counts, Siddharth Suri

Main category: cs.AI

TL;DR: Analysis of 200k AI conversations reveals AI’s strongest impact on knowledge work occupations like computer/mathematical fields and office support, with information gathering and writing being the most common assisted activities.

DetailsMotivation: To understand AI's economic impact by analyzing how people use AI for work activities and which occupations are most affected by AI assistance.

Method: Analyzed 200k anonymized conversations from Microsoft Bing Copilot users, classified work activities, measured task success and scope of impact, then computed AI applicability scores for occupations.

Result: Highest AI applicability scores found for knowledge work occupations (computer/mathematical, office/administrative support) and sales roles. Most common AI-assisted activities are information gathering and writing.

Conclusion: AI has significant applicability in knowledge-intensive occupations, particularly those involving information processing and communication, with real-world usage patterns differing from theoretical predictions of AI impact.

Abstract: Given the rapid adoption of generative AI and its potential to impact a wide range of tasks, understanding the effects of AI on the economy is one of society’s most important questions. In this work, we take a step toward that goal by analyzing the work activities people do with AI, how successfully and broadly those activities are done, and combine that with data on what occupations do those activities. We analyze a dataset of 200k anonymized and privacy-scrubbed conversations between users and Microsoft Bing Copilot, a publicly available generative AI system. We find the most common work activities people seek AI assistance for involve gathering information and writing, while the most common activities that AI itself is performing are providing information and assistance, writing, teaching, and advising. Combining these activity classifications with measurements of task success and scope of impact, we compute an AI applicability score for each occupation. We find the highest AI applicability scores for knowledge work occupation groups such as computer and mathematical, and office and administrative support, as well as occupations such as sales whose work activities involve providing and communicating information. Additionally, we characterize the types of work activities performed most successfully, how wage and education correlate with AI applicability, and how real-world usage compares to predictions of occupational AI impact.

[564] RL-PLUS: Countering Capability Boundary Collapse of LLMs in Reinforcement Learning with Hybrid-policy Optimization

Yihong Dong, Xue Jiang, Yongding Tao, Huanyu Liu, Kechi Zhang, Lili Mou, Rongyu Cao, Yingwei Ma, Jue Chen, Binhua Li, Zhi Jin, Fei Huang, Yongbin Li, Ge Li

Main category: cs.AI

TL;DR: RL-PLUS is a hybrid-policy optimization method that overcomes RLVR’s limitations by combining internal exploitation with external data, achieving superior reasoning performance and preventing capability boundary collapse in LLMs.

DetailsMotivation: RLVR struggles to surpass base LLM capability boundaries due to on-policy strategy, large action space, and sparse rewards, leading to capability boundary collapse that narrows problem-solving scope.

Method: Integrates Multiple Importance Sampling to handle distributional mismatch from external data and Exploration-Based Advantage Function to guide models toward high-value unexplored reasoning paths.

Result: Achieves SOTA performance on 6 math reasoning benchmarks, superior results on 6 OOD tasks, up to 69.2% relative improvements across model families, and resolves capability boundary collapse.

Conclusion: RL-PLUS effectively surpasses base model boundaries through hybrid-policy optimization, demonstrating strong generalizability and solving the capability collapse problem in RLVR methods.

Abstract: Reinforcement Learning with Verifiable Reward (RLVR) has significantly advanced the complex reasoning abilities of Large Language Models (LLMs). However, it struggles to break through the inherent capability boundaries of the base LLM, due to its essentially on-policy strategy coupled with LLM’s immense action space and sparse reward. Critically, RLVR can lead to the capability boundary collapse, narrowing the LLM’s problem-solving scope. To address this problem, we propose RL-PLUS, a novel hybrid-policy optimization approach for LLMs that synergizes internal exploitation with external data to achieve stronger reasoning capabilities and surpass the boundaries of base models. RL-PLUS integrates two core components, i.e., Multiple Importance Sampling to address distributional mismatch from external data, and Exploration-Based Advantage Function to guide the model towards high-value, unexplored reasoning paths. We provide both theoretical analysis and extensive experiments to demonstrate the superiority and generalizability of our approach. Compared with existing RLVR methods, RL-PLUS achieves 1) state-of-the-art performance on six math reasoning benchmarks; 2) superior performance on six out-of-distribution reasoning tasks; 3) consistent and significant gains across diverse model families, with average relative improvements up to 69.2%. Moreover, the analysis of Pass@k curves indicates that RL-PLUS effectively resolves the capability boundary collapse problem.

[565] Trainable Dynamic Mask Sparse Attention

Jingze Shi, Yifan Wu, Yiran Peng, Bingheng Wu, Liangdong Wang, Guang Liu, Yuyu Luo

Main category: cs.AI

TL;DR: Dynamic Mask Attention (DMA) is a trainable sparse attention mechanism that combines position-aware and content-aware approaches to address the quadratic complexity of self-attention in LLMs, achieving up to 10x speedup while maintaining performance.

DetailsMotivation: Standard self-attention has quadratic complexity that bottlenecks long-context modeling in LLMs. Existing sparse attention methods either use static sparse structures lacking adaptability or rely on heuristic key-value selection that hinders differentiability.

Method: DMA uses three innovations: 1) value vector representations to generate content-aware dynamic masks, 2) position-aware sparse weights computed in hardware-friendly manner, 3) end-to-end trainable design that doesn’t obstruct gradients.

Result: DMA consistently outperforms state-of-the-art sparse attention baselines in scaling laws, multi-query associative recall, standard benchmarks, and needle in a haystack tests, with up to 10x overall speedup.

Conclusion: DMA effectively balances model efficiency with long-context modeling capabilities, demonstrating Pareto advantage over existing methods while being fully differentiable and hardware-friendly.

Abstract: The increasing demand for long-context modeling in large language models (LLMs) is bottlenecked by the quadratic complexity of the standard self-attention mechanism. The community has proposed sparse attention to mitigate this issue. However, position-aware sparse attention methods rely on static sparse structures that lack adaptability to diverse query contexts, while content-aware sparse attention methods depend on heuristic key-value selection, hindering full differentiability. We introduce a trainable dynamic mask sparse attention mechanism, a method that merges the advantages of both position-aware and content-aware approaches. Dynamic Mask Attention (DMA) achieves this through three key innovations: First, it leverages value vector representations to generate content-aware dynamic masks, enabling the model to adaptively identify and attend to critical information. Second, it computes position-aware sparse weights in a hardware-friendly manner, efficiently skipping unnecessary computational regions. Finally, we demonstrate that the introduced dynamic mask and sparse weights do not obstruct gradients, supporting end-to-end training. We have validated the performance of DMA through comprehensive experiments. A large body of experimental evidence shows that DMA consistently holds a Pareto advantage over state-of-the-art sparse attention baselines in tasks including scaling laws, multi-query associative recall, standard benchmarks, and needle in a haystack tests, while also delivering up to a 10x overall speedup. These results highlight its ability to effectively balance model efficiency with long-context modeling capabilities. Our computational kernel code is now open-source at https://github.com/SmallDoges/flash-dmattn to encourage further research and application by the community.

[566] LLM Collaboration With Multi-Agent Reinforcement Learning

Shuo Liu, Tianle Chen, Zeyu Liang, Xueguang Lyu, Christopher Amato

Main category: cs.AI

TL;DR: The paper proposes MAGRPO, a multi-agent reinforcement learning approach to fine-tune LLMs for better collaboration, addressing the gap in current LLM training that lacks coordination optimization.

DetailsMotivation: Most LLMs are pretrained independently without coordination optimization, and existing fine-tuning frameworks rely on complex individual reward designs that don't effectively encourage collaboration.

Method: Model LLM collaboration as cooperative MARL problem and develop Multi-Agent Group Relative Policy Optimization (MAGRPO) algorithm, building on RL approaches for LLMs and MARL techniques.

Result: Experiments on LLM writing and coding collaboration show that fine-tuning with MAGRPO enables agents to generate high-quality responses efficiently through effective cooperation.

Conclusion: The approach opens doors to using other MARL methods for LLMs and highlights associated challenges in multi-agent LLM coordination.

Abstract: A large amount of work has been done in Multi-Agent Systems (MAS) for modeling and solving problems with multiple interacting agents. However, most LLMs are pretrained independently and not specifically optimized for coordination. Existing LLM fine-tuning frameworks rely on individual rewards, which require complex reward designs for each agent to encourage collaboration. To address these challenges, we model LLM collaboration as a cooperative Multi-Agent Reinforcement Learning (MARL) problem. We develop a multi-agent, multi-turn algorithm, Multi-Agent Group Relative Policy Optimization (MAGRPO), to solve it, building on current RL approaches for LLMs as well as MARL techniques. Our experiments on LLM writing and coding collaboration demonstrate that fine-tuning MAS with MAGRPO enables agents to generate high-quality responses efficiently through effective cooperation. Our approach opens the door to using other MARL methods for LLMs and highlights the associated challenges.

[567] BASIL: Bayesian Assessment of Sycophancy in LLMs

Katherine Atwell, Pedram Heydari, Anthony Sicilia, Malihe Alikhani

Main category: cs.AI

TL;DR: The paper introduces a Bayesian framework to study sycophancy in LLMs without requiring ground-truth data, finding significant evidence of sycophantic behavior that often reduces rationality in decision-making.

DetailsMotivation: Existing methods for studying sycophancy in LLMs are limited to objective tasks with ground-truth data, ignoring the natural subjectivity in many NLP tasks. Understanding sycophancy is critical for human-AI collaboration in decision-making settings like health, law, and education.

Method: Drawing from behavioral economics and rational decision theory, the authors introduce a Bayesian framework to study normative effects of sycophancy on rationality in LLMs. They experiment with various methods for eliciting sycophancy and obtaining probability judgments across multiple LLM baselines in three different tasks.

Result: Significant evidence of sycophancy was found (7 of 8 baselines for one probing technique). Sycophancy was more likely to reduce rationality than increase it in LLMs’ decisions when directly probed for probabilities (2 out of 4 baselines showed significant increases overall).

Conclusion: The Bayesian framework successfully studies sycophancy’s normative effects on rationality without requiring labeled ground-truth data, revealing that sycophantic behavior often compromises rational decision-making in LLMs.

Abstract: Sycophancy (overly agreeable or flattering behavior) is critical to understand in the context of human-AI collaboration, especially in decision-making settings like health, law, and education. Existing methods for studying sycophancy in LLMs are either descriptive (study behavior change when sycophancy is elicited) or normative (provide values-based judgment on behavior change). Together, these approaches help us understand the extent, and impacts, of sycophancy. However, existing normative approaches only apply for objective tasks where ground-truth data exists, ignoring the natural subjectivity in many NLP tasks. Drawing from behavioral economics and rational decision theory, we introduce an Bayesian framework to study the normative effects of sycophancy on rationality in LLMs, without requiring labeled ground-truth. Using this interdisciplinary framework, we study sycophantic behavior in multiple LLM baselines across three different tasks, experimenting with various methods for eliciting sycophancy and obtaining probability judgments from LLMs. We find significant evidence of sycophancy in our experiments (7 of 8 baselines for one of our probing techniques), and observe that sycophancy is more likely to reduce rationality than it is to increase rationality in LLMs’ decisions when they are directly probed for probabilities (2 out of 4 baselines show significant increases overall).

[568] A Markovian Framing of WaveFunctionCollapse for Procedurally Generating Aesthetically Complex Environments

Franklin Yiu, Mohan Lu, Nina Li, Kevin Joseph, Tianxu Zhang, Julian Togelius, Timothy Merino, Sam Earle

Main category: cs.AI

TL;DR: Reformulating WaveFunctionCollapse as a Markov Decision Process to decouple constraint satisfaction from objective optimization, showing superior performance over joint optimization approaches.

DetailsMotivation: Procedural content generation needs to satisfy both designer objectives and tile adjacency constraints, requiring joint optimization which becomes challenging as complexity increases.

Method: Reformulate WaveFunctionCollapse as a Markov Decision Process (WFC-MDP), allowing optimization algorithms to focus on objectives while WFC handles constraint satisfaction through propagation.

Result: WFC-MDP consistently outperforms traditional evolutionary approaches across multiple domains and difficulty levels, especially as task complexity increases.

Conclusion: Decoupling local constraint satisfaction from global objective optimization provides significant advantages over joint optimization approaches in procedural content generation.

Abstract: Procedural content generation often requires satisfying both designer-specified objectives and adjacency constraints implicitly imposed by the underlying tile set. To address the challenges of jointly optimizing both constraints and objectives, we reformulate WaveFunctionCollapse (WFC) as a Markov Decision Process (MDP), enabling external optimization algorithms to focus exclusively on objective maximization while leveraging WFC’s propagation mechanism to enforce constraint satisfaction. We empirically compare optimizing this MDP to traditional evolutionary approaches that jointly optimize global metrics and local tile placement. Across multiple domains with various difficulties, we find that joint optimization not only struggles as task complexity increases, but consistently underperforms relative to optimization over the WFC-MDP, underscoring the advantages of decoupling local constraint satisfaction from global objective optimization.

[569] From Next Token Prediction to (STRIPS) World Models – Preliminary Results

Carlos Núñez-Molina, Vicenç Gómez, Hector Geffner

Main category: cs.AI

TL;DR: Learning propositional STRIPS world models from action traces using transformers and gradient descent, framed as next token prediction.

DetailsMotivation: To learn world models from action traces alone without explicit state information, using deep learning approaches.

Method: Use transformer architecture for next token prediction where tokens are actions, ensuring action preconditions are satisfied by previous hidden effects.

Result: Transformers can faithfully represent STRIPS models and learn them from random valid/invalid action sequences.

Conclusion: Suitable transformer architectures can effectively learn propositional STRIPS world models from action traces through supervised learning.

Abstract: We consider the problem of learning propositional STRIPS world models from action traces alone, using a deep learning architecture (transformers) and gradient descent. The task is cast as a supervised next token prediction problem where the tokens are the actions, and an action $a$ may follow an action sequence if the hidden effects of the previous actions do not make an action precondition of $a$ false. We show that a suitable transformer architecture can faithfully represent propositional STRIPS world models, and that the models can be learned from sets of random valid (positive) and invalid (negative) action sequences alone. A number of experiments are reported.

[570] Programmable Cognitive Bias in Social Agents

Xuan Liu, Haoyang Shang, Haojian Jin

Main category: cs.AI

TL;DR: CoBRA is a toolkit for systematically programming cognitive biases in LLM-based social agents by grounding behaviors in classic social science experiments, enabling consistent and nuanced agent behaviors across different models.

DetailsMotivation: Conventional approaches using natural language descriptions fail to produce consistent agent behaviors across models and cannot capture behavioral nuances effectively.

Method: CoBRA has two components: Cognitive Bias Index that measures bias through social science experiments, and Behavioral Regulation Engine that aligns agent behavior to demonstrate controlled cognitive bias.

Result: CoBRA can precisely program cognitive bias in social agents in a model-agnostic manner, as demonstrated through HCI toolkit evaluation and technical benchmarks.

Conclusion: CoBRA provides a systematic approach to specifying agent behavior by explicitly programming cognitive biases, overcoming limitations of natural language-based methods.

Abstract: This paper introduces CoBRA, a novel toolkit for systematically specifying agent behavior in LLM-based social simulation. We found that conventional approaches that specify agent behaviors through implicit natural language descriptions cannot yield consistent behaviors across models, and the produced agent behaviors do not capture the nuances of the descriptions. In contrast, CoBRA presents a new approach to program agents’ cognitive biases explicitly, by grounding agents’ expected behaviors using classic social science experiments. CoBRA has two components: (1) Cognitive Bias Index that measures the cognitive bias of a social agent, by quantifying the agent’s reactions in a set of validated classical social science experiments; (2) Behavioral Regulation Engine that aligns the agent’s behavior to demonstrate controlled cognitive bias. We evaluated CoBRA as an HCI toolkit through demonstration and technical benchmarks. Our results suggest that CoBRA can precisely program the cognitive bias demonstrated in a social agent in a model-agnostic manner.

[571] Chain-in-Tree: Back to Sequential Reasoning in LLM Tree Search

Xinzhe Li

Main category: cs.AI

TL;DR: Chain-in-Tree (CiT) is a plug-in framework that reduces computational costs in LLM tree search methods by selectively branching only when necessary, achieving 75-85% reductions in tokens, model calls, and runtime with minimal accuracy loss.

DetailsMotivation: Existing LLM Inference via Tree Search (LITS) methods achieve strong performance but are highly inefficient, often running an order of magnitude slower than iterative approaches, creating a need for more efficient inference methods.

Method: CiT introduces lightweight Branching Necessity (BN) evaluations: BN-DP uses an auxiliary LLM to judge branching needs, and BN-SC clusters candidate actions to assess agreement. This framework decides when to branch during search rather than expanding at every step.

Result: BN-DP achieves 75-85% reductions in token generation, model calls, and runtime on GSM8K and Math500 with negligible or no accuracy loss. BN-SC typically yields substantial savings (up to 80%) but shows instability in some settings due to extremely long reasoning steps.

Conclusion: CiT provides an effective plug-in framework that significantly improves the efficiency of LITS methods while maintaining performance, with theoretical guarantees that BN-DP never increases policy invocations.

Abstract: Test-time scaling improves large language models (LLMs) on long-horizon reasoning tasks by allocating more compute at inference. LLM Inference via Tree Search (LITS) methods achieve strong performance but are highly inefficient, often running an order of magnitude slower than iterative approaches. We propose Chain-in-Tree (CiT), a plug-in framework that decides when to branch during search rather than expanding at every step. CiT introduces lightweight Branching Necessity (BN) evaluations: BN-DP (Direct Prompting), where an auxiliary LLM judges branching needs, and BN-SC (Self-Consistency), which clusters candidate actions to assess agreement. Integrated into Tree of Thoughts, ReST-MCTS, and RAP, BN-DP achieves 75-85% reductions in token generation, model calls, and runtime on GSM8K and Math500, with often negligible or no accuracy loss. BN-SC typically yields substantial savings (up to 80%) generally but shows instability in 1-4 out of 14 settings, caused by a small subset of examples that produce extremely long reasoning steps. We theoretically prove that BN-DP never increases policy invocations and release both modular LITS implementations and a lightweight CiT function applicable across all LITS variants. The full codebase is publicly available at https://github.com/xinzhel/chain_in_tree.

[572] DualTune: Decoupled Fine-Tuning for On-Device Agentic Systems

Rohan Kadekodi, Zhan Jin, Keisuke Kamahori, Yile Gu, Sean Khatiri, Noah H. Bayindirli, Sergey Gorbunov, Baris Kasikci

Main category: cs.AI

TL;DR: The paper introduces decoupled fine-tuning and DualTune framework to improve local LLMs’ tool-calling performance by separating tool selection and argument generation into distinct subtasks with dedicated LoRA adapters.

DetailsMotivation: Local LLMs underperform frontier models in tool calling scenarios, struggling with tool selection from large sets and accurate argument generation for complex parameters, while privacy and cost concerns demand on-device inference capabilities.

Method: Decoupled fine-tuning using LoRA adapters with separate loss masking for tool selection and argument generation subtasks, plus DualTune inference framework that dynamically loads corresponding adapters and implements hierarchical orchestration.

Result: Qwen-2.5-7B model with decoupled fine-tuning improves tool calling accuracy by 46% over base model, outperforms similar-sized models in all cases and often beats 2x larger models on MCP-Bench benchmark.

Conclusion: The proposed methodology effectively addresses local LLMs’ tool-calling limitations through task disaggregation and specialized fine-tuning, enabling privacy-preserving, cost-effective agent orchestration on end-user devices.

Abstract: The deployment of Large Language Models (LLMs) as agentic orchestrators has revolutionized task automation, but the need for privacy-preserving, cost-effective solutions demands on-device inference capabilities. However, local LLMs consistently underperform compared to frontier models in tool calling scenarios, struggling with both tool selection from large tool sets and accurate argument generation for complex parameter structures. We introduce a methodology that disaggregates a tool-calling task into two distinct subtasks: tool selection and argument generation. We propose “decoupled fine-tuning”, a novel post-training approach that employs LoRA fine-tuning to create dedicated LoRA adapters for tool selection and tool-specific argument generation using separate loss masking for each of the subtasks. Furthermore, we present DualTune, an inference framework that leverages the LoRA adapters created using decoupled fine-tuning to perform efficient agent orchestration with the help of local models on end-user devices. DualTune decomposes the tool-call generation step into tool selection and argument generation, and dynamically loads the corresponding LoRA adapters to generate tool calls. Additionally, DualTune implements hierarchical orchestration to restrict the number of tools required for tool selection. Our experiments on the MCP-Bench benchmark demonstrate that the Qwen-2.5-7B model trained using decoupled fine-tuning improves the tool calling accuracy of the base model by 46%, and outperforms other local reasoning, non-reasoning and fine-tuned models of similar size in all cases, and models that are 2x larger, in most cases.

[573] PsychCounsel-Bench: Evaluating the Psychology Intelligence of Large Language Models

Min Zeng

Main category: cs.AI

TL;DR: This paper introduces PsychCounsel-Bench, a benchmark based on US counselor certification exams to evaluate if LLMs can qualify as psychological counselors by testing their psychological knowledge.

DetailsMotivation: While LLMs show impressive generative abilities, their potential in cognitive applications like psychological counseling remains largely unexplored. The paper aims to determine if LLMs can effectively serve as psychological counselors by assessing if they meet professional certification standards.

Method: The authors created PsychCounsel-Bench, a benchmark comprising approximately 2,252 single-choice questions from US national counselor examinations, which requires about 70% accuracy to pass. This benchmark tests deep understanding across various psychology sub-disciplines.

Result: Advanced models like GPT-4o, Llama3.3-70B, and Gemma3-27B achieved well above the passing threshold, while smaller open-source models (Qwen2.5-7B, Mistral-7B) remained far below it.

Conclusion: Only frontier LLMs currently meet counseling exam standards, highlighting both the promise and challenges of developing psychology-oriented LLMs. The dataset is publicly released for further research.

Abstract: Large Language Models (LLMs) have demonstrated remarkable success across a wide range of industries, primarily due to their impressive generative abilities. Yet, their potential in applications requiring cognitive abilities, such as psychological counseling, remains largely untapped. This paper investigates the key question: \textit{Can LLMs be effectively applied to psychological counseling?} To determine whether an LLM can effectively take on the role of a psychological counselor, the first step is to assess whether it meets the qualifications required for such a role, namely the ability to pass the U.S. National Counselor Certification Exam (NCE). This is because, just as a human counselor must pass a certification exam to practice, an LLM must demonstrate sufficient psychological knowledge to meet the standards required for such a role. To address this, we introduce PsychCounsel-Bench, a benchmark grounded in U.S.national counselor examinations, a licensure test for professional counselors that requires about 70% accuracy to pass. PsychCounsel-Bench comprises approximately 2,252 carefully curated single-choice questions, crafted to require deep understanding and broad enough to cover various sub-disciplines of psychology. This benchmark provides a comprehensive assessment of an LLM’s ability to function as a counselor. Our evaluation shows that advanced models such as GPT-4o, Llama3.3-70B, and Gemma3-27B achieve well above the passing threshold, while smaller open-source models (e.g., Qwen2.5-7B, Mistral-7B) remain far below it. These results suggest that only frontier LLMs are currently capable of meeting counseling exam standards, highlighting both the promise and the challenges of developing psychology-oriented LLMs. We release the proposed dataset for public use: https://github.com/cloversjtu/PsychCounsel-Bench

[574] Aligning Perception, Reasoning, Modeling and Interaction: A Survey on Physical AI

Kun Xiang, Terry Jingchen Zhang, Yinya Huang, Jixi He, Zirong Liu, Yueling Tang, Ruizhe Zhou, Lijing Luo, Youpeng Wen, Xiuwei Chen, Bingqian Lin, Jianhua Han, Hang Xu, Hanhui Li, Bin Dong, Xiaodan Liang

Main category: cs.AI

TL;DR: This paper provides a comprehensive overview of Physical AI, bridging the gap between theoretical physics reasoning and applied physical understanding in AI systems.

DetailsMotivation: The rapid advancement of embodied intelligence and world models has intensified efforts to integrate physical laws into AI systems, but physical perception and symbolic physics reasoning have developed separately without a unified framework.

Method: Systematic examination of how physics-grounded methods enhance AI’s real-world comprehension across structured symbolic reasoning, embodied systems, and generative models through rigorous analysis of recent advances.

Result: The paper establishes clear distinctions between theoretical physics reasoning and applied physical understanding, advocating for intelligent systems that ground learning in both physical principles and embodied reasoning processes.

Conclusion: The synthesis envisions next-generation world models capable of explaining physical phenomena and predicting future states, advancing safe, generalizable, and interpretable AI systems.

Abstract: The rapid advancement of embodied intelligence and world models has intensified efforts to integrate physical laws into AI systems, yet physical perception and symbolic physics reasoning have developed along separate trajectories without a unified bridging framework. This work provides a comprehensive overview of physical AI, establishing clear distinctions between theoretical physics reasoning and applied physical understanding while systematically examining how physics-grounded methods enhance AI’s real-world comprehension across structured symbolic reasoning, embodied systems, and generative models. Through rigorous analysis of recent advances, we advocate for intelligent systems that ground learning in both physical principles and embodied reasoning processes, transcending pattern recognition toward genuine understanding of physical laws. Our synthesis envisions next-generation world models capable of explaining physical phenomena and predicting future states, advancing safe, generalizable, and interpretable AI systems. We maintain a continuously updated resource at https://github.com/AI4Phys/Awesome-AI-for-Physics.

[575] Optimizing for Persuasion Improves LLM Generalization: Evidence from Quality-Diversity Evolution of Debate Strategies

Aksel Joonas Reedi, Corentin Léger, Julien Pourcel, Loris Gaven, Perrine Charriau, Guillaume Pourcel

Main category: cs.AI

TL;DR: DebateQD introduces a Quality-Diversity evolutionary algorithm that evolves diverse debate strategies through tournament-style competitions, showing that persuasion-based optimization achieves better generalization than truth-based approaches.

DetailsMotivation: LLMs optimized for truthfulness often overfit and produce brittle reasoning that fails to generalize. Persuasion-based optimization has shown promise but hasn't been systematically compared against truth-based approaches.

Method: DebateQD uses a minimal QD evolutionary algorithm with tournament-style debates where two LLMs debate while a third judges. It evolves diverse strategies across categories (rationality, authority, emotional appeal) using prompt-based strategies within a single LLM architecture.

Result: Persuasion-optimized strategies achieve up to 13.94% smaller train-test generalization gaps while matching or exceeding truth optimization’s test performance across three model scales (7B, 32B, 72B parameters) and multiple dataset sizes.

Conclusion: Competitive pressure to persuade, rather than seeking truth collaboratively, fosters more transferable reasoning skills, offering a promising path for improving LLM generalization.

Abstract: Large Language Models (LLMs) optimized to output truthful answers often overfit, producing brittle reasoning that fails to generalize. While persuasion-based optimization has shown promise in debate settings, it has not been systematically compared against mainstream truth-based approaches. We introduce DebateQD, a minimal Quality-Diversity (QD) evolutionary algorithm that evolves diverse debate strategies across different categories (rationality, authority, emotional appeal, etc.) through tournament-style competitions where two LLMs debate while a third judges. Unlike previously proposed methods that require a population of LLMs, our approach maintains diversity of opponents through prompt-based strategies within a single LLM architecture, making it more accessible for experiments while preserving the key benefits of population-based optimization. In contrast to prior work, we explicitly isolate the role of the optimization objective by fixing the debate protocol and swapping only the fitness function: persuasion rewards strategies that convince the judge irrespective of truth, whereas truth rewards collaborative correctness. Across three model scales (7B, 32B, 72B parameters) and multiple dataset sizes from the QuALITY benchmark, persuasion-optimized strategies achieve up to 13.94% smaller train-test generalization gaps, while matching or exceeding truth optimization’s test performance. These results provide the first controlled evidence that competitive pressure to persuade, rather than seek the truth collaboratively, fosters more transferable reasoning skills, offering a promising path for improving LLM generalization.

[576] DRIFT: Decompose, Retrieve, Illustrate, then Formalize Theorems

Meiru Zhang, Philipp Borchert, Milan Gritta, Gerasimos Lampouras

Main category: cs.AI

TL;DR: DRIFT is a framework that improves mathematical autoformalization by decomposing informal statements into sub-components for better premise retrieval from math libraries, achieving significant performance gains across multiple benchmarks.

DetailsMotivation: Current LLMs struggle with formalizing mathematical statements because they can't effectively identify and use prerequisite mathematical knowledge from formal libraries. Direct retrieval methods fail due to the complexity and limited context of informal statements.

Method: DRIFT decomposes informal mathematical statements into smaller sub-components to enable targeted retrieval of premises from libraries like Mathlib. It also retrieves illustrative theorems to help models use premises more effectively in formalization.

Result: DRIFT nearly doubles the F1 score compared to DPR baseline on ProofNet, and shows strong out-of-distribution performance on ConNF with BEq+@10 improvements of 37.14% (GPT-4.1) and 42.25% (DeepSeek-V3.1).

Conclusion: Retrieval effectiveness in mathematical autoformalization depends on model-specific knowledge boundaries, requiring adaptive retrieval strategies aligned with each model’s capabilities. DRIFT demonstrates the value of decomposition for improving premise retrieval.

Abstract: Automating the formalization of mathematical statements for theorem proving remains a major challenge for Large Language Models (LLMs). LLMs struggle to identify and utilize the prerequisite mathematical knowledge and its corresponding formal representation in languages like Lean. Current retrieval-augmented autoformalization methods query external libraries using the informal statement directly, but overlook a fundamental limitation: informal mathematical statements are often complex and offer limited context on the underlying math concepts. To address this, we introduce DRIFT, a novel framework that enables LLMs to decompose informal mathematical statements into smaller, more tractable ‘‘sub-components’’. This facilitates targeted retrieval of premises from mathematical libraries such as Mathlib. Additionally, DRIFT retrieves illustrative theorems to help models use premises more effectively in formalization tasks. We evaluate DRIFT across diverse benchmarks (ProofNet, ConNF, and MiniF2F-test) and find that it consistently improves premise retrieval, nearly doubling the F1 score compared to the DPR baseline on ProofNet. Notably, DRIFT demonstrates strong performance on the out-of-distribution ConNF benchmark, with BEq+@10 improvements of 37.14% and 42.25% using GPT-4.1 and DeepSeek-V3.1, respectively. Our analysis shows that retrieval effectiveness in mathematical autoformalization depends heavily on model-specific knowledge boundaries, highlighting the need for adaptive retrieval strategies aligned with each model’s capabilities.

[577] Agentic Design of Compositional Machines

Wenqian Zhang, Weiyang Liu, Zhen Liu

Main category: cs.AI

TL;DR: This paper investigates whether large language models can learn machine design by assembling standardized components in a simulated physical environment, using a new testbed called BesiegeField.

DetailsMotivation: To explore if large language models can demonstrate creative intelligence in engineering design tasks, specifically compositional machine design where machines are built from parts to meet functional requirements.

Method: Introduced BesiegeField testbed built on Besiege game for part-based construction and physical simulation. Benchmarked LLMs with agentic workflows and used reinforcement learning finetuning with a curated cold-start dataset.

Result: Identified key capabilities needed for success (spatial reasoning, strategic assembly, instruction-following) and found current open-source models fall short, highlighting challenges in language, machine design, and physical reasoning.

Conclusion: Reinforcement learning shows promise for improving LLMs in machine design tasks, but significant challenges remain at the intersection of language, design, and physical reasoning.

Abstract: The design of complex machines stands as both a marker of human intelligence and a foundation of engineering practice. Given recent advances in large language models (LLMs), we ask whether they, too, can learn to create. We approach this question through the lens of compositional machine design: a task in which machines are assembled from standardized components to meet functional demands like locomotion or manipulation in a simulated physical environment. With this simplification, machine design is expressed as writing XML-like code that explicitly specifies pairwise part connections. To support this investigation, we introduce BesiegeField, a testbed built on the machine-building game Besiege, which enables part-based construction, physical simulation and reward-driven evaluation. Using BesiegeField, we benchmark state-of-the-art LLMs with agentic workflows and identify key capabilities required for success, including spatial reasoning, strategic assembly, and instruction-following. As current open-source models fall short, we explore reinforcement learning (RL) as a path to improvement: we curate a cold-start dataset, conduct RL finetuning experiments, and highlight open challenges at the intersection of language, machine design, and physical reasoning.

[578] PokeeResearch: Effective Deep Research via Reinforcement Learning from AI Feedback and Robust Reasoning Scaffold

Yi Wan, Jiuqi Wang, Liam Li, Jinsong Liu, Ruihao Zhu, Zheqing Zhu

Main category: cs.AI

TL;DR: PokeeResearch-7B is a 7B-parameter deep research agent that uses reinforcement learning and chain-of-thought reasoning to achieve state-of-the-art performance on research benchmarks.

DetailsMotivation: Current research agents have limitations including shallow retrieval, weak alignment metrics, and brittle tool-use behavior that need to be addressed.

Method: Uses annotation-free Reinforcement Learning from AI Feedback (RLAIF) with LLM-based reward signals for factual accuracy, citation faithfulness, and instruction adherence. Also employs chain-of-thought-driven multi-call reasoning scaffold with self-verification and adaptive recovery from tool failures.

Result: Achieves state-of-the-art performance among 7B-scale deep research agents across 10 popular deep research benchmarks.

Conclusion: Careful reinforcement learning and reasoning design can produce efficient, resilient, and research-grade AI agents.

Abstract: Tool-augmented large language models (LLMs) are emerging as deep research agents, systems that decompose complex queries, retrieve external evidence, and synthesize grounded responses. Yet current agents remain limited by shallow retrieval, weak alignment metrics, and brittle tool-use behavior. We introduce PokeeResearch-7B, a 7B-parameter deep research agent built under a unified reinforcement learning framework for robustness, alignment, and scalability. PokeeResearch-7B is trained by an annotation-free Reinforcement Learning from AI Feedback (RLAIF) framework to optimize policies using LLM-based reward signals that capture factual accuracy, citation faithfulness, and instruction adherence. A chain-of-thought-driven multi-call reasoning scaffold further enhances robustness through self-verification and adaptive recovery from tool failures. Among 10 popular deep research benchmarks, PokeeResearch-7B achieves state-of-the-art performance among 7B-scale deep research agents. This highlights that careful reinforcement learning and reasoning design can produce efficient, resilient, and research-grade AI agents. The model and inference code is open-sourced under MIT license at https://github.com/Pokee-AI/PokeeResearchOSS.

cs.SD

[579] MuseTok: Symbolic Music Tokenization for Generation and Semantic Understanding

Jingyue Huang, Zachary Novack, Phillip Long, Yupeng Hou, Ke Chen, Taylor Berg-Kirkpatrick, Julian McAuley

Main category: cs.SD

TL;DR: MuseTok is a tokenization method for symbolic music using RQ-VAE in a Transformer framework, achieving high-fidelity reconstruction and accurate music theory understanding across generation and understanding tasks.

DetailsMotivation: To develop effective discrete representation learning for symbolic music that works well for both generation and understanding tasks, inspired by advances in image, speech, and language domains.

Method: Uses residual vector quantized-variational autoencoder (RQ-VAE) on bar-wise music segments within a Transformer-based encoder-decoder framework to produce music codes.

Result: Outperforms previous representation learning baselines in semantic understanding tasks (melody extraction, chord recognition, emotion recognition) while maintaining comparable performance in content generation. Qualitative analysis shows effective capture of musical concepts.

Conclusion: MuseTok successfully creates discrete representations that enable both high-quality music generation and accurate music understanding, demonstrating its effectiveness across multiple musical tasks.

Abstract: Discrete representation learning has shown promising results across various domains, including generation and understanding in image, speech and language. Inspired by these advances, we propose MuseTok, a tokenization method for symbolic music, and investigate its effectiveness in both music generation and understanding tasks. MuseTok employs the residual vector quantized-variational autoencoder (RQ-VAE) on bar-wise music segments within a Transformer-based encoder-decoder framework, producing music codes that achieve high-fidelity music reconstruction and accurate understanding of music theory. For comprehensive evaluation, we apply MuseTok to music generation and semantic understanding tasks, including melody extraction, chord recognition, and emotion recognition. Models incorporating MuseTok outperform previous representation learning baselines in semantic understanding while maintaining comparable performance in content generation. Furthermore, qualitative analyses on MuseTok codes, using ground-truth categories and synthetic datasets, reveal that MuseTok effectively captures underlying musical concepts from large music collections.

[580] Transmission of High-Amplitude Sound through Leakages of Ill-fitting Earplugs

Haocheng Yu, Krishan K. Ahuja, Lakshmi N. Sankar, Spencer H. Bryngelson

Main category: cs.SD

TL;DR: This study investigates sound leakage in ill-fitting earplugs under high sound pressure levels (120-150 dB), showing significant transmission loss reduction and acoustic energy dissipation through vorticity conversion.

DetailsMotivation: High sound pressure levels in loud environments pose risks of noise-induced hearing loss, and ill-fitting earplugs often lead to sound leakage that compromises hearing protection.

Method: Used computational and experimental approaches to analyze acoustic transmission data for various leakage geometries, including different orifice diameters, across 1-5 kHz frequency range under 120-150 dB SPL conditions.

Result: Unsealed silicone rubber earplugs showed average transmission loss reduction of ~18 dB at 120 dB OISPL. Numerical simulations revealed SPL-dependent acoustic dissipation with acoustic energy converting to vorticity at 150 dB OISPL in ill-fitting models.

Conclusion: Earplug design plays a critical role in effective hearing protection in high-sound-pressure-level environments, with proper sealing being essential to prevent sound leakage and maintain adequate transmission loss.

Abstract: High sound pressure levels (SPL) pose notable risks in loud environments, particularly due to noise-induced hearing loss. Ill-fitting earplugs often lead to sound leakage, a phenomenon this study seeks to investigate. To validate our methodology, we first obtained computational and experimental acoustic transmission data for stand-alone slit resonators and orifices, for which extensive published data are readily available for comparison. We then examined the frequency-dependent acoustic power absorption coefficient and transmission loss (TL) across various leakage geometries, modeled using different orifice diameters. Experimental approaches spanned a frequency range of 1–5 kHz under SPL conditions of 120–150 dB. Key findings reveal that unsealed silicone rubber earplugs demonstrate an average TL reduction of approximately 18 dB at an overall incident SPL (OISPL) of 120 dB. Direct numerical simulations further highlight SPL-dependent acoustic dissipation mechanisms, showing the conversion of acoustic energy into vorticity in ill-fitting earplug models at an OISPL of 150 dB. These results highlight the role of earplug design for high-sound-pressure-level environments.

[581] Interpreting the Dimensions of Speaker Embedding Space

Mark Huckvale

Main category: cs.SD

TL;DR: Speaker embeddings capture acoustic characteristics and gender but poorly represent speaker age, suggesting room for improvement.

DetailsMotivation: To understand how speaker embeddings relate to conventional acoustic dimensions, age, and gender, as they are often treated as black-box encodings.

Method: Analyzed 10,000 speakers using three embedding systems, comparing 9 interpretable acoustic parameters with 7 principal components to predict embeddings.

Result: 9 acoustic parameters predict embeddings similarly to 7 principal components (over 50% variance explained). Embeddings capture gender implicitly but perform poorly on age.

Conclusion: Speaker embeddings effectively represent acoustic characteristics and gender but fail to capture age, indicating potential for enhancement in embedding calculation methods.

Abstract: Speaker embeddings are widely used in speaker verification systems and other applications where it is useful to characterise the voice of a speaker with a fixed-length vector. These embeddings tend to be treated as “black box” encodings, and how they relate to conventional acoustic and phonetic dimensions of voices has not been widely studied. In this paper we investigate how state-of-the-art speaker embedding systems represent the acoustic characteristics of speakers as described by conventional acoustic descriptors, age, and gender. Using a large corpus of 10,000 speakers and three embedding systems we show that a small set of 9 acoustic parameters chosen to be “interpretable” predict embeddings about the same as 7 principal components, corresponding to over 50% of variance in the data. We show that some principal dimensions operate differently for male and female speakers, suggesting there is implicit gender recognition within the embedding systems. However we show that speaker age is not well captured by embeddings, suggesting opportunities exist for improvements in their calculation.

[582] Zero- and One-Shot Data Augmentation for Sentence-Level Dysarthric Speech Recognition in Constrained Scenarios

Shiyao Wang, Shiwan Zhao, Jiaming Zhou, Yong Qin

Main category: cs.SD

TL;DR: A novel text-coverage strategy for dysarthric data augmentation that enables efficient zero/one-shot learning, significantly improving dysarthric speech recognition performance for unseen speakers.

DetailsMotivation: The scarcity of dysarthric speech data poses a major challenge for developing effective sentence-level dysarthric speech recognition systems, especially in zero/one-shot learning scenarios where models struggle to generalize to new speakers due to wide pronunciation variability.

Method: Proposed a text-coverage strategy specifically designed for text-matching data synthesis, enabling efficient zero/one-shot dysarthric data augmentation.

Result: The approach leads to substantial enhancements in dysarthric speech recognition performance when dealing with unseen dysarthric speakers.

Conclusion: The improvements are significant for practical applications including dysarthria rehabilitation programs and daily communication scenarios, addressing the data scarcity problem in dysarthric speech recognition.

Abstract: Dysarthric speech recognition (DSR) research has witnessed remarkable progress in recent years, evolving from the basic understanding of individual words to the intricate comprehension of sentence-level expressions, all driven by the pressing communication needs of individuals with dysarthria. Nevertheless, the scarcity of available data remains a substantial hurdle, posing a significant challenge to the development of effective sentence-level DSR systems. In response to this issue, dysarthric data augmentation (DDA) has emerged as a highly promising approach. Generative models are frequently employed to generate training data for automatic speech recognition tasks. However, their effectiveness hinges on the ability of the synthesized data to accurately represent the target domain. The wide-ranging variability in pronunciation among dysarthric speakers makes it extremely difficult for models trained on data from existing speakers to produce useful augmented data, especially in zero-shot or one-shot learning settings. To address this limitation, we put forward a novel text-coverage strategy specifically designed for text-matching data synthesis. This innovative strategy allows for efficient zero/one-shot DDA, leading to substantial enhancements in the performance of DSR when dealing with unseen dysarthric speakers. Such improvements are of great significance in practical applications, including dysarthria rehabilitation programs and day-to-day common-sentence communication scenarios.

[583] U-Codec: Ultra Low Frame-rate Neural Speech Codec for Fast High-fidelity Speech Generation

Xusheng Yang, Long Zhou, Wenfu Wang, Kai Hu, Shulin Feng, Chenxing Li, Meng Yu, Dong Yu, Yuexian Zou

Main category: cs.SD

TL;DR: U-Codec is an ultra low frame-rate neural speech codec that achieves high-fidelity reconstruction and fast speech generation at 5Hz frame rate, improving TTS inference speed by 3x while maintaining quality.

DetailsMotivation: Extreme compression at 5Hz typically causes severe intelligibility and spectral detail loss, so there's a need for better methods to maintain quality at ultra low frame rates.

Method: Uses Transformer-based inter-frame long-term dependency module, explores RVQ depth and codebook size optimization, and integrates into LLM-based TTS with global and local hierarchical architecture for multi-layer token dependencies.

Result: Extends LLM-based TTS from 3-layer RVQ at 50Hz to 32-layer RVQ at 5Hz, achieving 3x faster inference speed while maintaining similarity and naturalness compared to high-frame-rate codecs.

Conclusion: Validates the feasibility of using highly compressed 5Hz discrete tokens for fast and high-fidelity speech synthesis.

Abstract: We propose \textbf{U-Codec}, an \textbf{U}ltra low frame-rate neural speech \textbf{Codec} that achieves high-fidelity reconstruction and fast speech generation at an extremely low frame-rate of 5Hz (5 frames per second). Extreme compression at 5Hz typically leads to severe intelligibility and spectral detail loss, we introduce a Transformer-based inter-frame long-term dependency module and systematically explore residual vector quantization (RVQ) depth and codebook size to identify optimal configurations. Moreover, we apply U-Codec into a large language model (LLM)-based auto-regressive TTS model, which leverages global and local hierarchical architecture to effectively capture dependencies across multi-layer tokens. We extend LLM-based TTS from 3-layer RVQ at 50Hz to 32-layer RVQ at 5Hz. Experimental results demonstrate that U-Codec improves LLM-based TTS inference speed by around 3 $\times$ over high-frame-rate codecs while maintaining similarity and naturalness. These results validate the feasibility of using highly compressed 5Hz discrete tokens for fast and high-fidelity speech synthesis.

[584] Schrödinger Bridge Mamba for One-Step Speech Enhancement

Jing Yang, Sirui Wang, Chao Wu, Fan Fan

Main category: cs.SD

TL;DR: Schrödinger Bridge Mamba (SBM) is a novel training-inference framework that combines Schrödinger Bridge training with Mamba’s selective state-space model, achieving superior speech enhancement performance with only 1-step inference.

DetailsMotivation: The motivation stems from the inherent compatibility between Schrödinger Bridge training paradigm and Mamba's selective state-space model architecture, suggesting a promising integration for efficient generative modeling.

Method: The method implements SBM for generative speech enhancement, combining Schrödinger Bridge training with Mamba’s selective state-space model to enable efficient 1-step inference.

Result: Experiments on joint denoising and dereverberation tasks across four benchmark datasets show SBM outperforms strong baselines with 1-step or iterative inference while achieving the best real-time factor (RTF).

Conclusion: The integration of Schrödinger Bridge paradigm with selective state-space model architecture represents a promising direction for developing new deep generative models applicable to a broad range of generative tasks.

Abstract: We propose Schr"odinger Bridge Mamba (SBM), a new concept of training-inference framework motivated by the inherent compatibility between Schr"odinger Bridge (SB) training paradigm and selective state-space model Mamba. We exemplify the concept of SBM with an implementation for generative speech enhancement. Experiments on a joint denoising and dereverberation task using four benchmark datasets demonstrate that SBM, with only 1-step inference, outperforms strong baselines with 1-step or iterative inference and achieves the best real-time factor (RTF). Beyond speech enhancement, we discuss the integration of SB paradigm and selective state-space model architecture based on their underlying alignment, which indicates a promising direction for exploring new deep generative models potentially applicable to a broad range of generative tasks. Demo page: https://sbmse.github.io

[585] Investigating Safety Vulnerabilities of Large Audio-Language Models Under Speaker Emotional Variations

Bo-Han Feng, Chien-Feng Liu, Yu-Hsuan Li Liang, Chih-Kai Yang, Szu-Wei Fu, Zhehuai Chen, Ke-Han Lu, Sung-Feng Huang, Chao-Han Huck Yang, Yu-Chiang Frank Wang, Yun-Nung Chen, Hung-yi Lee

Main category: cs.SD

TL;DR: LALMs show safety inconsistencies across different speaker emotions and intensities, with medium emotional expressions posing the greatest risk.

DetailsMotivation: To investigate the safety alignment of Large Audio-Language Models under paralinguistic variation, specifically speaker emotion, which remains underexplored despite widespread study of their other capabilities.

Method: Constructed a dataset of malicious speech instructions expressed across multiple emotions and intensities, then evaluated several state-of-the-art LALMs on this dataset.

Result: Revealed substantial safety inconsistencies: different emotions elicit varying levels of unsafe responses, and the effect of intensity is non-monotonic with medium expressions often posing the greatest risk.

Conclusion: Highlights an overlooked vulnerability in LALMs and calls for alignment strategies explicitly designed to ensure robustness under emotional variation for trustworthy real-world deployment.

Abstract: Large audio-language models (LALMs) extend text-based LLMs with auditory understanding, offering new opportunities for multimodal applications. While their perception, reasoning, and task performance have been widely studied, their safety alignment under paralinguistic variation remains underexplored. This work systematically investigates the role of speaker emotion. We construct a dataset of malicious speech instructions expressed across multiple emotions and intensities, and evaluate several state-of-the-art LALMs. Our results reveal substantial safety inconsistencies: different emotions elicit varying levels of unsafe responses, and the effect of intensity is non-monotonic, with medium expressions often posing the greatest risk. These findings highlight an overlooked vulnerability in LALMs and call for alignment strategies explicitly designed to ensure robustness under emotional variation, a prerequisite for trustworthy deployment in real-world settings.

[586] SAKE: Towards Editing Auditory Attribute Knowledge of Large Audio-Language Models

Chih-Kai Yang, Yen-Ting Piao, Tzu-Wen Hsu, Szu-Wei Fu, Zhehuai Chen, Ke-Han Lu, Sung-Feng Huang, Chao-Han Huck Yang, Yu-Chiang Frank Wang, Yun-Nung Chen, Hung-yi Lee

Main category: cs.SD

TL;DR: SAKE is the first benchmark for editing auditory attribute knowledge in Large Audio-Language Models, addressing challenges in multimodal knowledge editing beyond textual and visual domains.

DetailsMotivation: Prior knowledge editing work focused mainly on textual or visual modalities, leaving auditory knowledge editing unexplored despite its importance for real-world multimodal applications.

Method: Created SAKE benchmark targeting abstract auditory attributes, benchmarked seven editing methods on two LALMs across four dimensions: reliability, generality, audio/text locality, and portability.

Result: Results revealed significant challenges including preserving intra-attribute knowledge unrelated to edits, generalizing edits to multimodal reasoning, and maintaining edits under sequential updates.

Conclusion: SAKE provides a principled framework for studying auditory knowledge editing, opening new directions for maintaining and adapting LALMs in diverse real-world scenarios.

Abstract: Knowledge editing offers an efficient way to update model knowledge without full retraining, but prior work has concentrated almost exclusively on textual or visual modalities. We introduce SAKE, the first benchmark specifically designed for editing auditory attribute knowledge in Large Audio-Language Models (LALMs). Unlike factual updates, SAKE targets several abstract auditory attributes, capturing knowledge types that go beyond conventional textual and visual domains. We benchmark seven editing methods on two LALMs along four dimensions: reliability, generality, audio/text locality, and portability. Results highlight challenges such as preserving intra-attribute knowledge unrelated to the edit, generalizing edits to multimodal reasoning, and maintaining edits under sequential updates. SAKE provides a principled framework to study how knowledge editing extends to the auditory modalities, opening new directions for maintaining and adapting LALMs in more diverse real-world scenarios.

[587] DDSC: Dynamic Dual-Signal Curriculum for Data-Efficient Acoustic Scene Classification under Domain Shift

Peihong Zhang, Yuxuan Liu, Rui Sang, Zhixin Li, Yiqiang Cai, Yizhou Tan, Shengchen Li

Main category: cs.SD

TL;DR: DDSC is a dynamic curriculum learning method that adaptively weights training examples using domain-invariance and learning-progress signals to address device-induced domain shift in acoustic scene classification.

DetailsMotivation: Existing curriculum learning methods for ASC are static and don't adapt to evolving example difficulty and marginal utility during training, limiting their effectiveness against device-induced domain shift.

Method: Proposes Dynamic Dual-Signal Curriculum (DDSC) that combines domain-invariance and learning-progress signals computed each epoch, fused through a time-varying scheduler to create per-example weights that prioritize domain-invariant examples early and device-specific cases later.

Result: DDSC consistently improves cross-device performance across diverse ASC baselines and label budgets, with largest gains on unseen-device splits under DCASE 2024 Task 1 protocol.

Conclusion: DDSC is an effective, lightweight, architecture-agnostic solution that adapts curriculum learning online to handle device-induced domain shift in ASC without additional inference overhead.

Abstract: Acoustic scene classification (ASC) suffers from device-induced domain shift, especially when labels are limited. Prior work focuses on curriculum-based training schedules that structure data presentation by ordering or reweighting training examples from easy-to-hard to facilitate learning; however, existing curricula are static, fixing the ordering or the weights before training and ignoring that example difficulty and marginal utility evolve with the learned representation. To overcome this limitation, we propose the Dynamic Dual-Signal Curriculum (DDSC), a training schedule that adapts the curriculum online by combining two signals computed each epoch: a domain-invariance signal and a learning-progress signal. A time-varying scheduler fuses these signals into per-example weights that prioritize domain-invariant examples in early epochs and progressively emphasize device-specific cases. DDSC is lightweight, architecture-agnostic, and introduces no additional inference overhead. Under the official DCASE 2024 Task~1 protocol, DDSC consistently improves cross-device performance across diverse ASC baselines and label budgets, with the largest gains on unseen-device splits.

[588] TopSeg: A Multi-Scale Topological Framework for Data-Efficient Heart Sound Segmentation

Peihong Zhang, Zhixin Li, Yuxuan Liu, Rui Sang, Yiqiang Cai, Yizhou Tan, Shengchen Li

Main category: cs.SD

TL;DR: TopSeg uses topological features for heart sound segmentation, achieving better data efficiency and cross-dataset generalization than spectrogram-based methods, especially with limited training data.

DetailsMotivation: Current deep learning methods for PCG segmentation rely on large labeled datasets and time-frequency features, limiting robustness and deployment in data-scarce scenarios.

Method: TopSeg framework encodes PCG dynamics with multi-scale topological features and decodes them using a lightweight TCN with order- and duration-constrained inference, trained exclusively on PhysioNet 2016 with external validation on CirCor.

Result: Topological features consistently outperform spectrogram and envelope inputs, with largest margins at low data budgets. Full system surpasses end-to-end baselines while remaining competitive at full data. Combining H_0 and H_1 features improves S1/S2 localization and boundary stability.

Conclusion: Topology-aware representations provide strong inductive bias for data-efficient, cross-dataset PCG segmentation, supporting practical use when labeled data are limited.

Abstract: Deep learning approaches for heart-sound (PCG) segmentation built on time–frequency features can be accurate but often rely on large expert-labeled datasets, limiting robustness and deployment. We present TopSeg, a topological representation-centric framework that encodes PCG dynamics with multi-scale topological features and decodes them using a lightweight temporal convolutional network (TCN) with an order- and duration-constrained inference step. To evaluate data efficiency and generalization, we train exclusively on PhysioNet 2016 dataset with subject-level subsampling and perform external validation on CirCor dataset. Under matched-capacity decoders, the topological features consistently outperform spectrogram and envelope inputs, with the largest margins at low data budgets; as a full system, TopSeg surpasses representative end-to-end baselines trained on their native inputs under the same budgets while remaining competitive at full data. Ablations at 10% training confirm that all scales contribute and that combining H_0 and H_1 yields more reliable S1/S2 localization and boundary stability. These results indicate that topology-aware representations provide a strong inductive bias for data-efficient, cross-dataset PCG segmentation, supporting practical use when labeled data are limited.

[589] AWARE: Audio Watermarking with Adversarial Resistance to Edits

Kosta Pavlović, Lazar Stanarević, Petar Nedić, Slavko Kovačević, Igor Djurović

Main category: cs.SD

TL;DR: AWARE is a learning-based audio watermarking system that uses adversarial optimization instead of attack simulation, with a time-order-agnostic detector that achieves robust performance against various audio edits.

DetailsMotivation: Current learning-based audio watermarking methods rely on simulated distortions during training, which are narrow and prone to overfitting, limiting their robustness to real-world attacks.

Method: Uses adversarial optimization in time-frequency domain with level-proportional perceptual budget. Detection employs time-order-agnostic detector with Bitwise Readout Head (BRH) that aggregates temporal evidence per watermark bit.

Result: Achieves high audio quality and speech intelligibility (PESQ/STOI) with consistently low BER across various audio edits, outperforming state-of-the-art learning-based systems.

Conclusion: AWARE provides a robust alternative to attack-simulation approaches by using adversarial optimization and temporal evidence aggregation, demonstrating superior performance against diverse audio edits.

Abstract: Prevailing practice in learning-based audio watermarking is to pursue robustness by expanding the set of simulated distortions during training. However, such surrogates are narrow and prone to overfitting. This paper presents AWARE (Audio Watermarking with Adversarial Resistance to Edits), an alternative approach that avoids reliance on attack-simulation stacks and handcrafted differentiable distortions. Embedding is obtained via adversarial optimization in the time-frequency domain under a level-proportional perceptual budget. Detection employs a time-order-agnostic detector with a Bitwise Readout Head (BRH) that aggregates temporal evidence into one score per watermark bit, enabling reliable watermark decoding even under desynchronization and temporal cuts. Empirically, AWARE attains high audio quality and speech intelligibility (PESQ/STOI) and consistently low BER across various audio edits, often surpassing representative state-of-the-art learning-based audio watermarking systems.

[590] Not All Deepfakes Are Created Equal: Triaging Audio Forgeries for Robust Deepfake Singer Identification

Davide Salvi, Hendrik Vincent Koops, Elio Quinton

Main category: cs.SD

TL;DR: A two-stage pipeline for singer identification in vocal deepfakes that filters low-quality forgeries first, then identifies singers in high-quality deepfakes using models trained only on authentic recordings.

DetailsMotivation: To protect artist likeness and content authenticity against highly realistic singing voice deepfakes by enabling automatic singer identification as a defense mechanism.

Method: Two-stage pipeline: 1) Discriminator model filters out low-quality forgeries that don’t accurately reproduce vocal likeness, 2) Subsequent model trained only on authentic recordings identifies singers in remaining high-quality deepfakes and authentic audio.

Result: The system consistently outperforms existing baselines on both authentic and synthetic content.

Conclusion: The proposed two-stage approach effectively addresses singer identification in vocal deepfakes by focusing on high-quality forgeries and leveraging authentic training data.

Abstract: The proliferation of highly realistic singing voice deepfakes presents a significant challenge to protecting artist likeness and content authenticity. Automatic singer identification in vocal deepfakes is a promising avenue for artists and rights holders to defend against unauthorized use of their voice, but remains an open research problem. Based on the premise that the most harmful deepfakes are those of the highest quality, we introduce a two-stage pipeline to identify a singer’s vocal likeness. It first employs a discriminator model to filter out low-quality forgeries that fail to accurately reproduce vocal likeness. A subsequent model, trained exclusively on authentic recordings, identifies the singer in the remaining high-quality deepfakes and authentic audio. Experiments show that this system consistently outperforms existing baselines on both authentic and synthetic content.

[591] SARSteer: Safeguarding Large Audio Language Models via Safe-Ablated Refusal Steering

Weilin Lin, Jianze Li, Hui Xiong, Li Liu

Main category: cs.SD

TL;DR: SARSteer is an inference-time defense framework for Large Audio-Language Models that uses text-derived refusal steering and safe-space ablation to improve safety alignment without causing over-refusal on benign queries.

DetailsMotivation: Audio inputs can more easily elicit harmful responses than text, creating new safety risks for LALMs. Existing safety alignment methods from LLMs and LVLMs fail when adapted to LALMs due to distributional gaps in activations and cause over-refusal on benign queries.

Method: Proposes Safe-Ablated Refusal Steering (SARSteer) which leverages text-derived refusal steering to enforce rejection without manipulating audio inputs and introduces decomposed safe-space ablation to mitigate over-refusal.

Result: Extensive experiments show SARSteer significantly improves harmful-query refusal while preserving benign responses.

Conclusion: SARSteer establishes a principled step toward safety alignment in LALMs, addressing the unique challenges of audio modality safety.

Abstract: Large Audio-Language Models (LALMs) are becoming essential as a powerful multimodal backbone for real-world applications. However, recent studies show that audio inputs can more easily elicit harmful responses than text, exposing new risks toward deployment. While safety alignment has made initial advances in LLMs and Large Vision-Language Models (LVLMs), we find that vanilla adaptation of these approaches to LALMs faces two key limitations: 1) LLM-based steering fails under audio input due to the large distributional gap between activations, and 2) prompt-based defenses induce over-refusals on benign-speech queries. To address these challenges, we propose Safe-Ablated Refusal Steering (SARSteer), the first inference-time defense framework for LALMs. Specifically, SARSteer leverages text-derived refusal steering to enforce rejection without manipulating audio inputs and introduces decomposed safe-space ablation to mitigate over-refusal. Extensive experiments demonstrate that SARSteer significantly improves harmful-query refusal while preserving benign responses, establishing a principled step toward safety alignment in LALMs.

[592] DELULU: Discriminative Embedding Learning Using Latent Units for Speaker-Aware Self-Supervised Speech Foundational Model

Massa Baali, Rita Singh, Bhiksha Raj

Main category: cs.SD

TL;DR: DELULU is a speaker-aware self-supervised model that integrates external speaker verification embeddings into pseudo-label generation, achieving significant improvements on speaker-centric tasks without task-specific fine-tuning.

DetailsMotivation: Current self-supervised speech models excel at content-driven tasks but struggle to capture speaker-discriminative features needed for verification, diarization, and profiling applications.

Method: Integrates frame-level embeddings from ReDimNet (state-of-the-art speaker verification model) to guide k-means clustering during pre-training, using dual objective of masked prediction and denoising for enhanced robustness.

Result: Achieves up to 62% relative improvement in EER for speaker verification and consistent gains on zero-shot profiling tasks (gender, age, accent, speaker counting).

Conclusion: DELULU serves as a strong universal encoder for speaker-aware speech processing, enabling superior performance across speaker-centric tasks without requiring task-specific fine-tuning.

Abstract: Self-supervised speech models have achieved remarkable success on content-driven tasks, yet they remain limited in capturing speaker-discriminative features critical for verification, diarization, and profiling applications. We introduce DELULU, a speaker-aware self-supervised foundational model that addresses this limitation by integrating external supervision into the pseudo-label generation process. DELULU leverages frame-level embeddings from ReDimNet, a state-of-the-art speaker verification model, to guide the k-means clustering step during pre-training, introducing a strong speaker-discriminative inductive bias that aligns representation learning with speaker identity. The model is trained using a dual objective that combines masked prediction and denoising, further enhancing robustness and generalization. DELULU significantly outperforms prior self-supervised learning (SSL) models across a range of speaker-centric tasks, achieving up to 62% relative improvement in equal error rate (EER) for speaker verification and consistent gains on zero-shot profiling tasks such as gender, age, accent, and speaker counting. Our findings demonstrate that DELULU is a strong universal encoder for speaker-aware speech processing, enabling superior performance even without task-specific fine-tuning.

[593] MGE-LDM: Joint Latent Diffusion for Simultaneous Music Generation and Source Extraction

Yunkee Chae, Kyogu Lee

Main category: cs.SD

TL;DR: MGE-LDM is a unified latent diffusion framework that enables simultaneous music generation, source imputation, and query-driven source separation in a single model, supporting flexible manipulation of arbitrary instrument sources without predefined categories.

DetailsMotivation: Prior approaches were constrained to fixed instrument classes, limiting flexibility in music source manipulation. The motivation is to create a unified framework that can handle multiple tasks (generation, imputation, separation) without relying on predefined instrument categories.

Method: Uses a latent diffusion model that learns a joint distribution over full mixtures, submixtures, and individual stems. Formulates separation and imputation as conditional inpainting tasks in the latent space, enabling class-agnostic manipulation of arbitrary instrument sources.

Result: MGE-LDM can perform complete mixture generation, partial generation (source imputation), and text-conditioned extraction of arbitrary sources. It can be trained jointly across heterogeneous multi-track datasets without predefined instrument categories.

Conclusion: The framework provides a unified solution for multiple music manipulation tasks, supporting flexible, class-agnostic handling of arbitrary instrument sources through latent space conditional inpainting.

Abstract: We present MGE-LDM, a unified latent diffusion framework for simultaneous music generation, source imputation, and query-driven source separation. Unlike prior approaches constrained to fixed instrument classes, MGE-LDM learns a joint distribution over full mixtures, submixtures, and individual stems within a single compact latent diffusion model. At inference, MGE-LDM enables (1) complete mixture generation, (2) partial generation (i.e., source imputation), and (3) text-conditioned extraction of arbitrary sources. By formulating both separation and imputation as conditional inpainting tasks in the latent space, our approach supports flexible, class-agnostic manipulation of arbitrary instrument sources. Notably, MGE-LDM can be trained jointly across heterogeneous multi-track datasets (e.g., Slakh2100, MUSDB18, MoisesDB) without relying on predefined instrument categories. Audio samples are available at our project page: https://yoongi43.github.io/MGELDM_Samples/.

[594] CoVoMix2: Advancing Zero-Shot Dialogue Generation with Fully Non-Autoregressive Flow Matching

Leying Zhang, Yao Qian, Xiaofei Wang, Manthan Thakker, Dongmei Wang, Jianwei Yu, Haibin Wu, Yuxuan Hu, Jinyu Li, Yanmin Qian, Sheng Zhao

Main category: cs.SD

TL;DR: CoVoMix2 is a non-autoregressive framework for zero-shot multi-speaker dialogue generation that directly predicts mel-spectrograms from multi-stream transcriptions using flow matching, achieving state-of-the-art performance in speech quality and speaker consistency.

DetailsMotivation: Existing systems struggle with maintaining speaker consistency, modeling overlapping speech, and synthesizing coherent conversations efficiently for applications like podcast creation and virtual agents.

Method: Uses flow-matching-based generative model to directly predict mel-spectrograms from multi-stream transcriptions, with transcription-level speaker disentanglement, sentence-level alignment, and prompt-level random masking strategies.

Result: Achieves state-of-the-art performance, outperforming baselines like MoonCast and Sesame in speech quality, speaker consistency, and inference speed. Supports controllable dialogue generation with overlapping speech and timing control.

Conclusion: CoVoMix2 demonstrates strong generalizability to real-world speech generation scenarios without requiring prompt transcriptions, enabling efficient and controllable multi-speaker dialogue synthesis.

Abstract: Generating natural-sounding, multi-speaker dialogue is crucial for applications such as podcast creation, virtual agents, and multimedia content generation. However, existing systems struggle to maintain speaker consistency, model overlapping speech, and synthesize coherent conversations efficiently. In this paper, we introduce CoVoMix2, a fully non-autoregressive framework for zero-shot multi-talker dialogue generation. CoVoMix2 directly predicts mel-spectrograms from multi-stream transcriptions using a flow-matching-based generative model, eliminating the reliance on intermediate token representations. To better capture realistic conversational dynamics, we propose transcription-level speaker disentanglement, sentence-level alignment, and prompt-level random masking strategies. Our approach achieves state-of-the-art performance, outperforming strong baselines like MoonCast and Sesame in speech quality, speaker consistency, and inference speed. Notably, CoVoMix2 operates without requiring transcriptions for the prompt and supports controllable dialogue generation, including overlapping speech and precise timing control, demonstrating strong generalizability to real-world speech generation scenarios.

cs.LG

[595] Lean Finder: Semantic Search for Mathlib That Understands User Intents

Jialin Lu, Kye Emond, Kaiyu Yang, Swarat Chaudhuri, Weiran Sun, Wuyang Chen

Main category: cs.LG

TL;DR: Lean Finder is a semantic search engine for Lean and mathlib that understands mathematician intents through fine-tuned embeddings and user feedback, achieving 30%+ improvement over previous methods.

DetailsMotivation: Progress in formal theorem proving is hindered by difficulty locating relevant theorems and Lean 4's steep learning curve, with existing search engines overlooking real-world user query mismatches.

Method: Analyze and cluster semantics of public Lean discussions, fine-tune text embeddings on synthesized queries emulating user intents, and align with mathematician preferences using diverse feedback signals.

Result: Achieves over 30% relative improvement on real-world queries, informalized statements, and proof states compared to previous search engines and GPT-4o.

Conclusion: Lean Finder provides user-centered semantic search tailored to mathematicians’ needs and is compatible with LLM-based theorem provers, bridging retrieval with formal reasoning.

Abstract: We present Lean Finder, a semantic search engine for Lean and mathlib that understands and aligns with the intents of mathematicians. Progress in formal theorem proving is often hindered by the difficulty of locating relevant theorems and the steep learning curve of the Lean 4 language, making advancement slow and labor-intensive. Existing Lean search engines, though helpful, rely primarily on informalizations (natural language translation of the formal statements), while largely overlooking the mismatch with real-world user queries. In contrast, we propose a user-centered semantic search tailored to the needs of mathematicians. Our approach begins by analyzing and clustering the semantics of public Lean discussions, then fine-tuning text embeddings on synthesized queries that emulate user intents. We further align Lean Finder with mathematicians’ preferences using diverse feedback signals, encoding it with a rich awareness of their goals from multiple perspectives. Evaluations on real-world queries, informalized statements, and proof states demonstrate that our Lean Finder achieves over $30%$ relative improvement compared to previous search engines and GPT-4o. In addition, Lean Finder is compatible with LLM-based theorem provers, bridging retrieval with formal reasoning. Lean Finder is available at: https://leanfinder.github.io

[596] Lyapunov-Stable Adaptive Control for Multimodal Concept Drift

Tianyu Bell Pan, Mengdi Zhu, Alexa Jordyn Cole, Ronald Wilson, Damon L. Woodard

Main category: cs.LG

TL;DR: LS-OGD is an adaptive control framework for robust multimodal learning that dynamically adjusts learning rates and fusion weights to handle concept drift, ensuring prediction error remains bounded and system resilience.

DetailsMotivation: Multimodal learning systems struggle with concept drift in non-stationary environments, particularly modality-specific drifts and lack of continuous adaptation mechanisms, leading to performance degradation.

Method: LS-OGD uses an online controller that dynamically adjusts the model’s learning rate and fusion weights between different data modalities based on detected drift and prediction errors.

Result: Under bounded drift conditions, LS-OGD ensures prediction error is uniformly ultimately bounded and converges to zero if drift ceases. The adaptive fusion strategy effectively isolates and mitigates severe modality-specific drift.

Conclusion: LS-OGD provides theoretical guarantees for developing reliable, continuously adapting multimodal learning systems with resilience and fault tolerance against concept drift.

Abstract: Multimodal learning systems often struggle in non-stationary environments due to concept drift, where changing data distributions can degrade performance. Modality-specific drifts and the lack of mechanisms for continuous, stable adaptation compound this challenge. This paper introduces LS-OGD, a novel adaptive control framework for robust multimodal learning in the presence of concept drift. LS-OGD uses an online controller that dynamically adjusts the model’s learning rate and the fusion weights between different data modalities in response to detected drift and evolving prediction errors. We prove that under bounded drift conditions, the LS-OGD system’s prediction error is uniformly ultimately bounded and converges to zero if the drift ceases. Additionally, we demonstrate that the adaptive fusion strategy effectively isolates and mitigates the impact of severe modality-specific drift, thereby ensuring system resilience and fault tolerance. These theoretical guarantees establish a principled foundation for developing reliable and continuously adapting multimodal learning systems.

[597] BEACON: Bayesian Optimal Stopping for Efficient LLM Sampling

Guangya Wan, Zixin Stephen Xu, Sasa Zorc, Manel Baucells, Mengxuan Hu, Hao Wang, Sheng Li

Main category: cs.LG

TL;DR: BEACON is a Bayesian adaptive sampling framework that dynamically determines when to stop generating LLM responses by balancing accuracy gains against computational costs, reducing sampling by up to 80% while maintaining quality.

DetailsMotivation: Sampling multiple responses improves LLM output quality but incurs significant computational costs. The challenge is deciding when to stop sampling to balance accuracy gains with efficiency.

Method: BEACON uses Sequential Search with Bayesian Learning to sequentially generate responses, update posterior belief over reward distributions in real time without training, and determine stopping points by weighing expected gains against computational cost.

Result: BEACON reduces average sampling by up to 80% while maintaining response quality, and demonstrates utility for cost-efficient preference data generation.

Conclusion: BEACON provides a principled framework for adaptive LLM sampling with theoretical optimality guarantees and practical tractability, offering actionable insights for efficient LLM deployment.

Abstract: Sampling multiple responses is a common way to improve LLM output quality, but it comes at the cost of additional computation. The key challenge is deciding when to stop generating new samples to balance accuracy gains against efficiency. To address this, we introduce BEACON (Bayesian Efficient Adaptive Criterion for Optimal N-stopping), a principled adaptive sampling framework grounded in Sequential Search with Bayesian Learning. BEACON sequentially generates responses from the policy LLM, updates posterior belief over reward distributions in real time without further training, and determines when to stop by weighing expected gains against computational cost. Sampling terminates once the marginal utility of further exploration no longer justifies the expense. We establish both theoretical optimality guarantees and practical tractability, and show empirically that BEACON reduces average sampling by up to 80% while maintaining response quality. We further demonstrate BEACON’s utility for cost-efficient preference data generation and outline practical extensions, offering actionable insights for future researchers.

[598] Learning from Mistakes: Enhancing Harmful Meme Detection via Misjudgment Risk Patterns

Wenshuo Wang, Ziyou Jiang, Junjie Wang, Mingyang Li, Jie Huang, Yuekai Huang, Zhiyuan Chang, Feiyan Duan, Qing Wang

Main category: cs.LG

TL;DR: PatMD improves harmful meme detection by proactively identifying and mitigating misjudgment risks through pattern-based guidance for MLLMs, achieving significant performance gains over state-of-the-art methods.

DetailsMotivation: Internet memes are increasingly weaponized with subtle harmful content using rhetorical devices like irony and metaphor, which existing detection approaches struggle to identify accurately, leading to frequent misjudgments.

Method: PatMD constructs a knowledge base where memes are deconstructed into misjudgment risk patterns, then retrieves relevant patterns to dynamically guide MLLM reasoning and avoid known misjudgment pitfalls.

Result: Experiments on 6,626 memes across 5 harmful detection tasks show PatMD outperforms state-of-the-art baselines with 8.30% F1-score improvement and 7.71% accuracy improvement.

Conclusion: PatMD demonstrates strong generalizability and improved detection capability for harmful memes by proactively addressing misjudgment risks through pattern-based guidance.

Abstract: Internet memes have emerged as a popular multimodal medium, yet they are increasingly weaponized to convey harmful opinions through subtle rhetorical devices like irony and metaphor. Existing detection approaches, including MLLM-based techniques, struggle with these implicit expressions, leading to frequent misjudgments. This paper introduces PatMD, a novel approach that improves harmful meme detection by learning from and proactively mitigating these potential misjudgment risks. Our core idea is to move beyond superficial content-level matching and instead identify the underlying misjudgment risk patterns, proactively guiding the MLLMs to avoid known misjudgment pitfalls. We first construct a knowledge base where each meme is deconstructed into a misjudgment risk pattern explaining why it might be misjudged, either overlooking harmful undertones (false negative) or overinterpreting benign content (false positive). For a given target meme, PatMD retrieves relevant patterns and utilizes them to dynamically guide the MLLM’s reasoning. Experiments on a benchmark of 6,626 memes across 5 harmful detection tasks show that PatMD outperforms state-of-the-art baselines, achieving an average of 8.30% improvement in F1-score and 7.71% improvement in accuracy, demonstrating strong generalizability and improved detection capability of harmful memes.

[599] WaveNet’s Precision in EEG Classification

Casper van Laar, Khubaib Ahmed

Main category: cs.LG

TL;DR: WaveNet-based deep learning model for automated EEG signal classification into physiological, pathological, artifact, and noise categories, achieving higher accuracy than CNN/LSTM approaches.

DetailsMotivation: Traditional EEG classification methods relying on expert visual review are impractical due to growing complexity and volume of EEG recordings.

Method: Used WaveNet architecture with dilated causal convolutions and residual connections on 209,232 EEG samples from Mayo Clinic and St. Anne’s University Hospital, with 70/20/10 train/validation/test split.

Result: Achieved classification accuracy exceeding previous CNN and LSTM approaches, with high precision for noise/artifact distinction but modest misclassification between physiological/pathological signals due to clinical overlap.

Conclusion: WaveNet is well-suited for EEG data analysis due to its ability to capture both fine-grained and long-range temporal dependencies, offering an effective automated classification solution.

Abstract: This study introduces a WaveNet-based deep learning model designed to automate the classification of EEG signals into physiological, pathological, artifact, and noise categories. Traditional methods for EEG signal classification, which rely on expert visual review, are becoming increasingly impractical due to the growing complexity and volume of EEG recordings. Leveraging a publicly available annotated dataset from Mayo Clinic and St. Anne’s University Hospital, the WaveNet model was trained, validated, and tested on 209,232 samples with a 70/20/10 percent split. The model achieved a classification accuracy exceeding previous CNN and LSTM-based approaches, and was benchmarked against a Temporal Convolutional Network (TCN) baseline. Notably, the model distinguishes noise and artifacts with high precision, although it reveals a modest but explainable degree of misclassification between physiological and pathological signals, reflecting inherent clinical overlap. WaveNet’s architecture, originally developed for raw audio synthesis, is well suited for EEG data due to its use of dilated causal convolutions and residual connections, enabling it to capture both fine-grained and long-range temporal dependencies. The research also details the preprocessing pipeline, including dynamic dataset partitioning and normalization steps that support model generalization.

[600] Cross-dataset Multivariate Time-series Model for Parkinson’s Diagnosis via Keyboard Dynamics

Arianna Francesconi, Donato Cappetta, Fabio Rebecchi, Paolo Soda, Valerio Guarrasi, Rosa Sicilia

Main category: cs.LG

TL;DR: A novel pipeline using keystroke dynamics for Parkinson’s disease screening achieves over 90% AUC-ROC in external validation, demonstrating strong potential as a digital biomarker for early detection and telemonitoring.

DetailsMotivation: Parkinson's disease affects over 10 million people globally with prevalence expected to double by 2040. Early diagnosis is challenging due to late motor symptom emergence and limitations of traditional clinical assessments, creating need for non-invasive, scalable remote screening methods.

Method: Three-stage pipeline: (1) preprocessing data from four datasets, extracting temporal signals and addressing class imbalance; (2) pre-training eight deep-learning architectures on largest datasets with hyperparameter optimization; (3) fine-tuning on intermediate dataset and external validation on independent cohort.

Result: Hybrid convolutional-recurrent and transformer models achieved strong external validation performance with AUC-ROC scores exceeding 90% and F1-Score over 70%. A temporal convolutional model attained 91.14% AUC-ROC in external validation, outperforming existing methods that rely solely on internal validation.

Conclusion: Keystroke dynamics show significant promise as a reliable digital biomarker for Parkinson’s disease, offering a viable approach for early detection and continuous remote monitoring through scalable, non-invasive technology.

Abstract: Parkinson’s disease (PD) presents a growing global challenge, affecting over 10 million individuals, with prevalence expected to double by 2040. Early diagnosis remains difficult due to the late emergence of motor symptoms and limitations of traditional clinical assessments. In this study, we propose a novel pipeline that leverages keystroke dynamics as a non-invasive and scalable biomarker for remote PD screening and telemonitoring. Our methodology involves three main stages: (i) preprocessing of data from four distinct datasets, extracting four temporal signals and addressing class imbalance through the comparison of three methods; (ii) pre-training eight state-of-the-art deep-learning architectures on the two largest datasets, optimizing temporal windowing, stride, and other hyperparameters; (iii) fine-tuning on an intermediate-sized dataset and performing external validation on a fourth, independent cohort. Our results demonstrate that hybrid convolutional-recurrent and transformer-based models achieve strong external validation performance, with AUC-ROC scores exceeding 90% and F1-Score over 70%. Notably, a temporal convolutional model attains an AUC-ROC of 91.14% in external validation, outperforming existing methods that rely solely on internal validation. These findings underscore the potential of keystroke dynamics as a reliable digital biomarker for PD, offering a promising avenue for early detection and continuous monitoring.

[601] Fire-EnSF: Wildfire Spread Data Assimilation using Ensemble Score Filter

Hongzheng Shi, Yuhang Wang, Xiao Liu

Main category: cs.LG

TL;DR: The paper applies Ensemble Score Filter (EnSF) to improve real-time wildfire spread predictions through data assimilation, showing superior accuracy and efficiency.

DetailsMotivation: Wildfires are increasingly destructive and expensive to control, requiring accurate real-time fire spread predictions. Data assimilation can enhance forecasting by integrating observations with numerical models.

Method: Uses Ensemble Score Filter (EnSF), a diffusion-model-based filtering algorithm that leverages score-based generative diffusion models for high-dimensional nonlinear filtering problems in wildfire spread models.

Result: Numerical investigations demonstrate that EnSF provides superior accuracy, stability, and computational efficiency for wildfire data assimilation compared to other methods.

Conclusion: EnSF is established as a robust and practical method for wildfire data assimilation, with publicly available code for implementation.

Abstract: As wildfires become increasingly destructive and expensive to control, effective management of active wildfires requires accurate, real-time fire spread predictions. To enhance the forecasting accuracy of active fires, data assimilation plays a vital role by integrating observations (such as remote-sensing data) and fire predictions generated from numerical models. This paper provides a comprehensive investigation on the application of a recently proposed diffusion-model-based filtering algorithm – the Ensemble Score Filter (EnSF) – to the data assimilation problem for real-time active wildfire spread predictions. Leveraging a score-based generative diffusion model, EnSF has been shown to have superior accuracy for high-dimensional nonlinear filtering problems, making it an ideal candidate for the filtering problems of wildfire spread models. Technical details are provided, and our numerical investigations demonstrate that EnSF provides superior accuracy, stability, and computational efficiency, establishing it as a robust and practical method for wildfire data assimilation. Our code has been made publicly available.

[602] How Good Are LLMs at Processing Tool Outputs?

Kiran Kate, Yara Rizk, Poulami Ghosh, Ashu Gulati, Tathagata Chakraborti, Zidane Wright, Mayank Agarwal

Main category: cs.LG

TL;DR: LLMs struggle with processing complex JSON responses from tool calls, with performance varying 3-50% based on processing strategy, output size, and reasoning complexity.

DetailsMotivation: Most realistic task automation requires LLMs to process complex JSON responses from tool calls, but this ability is under-studied despite being crucial for task completion.

Method: Created a dataset for tool response processing and evaluated 15 open/closed weight models using multiple prompting approaches to analyze JSON processing capabilities.

Result: JSON processing remains difficult even for frontier models across all prompting strategies. Performance varies significantly (3-50%) based on processing approach, output size, and reasoning complexity.

Conclusion: The optimal response processing strategy depends on tool output characteristics and reasoning complexity, highlighting the need for better JSON processing capabilities in LLMs for effective task automation.

Abstract: Most realistic task automation problems require large language models (LLMs) to call tools, which often return complex JSON responses. These responses must be further processed to derive the information necessary for task completion. The ability of LLMs to do so is under-studied. In this paper, we study the tool response processing task and LLMs’ abilities to process structured (JSON) responses. We created a dataset for this task, and evaluated 15 open and closed weight models using multiple prompting approaches. Our results show that JSON processing remains a difficult task even for frontier models across multiple prompting strategies. The optimal response processing strategy depends on both the nature and size of the tool outputs, as well as the complexity of the required reasoning. Variations in processing approaches can lead to performance differences ranging from 3% to 50%.

[603] Hydrogen production from blended waste biomass: pyrolysis, thermodynamic-kinetic analysis and AI-based modelling

Sana Kordoghli, Abdelhakim Settar, Oumayma Belaati, Mohammad Alkhatib

Main category: cs.LG

TL;DR: This study investigates thermochemical conversion of food biomass (spent coffee grounds and date seeds) through pyrolysis for sustainable hydrogen production, using AI to enhance process modeling and optimization.

DetailsMotivation: To advance sustainable energy and waste management by exploring underutilized biomass resources for hydrogen production and improving pyrolysis process efficiency through AI integration.

Method: Conducted proximate, ultimate, fiber, TGA/DTG, kinetic, thermodynamic, and Py-Micro GC analyses on pure DS, SCG, and their blends. Used isoconversional methods (KAS, FWO, Friedman) for kinetic modeling and trained an LSTM model with lignocellulosic data.

Result: Blend 3 (25% DS - 75% SCG) offered superior hydrogen yield potential but highest activation energy (313.24 kJ/mol), while Blend 1 (75% DS - 25% SCG) had best activation energy (161.75 kJ/mol). KAS method was most accurate for kinetic modeling. LSTM model predicted TGA curves with exceptional accuracy (R²: 0.9996-0.9998).

Conclusion: The integration of AI, particularly LSTM models, significantly enhances pyrolysis process modeling accuracy, enabling better optimization of biomass blends for sustainable hydrogen production from food waste resources.

Abstract: This work contributes to advancing sustainable energy and waste management strategies by investigating the thermochemical conversion of food-based biomass through pyrolysis, highlighting the role of artificial intelligence (AI) in enhancing process modelling accuracy and optimization efficiency. The main objective is to explore the potential of underutilized biomass resources, such as spent coffee grounds (SCG) and date seeds (DS), for sustainable hydrogen production. Specifically, it aims to optimize the pyrolysis process while evaluating the performance of these resources both individually and as blends. Proximate, ultimate, fibre, TGA/DTG, kinetic, thermodynamic, and Py-Micro GC analyses were conducted for pure DS, SCG, and blends (75% DS - 25% SCG, 50% DS

  • 50% SCG, 25% DS - 75% SCG). Blend 3 offered superior hydrogen yield potential but had the highest activation energy (Ea: 313.24 kJ/mol), while Blend 1 exhibited the best activation energy value (Ea: 161.75 kJ/mol). The kinetic modelling based on isoconversional methods (KAS, FWO, Friedman) identified KAS as the most accurate. These approaches provide a detailed understanding of the pyrolysis process, with particular emphasis on the integration of artificial intelligence. An LSTM model trained with lignocellulosic data predicted TGA curves with exceptional accuracy (R^2: 0.9996-0.9998).

[604] Interpretable Graph-Language Modeling for Detecting Youth Illicit Drug Use

Yiyang Li, Zehong Wang, Zhengqing Yuan, Zheyuan Zhang, Keerthiram Murugesan, Chuxu Zhang, Yanfang Ye

Main category: cs.LG

TL;DR: LAMI is a joint graph-language modeling framework that detects illicit drug use among teenagers and young adults by learning latent connections in survey data and generating natural language explanations.

DetailsMotivation: Existing methods treat survey variables independently, overlooking latent and interconnected structures among demographic, psychological, and environmental factors related to substance use.

Method: LAMI represents individual survey responses as relational graphs, learns latent connections through a specialized graph structure learning layer, and integrates a large language model to generate natural language explanations.

Result: Experiments on YRBS and NSDUH datasets show LAMI outperforms competitive baselines in predictive accuracy and reveals meaningful behavioral substructures and psychosocial pathways.

Conclusion: LAMI effectively detects illicit drug use while providing interpretable insights into established risk factors like family dynamics, peer influence, and school-related distress.

Abstract: Illicit drug use among teenagers and young adults (TYAs) remains a pressing public health concern, with rising prevalence and long-term impacts on health and well-being. To detect illicit drug use among TYAs, researchers analyze large-scale surveys such as the Youth Risk Behavior Survey (YRBS) and the National Survey on Drug Use and Health (NSDUH), which preserve rich demographic, psychological, and environmental factors related to substance use. However, existing modeling methods treat survey variables independently, overlooking latent and interconnected structures among them. To address this limitation, we propose LAMI (LAtent relation Mining with bi-modal Interpretability), a novel joint graph-language modeling framework for detecting illicit drug use and interpreting behavioral risk factors among TYAs. LAMI represents individual responses as relational graphs, learns latent connections through a specialized graph structure learning layer, and integrates a large language model to generate natural language explanations grounded in both graph structures and survey semantics. Experiments on the YRBS and NSDUH datasets show that LAMI outperforms competitive baselines in predictive accuracy. Interpretability analyses further demonstrate that LAMI reveals meaningful behavioral substructures and psychosocial pathways, such as family dynamics, peer influence, and school-related distress, that align with established risk factors for substance use.

[605] CTR-LoRA: Curvature-Aware and Trust-Region Guided Low-Rank Adaptation for Large Language Models

Zhuxuanzi Wang, Mingqiao Mo, Xi Xiao, Chen Liu, Chenrui Ma, Yunbei Zhang, Xiao Wang, Smita Krishnaswamy, Tianyang Wang

Main category: cs.LG

TL;DR: CTR-LoRA is a parameter-efficient fine-tuning framework that integrates rank scheduling with stability-aware optimization using curvature trust regions, achieving better performance and efficiency than existing PEFT methods.

DetailsMotivation: Previous PEFT methods often decouple capacity allocation from update evolution during training, leading to suboptimal efficiency and performance.

Method: CTR-LoRA allocates parameters based on marginal utility from lightweight second-order proxies and constrains updates using a Fisher/Hessian-metric trust region.

Result: Experiments on 7B-13B models show consistent improvements over strong PEFT baselines on both in-distribution and out-of-distribution benchmarks, with enhanced training stability, reduced memory, and higher throughput.

Conclusion: CTR-LoRA provides a principled path toward more robust and deployable PEFT methods, positioning it on the Pareto frontier of performance and efficiency.

Abstract: Parameter-efficient fine-tuning (PEFT) has become the standard approach for adapting large language models under limited compute and memory budgets. Although previous methods improve efficiency through low-rank updates, quantization, or heuristic budget reallocation, they often decouple the allocation of capacity from the way updates evolve during training. In this work, we introduce CTR-LoRA, a framework guided by curvature trust region that integrates rank scheduling with stability-aware optimization. CTR-LoRA allocates parameters based on marginal utility derived from lightweight second-order proxies and constrains updates using a Fisher/Hessian-metric trust region. Experiments on multiple open-source backbones (7B-13B), evaluated on both in-distribution and out-of-distribution benchmarks, show consistent improvements over strong PEFT baselines. In addition to increased accuracy, CTR-LoRA enhances training stability, reduces memory requirements, and achieves higher throughput, positioning it on the Pareto frontier of performance and efficiency. These results highlight a principled path toward more robust and deployable PEFT.

[606] Long Exposure: Accelerating Parameter-Efficient Fine-Tuning for LLMs under Shadowy Sparsity

Tuowei Wang, Kun Li, Zixu Hao, Donglin Bai, Ju Ren, Yaoxue Zhang, Ting Cao, Mao Yang

Main category: cs.LG

TL;DR: Long Exposure is an efficient system that accelerates parameter-efficient fine-tuning (PEFT) for LLMs by addressing Shadowy Sparsity through three components: Shadowy-sparsity Exposer, Sequence-oriented Predictor, and Dynamic-aware Operator.

DetailsMotivation: The inefficiency of PEFT techniques presents significant challenges in terms of time investments and operational costs for adapting pre-trained LLMs to downstream tasks.

Method: The system addresses Shadowy Sparsity with three key components: 1) Shadowy-sparsity Exposer to capture sparsity details, 2) Sequence-oriented Predictor for efficient predictions on large sequences, and 3) Dynamic-aware Operator for structured computational patterns and coalesced memory accesses.

Result: Long Exposure achieves up to 2.49x speedup in end-to-end fine-tuning compared to state-of-the-art methods.

Conclusion: The system offers promising advancements in accelerating PEFT for LLMs by effectively addressing the unique challenges of Shadowy Sparsity in fine-tuning.

Abstract: The adaptation of pre-trained large language models (LLMs) to diverse downstream tasks via fine-tuning is critical for numerous applications. However, the inefficiency of parameter-efficient fine-tuning (PEFT) techniques presents significant challenges in terms of time investments and operational costs. In this paper, we first introduce a nuanced form of sparsity, termed Shadowy Sparsity, which is distinctive in fine-tuning and has not been adequately addressed for acceleration. Under Shadowy Sparsity, we propose Long Exposure, an efficient system to accelerate PEFT for LLMs. Long Exposure comprises three key components: Shadowy-sparsity Exposer employs a prolonged sensing range to capture more sparsity details under shadowy sparsity; Sequence-oriented Predictor provides efficient yet accurate predictions to handle large sequence inputs and constantly-evolving parameters; and Dynamic-aware Operator facilitates more structured computational patterns and coalesced memory accesses, addressing dynamic sparse operations. Extensive evaluations show that Long Exposure outperforms state-of-the-arts with up to a $2.49\times$ speedup in end-to-end fine-tuning, offering promising advancements in accelerating PEFT for LLMs.

[607] One Token Embedding Is Enough to Deadlock Your Large Reasoning Model

Mohan Zhang, Yihua Zhang, Jinghan Jia, Zhangyang Wang, Sijia Liu, Tianlong Chen

Main category: cs.LG

TL;DR: The Deadlock Attack is a resource exhaustion method that hijacks large reasoning models’ generative control flow using adversarial embeddings to induce perpetual reasoning loops, achieving 100% success rate across multiple models.

DetailsMotivation: To expose a critical security vulnerability in large reasoning models where chain-of-thought reasoning introduces a new attack surface for resource exhaustion through perpetual thinking loops.

Method: Training malicious adversarial embeddings that encourage transitional tokens after reasoning steps, combined with a backdoor implantation strategy to overcome the continuous-to-discrete projection gap and ensure reliable activation through trigger tokens.

Result: Achieved 100% attack success rate across four advanced LRMs (Phi-RM, Nemotron-Nano, R1-Qwen, R1-Llama) and three math reasoning benchmarks, forcing models to generate up to their maximum token limits while remaining stealthy and robust against mitigation strategies.

Conclusion: The findings reveal a critical and underexplored security vulnerability in large reasoning models from the perspective of reasoning inefficiency, exposing the risks of chain-of-thought reasoning mechanisms.

Abstract: Modern large reasoning models (LRMs) exhibit impressive multi-step problem-solving via chain-of-thought (CoT) reasoning. However, this iterative thinking mechanism introduces a new vulnerability surface. We present the Deadlock Attack, a resource exhaustion method that hijacks an LRM’s generative control flow by training a malicious adversarial embedding to induce perpetual reasoning loops. Specifically, the optimized embedding encourages transitional tokens (e.g., “Wait”, “But”) after reasoning steps, preventing the model from concluding its answer. A key challenge we identify is the continuous-to-discrete projection gap: na"ive projections of adversarial embeddings to token sequences nullify the attack. To overcome this, we introduce a backdoor implantation strategy, enabling reliable activation through specific trigger tokens. Our method achieves a 100% attack success rate across four advanced LRMs (Phi-RM, Nemotron-Nano, R1-Qwen, R1-Llama) and three math reasoning benchmarks, forcing models to generate up to their maximum token limits. The attack is also stealthy (in terms of causing negligible utility loss on benign user inputs) and remains robust against existing strategies trying to mitigate the overthinking issue. Our findings expose a critical and underexplored security vulnerability in LRMs from the perspective of reasoning (in)efficiency.

[608] Gains: Fine-grained Federated Domain Adaptation in Open Set

Zhengyi Zhong, Wenzheng Jiang, Weidong Bao, Ji Wang, Cheems Wang, Guanbo Wang, Yongheng Deng, Ju Ren

Main category: cs.LG

TL;DR: Gains is a fine-grained federated domain adaptation approach for open-set scenarios where new clients continuously join, enabling knowledge discovery and adaptation while preserving source domain performance.

DetailsMotivation: Real-world FL scenarios involve new clients joining continuously with new knowledge, but existing methods have coarse-grained discovery and sacrifice source domain performance and adaptation efficiency.

Method: Splits model into encoder and classifier, uses fine-grained knowledge discovery and contribution-driven aggregation to identify new knowledge, and includes anti-forgetting mechanism to preserve source domain performance.

Result: Significantly outperforms other baselines on multi-domain datasets across three data-shift scenarios for both source-domain and target-domain clients.

Conclusion: Gains effectively addresses open-set FL by providing fine-grained knowledge discovery and balanced adaptation while maintaining source domain performance.

Abstract: Conventional federated learning (FL) assumes a closed world with a fixed total number of clients. In contrast, new clients continuously join the FL process in real-world scenarios, introducing new knowledge. This raises two critical demands: detecting new knowledge, i.e., knowledge discovery, and integrating it into the global model, i.e., knowledge adaptation. Existing research focuses on coarse-grained knowledge discovery, and often sacrifices source domain performance and adaptation efficiency. To this end, we propose a fine-grained federated domain adaptation approach in open set (Gains). Gains splits the model into an encoder and a classifier, empirically revealing features extracted by the encoder are sensitive to domain shifts while classifier parameters are sensitive to class increments. Based on this, we develop fine-grained knowledge discovery and contribution-driven aggregation techniques to identify and incorporate new knowledge. Additionally, an anti-forgetting mechanism is designed to preserve source domain performance, ensuring balanced adaptation. Experimental results on multi-domain datasets across three typical data-shift scenarios demonstrate that Gains significantly outperforms other baselines in performance for both source-domain and target-domain clients. Code is available at: https://github.com/Zhong-Zhengyi/Gains.

[609] Self-Attention to Operator Learning-based 3D-IC Thermal Simulation

Zhen Huang, Hong Wang, Wenkai Yang, Muxi Tang, Depeng Xie, Ting-Jung Lin, Yu Zhang, Wei W. Xing, Lei He

Main category: cs.LG

TL;DR: SAU-FNO combines self-attention, U-Net, and Fourier Neural Operator to achieve fast and accurate thermal prediction in 3D ICs with 842x speedup over FEM methods.

DetailsMotivation: Thermal management in 3D ICs is challenging due to high power densities. Traditional PDE methods are accurate but slow, while existing ML approaches suffer from high-frequency information loss and require extensive high-fidelity data.

Method: Proposed SAU-FNO framework that integrates self-attention and U-Net with FNO to capture long-range dependencies and local high-frequency features. Uses transfer learning to fine-tune low-fidelity data, reducing dependency on high-fidelity datasets.

Result: SAU-FNO achieves state-of-the-art thermal prediction accuracy and provides 842x speedup compared to traditional FEM methods.

Conclusion: SAU-FNO is an efficient tool for advanced 3D IC thermal simulations, offering both accuracy and significant computational speed improvements.

Abstract: Thermal management in 3D ICs is increasingly challenging due to higher power densities. Traditional PDE-solving-based methods, while accurate, are too slow for iterative design. Machine learning approaches like FNO provide faster alternatives but suffer from high-frequency information loss and high-fidelity data dependency. We introduce Self-Attention U-Net Fourier Neural Operator (SAU-FNO), a novel framework combining self-attention and U-Net with FNO to capture long-range dependencies and model local high-frequency features effectively. Transfer learning is employed to fine-tune low-fidelity data, minimizing the need for extensive high-fidelity datasets and speeding up training. Experiments demonstrate that SAU-FNO achieves state-of-the-art thermal prediction accuracy and provides an 842x speedup over traditional FEM methods, making it an efficient tool for advanced 3D IC thermal simulations.

[610] LinearizeLLM: An Agent-Based Framework for LLM-Driven Exact Linear Reformulation of Nonlinear Optimization Problems

Paul-Niklas Ken Kandora, Simon Caspar Zeller, Aaron Jeremias Elsing, Elena Kuss, Steffen Rebennack

Main category: cs.LG

TL;DR: LinearizeLLM is an agent-based framework that uses specialized LLM agents to automatically reformulate nonlinear optimization problems into linear equivalents, enabling solver compatibility.

DetailsMotivation: Manual reformulation of nonlinear optimization problems is expertise-intensive but essential for using linear solvers or special-purpose algorithms. This process needs automation.

Method: An agent-based framework where each nonlinear pattern (e.g., absolute-value terms, bilinear products) is assigned to a specialized reformulation agent that derives exact linear reformulations, then coordinates to assemble a solver-ready linear model.

Result: Evaluated on 20 real-world nonlinear problems from ComplexOR dataset. Results show specialized LLM agents can successfully automate linearization tasks.

Conclusion: The framework enables automated linearization of nonlinear optimization problems, paving the way for fully conversational modeling pipelines in optimization.

Abstract: Reformulating nonlinear optimization problems is largely manual and expertise-intensive, yet it remains essential for solving such problems with linear optimization solvers or applying special-purpose algorithms. We introduce \textit{LinearizeLLM}, an agent-based framework that solves this task by leveraging Large Language Models (LLMs). The framework assigns each nonlinear pattern to a \textit{reformulation agent} that is explicitly instructed to derive an exact linear reformulation for its nonlinearity pattern, for instance, absolute-value terms or bilinear products of decision variables. The agents then coordinate to assemble a solver-ready linear model equivalent to the original problem. To benchmark the approach, we create a dataset of 20 real-world nonlinear optimization problems derived from the established ComplexOR dataset of linear optimization problems. We evaluate our approach with several LLMs. Our results indicate that specialized LLM agents can automate linearization tasks, opening a path toward fully conversational modeling pipelines for nonlinear optimization.

[611] Speech Foundation Models Generalize to Time Series Tasks from Wearable Sensor Data

Jaya Narain, Zakaria Aldeneh, Shirley Ren

Main category: cs.LG

TL;DR: Speech foundation models like HuBERT and wav2vec 2.0 can generalize beyond speech to achieve state-of-the-art performance on wearable sensor time-series tasks through simple probing methods.

DetailsMotivation: To develop generalized time-series models that unify speech and sensor modalities, leveraging the fact that both encode information in time- and frequency-domains.

Method: Extract features from speech foundation models (HuBERT and wav2vec 2.0) and train probes on these features for wearable sensor tasks, focusing on their convolutional feature encoders.

Result: Speech model features outperform self-supervised models trained directly on modality-specific datasets for mood classification, arrhythmia detection, and activity classification tasks.

Conclusion: Speech foundation models provide effective representations for data-scarce time-series tasks, enabling performance enhancement through simple probing methods and advancing generalized time-series modeling.

Abstract: Both speech and sensor time series data encode information in both the time- and frequency- domains, like spectral powers and waveform shapelets. We show that speech foundation models learn representations that generalize beyond the speech domain and achieve state-of-the-art performance on diverse time-series tasks from wearable sensors. Probes trained on features extracted from HuBERT and wav2vec 2.0 outperform those extracted from self-supervised models trained directly on modality-specific datasets for mood classification, arrhythmia detection, and activity classification tasks. We find that the convolutional feature encoders of speech models are particularly relevant for wearable sensor applications. The proposed approach enhances performance on data-scarce time-series tasks using simple probing methods. This work takes a step toward developing generalized time-series models that unify speech and sensor modalities.

[612] Predict Training Data Quality via Its Geometry in Metric Space

Yang Ba, Mohammad Sadeq Abolhasani, Rong Pan

Main category: cs.LG

TL;DR: The paper proposes using persistent homology to analyze the geometric structure of training data, showing that both representation richness and redundancy elimination critically impact machine learning model performance.

DetailsMotivation: While data quality is known to be crucial for ML/AI, the impact of data's geometric structure on model performance remains underexplored, particularly how representation richness and redundancy elimination affect learning outcomes.

Method: Employ persistent homology to extract topological features from data within a metric space, providing a principled way to quantify diversity beyond traditional entropy-based measures.

Result: Persistent homology proves to be a powerful tool for analyzing and enhancing training data quality by capturing geometric and topological properties that influence learning outcomes.

Conclusion: Persistent homology offers a novel approach to understand and improve training data quality by focusing on geometric structure, representation richness, and redundancy elimination, which are critical factors for AI system performance.

Abstract: High-quality training data is the foundation of machine learning and artificial intelligence, shaping how models learn and perform. Although much is known about what types of data are effective for training, the impact of the data’s geometric structure on model performance remains largely underexplored. We propose that both the richness of representation and the elimination of redundancy within training data critically influence learning outcomes. To investigate this, we employ persistent homology to extract topological features from data within a metric space, thereby offering a principled way to quantify diversity beyond entropy-based measures. Our findings highlight persistent homology as a powerful tool for analyzing and enhancing the training data that drives AI systems.

[613] Bolster Hallucination Detection via Prompt-Guided Data Augmentation

Wenyun Li, Zheng Zhang, Dongmei Jiang, Xiangyuan Lan

Main category: cs.LG

TL;DR: PALE is a novel framework for hallucination detection in LLMs that uses prompt-guided data augmentation and a Contrastive Mahalanobis Score metric, achieving 6.55% improvement over baselines without requiring human annotations.

DetailsMotivation: Address the scarcity of well-labeled datasets for hallucination detection in LLMs and the need for reliable methods to detect misleading or fabricated information in LLM-generated content.

Method: Uses prompt-guided responses from LLMs as data augmentation to generate both truthful and hallucinated data. Introduces Contrastive Mahalanobis Score (CM Score) based on modeling distributions of truthful and hallucinated data in activation space using matrix decomposition.

Result: Achieves superior hallucination detection performance, outperforming competitive baseline by 6.55% margin in extensive experiments.

Conclusion: PALE framework offers strong generalizability and practicality for real-world applications without requiring additional human annotations, effectively addressing hallucination detection challenges in LLMs.

Abstract: Large language models (LLMs) have garnered significant interest in AI community. Despite their impressive generation capabilities, they have been found to produce misleading or fabricated information, a phenomenon known as hallucinations. Consequently, hallucination detection has become critical to ensure the reliability of LLM-generated content. One primary challenge in hallucination detection is the scarcity of well-labeled datasets containing both truthful and hallucinated outputs. To address this issue, we introduce Prompt-guided data Augmented haLlucination dEtection (PALE), a novel framework that leverages prompt-guided responses from LLMs as data augmentation for hallucination detection. This strategy can generate both truthful and hallucinated data under prompt guidance at a relatively low cost. To more effectively evaluate the truthfulness of the sparse intermediate embeddings produced by LLMs, we introduce an estimation metric called the Contrastive Mahalanobis Score (CM Score). This score is based on modeling the distributions of truthful and hallucinated data in the activation space. CM Score employs a matrix decomposition approach to more accurately capture the underlying structure of these distributions. Importantly, our framework does not require additional human annotations, offering strong generalizability and practicality for real-world applications. Extensive experiments demonstrate that PALE achieves superior hallucination detection performance, outperforming the competitive baseline by a significant margin of 6.55%.

[614] DAWP: A framework for global observation forecasting via Data Assimilation and Weather Prediction in satellite observation space

Junchao Gong, Jingyi Xu, Ben Fei, Fenghua Ling, Wenlong Zhang, Kun Chen, Wanghan Xu, Weidong Yang, Xiaokang Yang, Lei Bai

Main category: cs.LG

TL;DR: DAWP is an AI weather prediction framework that operates directly in observation space using an AI data assimilation module, eliminating reliance on reanalysis data and enabling global observation forecasting.

DetailsMotivation: Current AI weather prediction methods rely on reanalysis data which has limitations including data assimilation biases and temporal discrepancies. Observation forecasting offers a transformative alternative but faces challenges in learning spatiotemporal dynamics from irregular high-resolution observation data.

Method: Proposes DAWP framework with two key components: 1) AI data assimilation (AIDA) module using mask multi-modality autoencoder (MMAE) with mask ViT-VAEs to assimilate irregular satellite observations, 2) Spatiotemporal decoupling transformer with cross-regional boundary conditioning for learning dynamics in observation space and enabling sub-image-based global forecasting.

Result: Comprehensive experiments show AIDA initialization significantly improves rollout and efficiency of AI weather prediction. The framework demonstrates promising potential for global precipitation forecasting applications.

Conclusion: DAWP successfully liberates AI weather prediction from reanalysis data dependencies by operating directly in observation space, providing improved performance and efficiency for weather forecasting tasks.

Abstract: Weather prediction is a critical task for human society, where impressive progress has been made by training artificial intelligence weather prediction (AIWP) methods with reanalysis data. However, reliance on reanalysis data limits the AIWPs with shortcomings, including data assimilation biases and temporal discrepancies. To liberate AIWPs from the reanalysis data, observation forecasting emerges as a transformative paradigm for weather prediction. One of the key challenges in observation forecasting is learning spatiotemporal dynamics across disparate measurement systems with irregular high-resolution observation data, which constrains the design and prediction of AIWPs. To this end, we propose our DAWP as an innovative framework to enable AIWPs to operate in a complete observation space by initialization with an artificial intelligence data assimilation (AIDA) module. Specifically, our AIDA module applies a mask multi-modality autoencoder(MMAE)for assimilating irregular satellite observation tokens encoded by mask ViT-VAEs. For AIWP, we introduce a spatiotemporal decoupling transformer with cross-regional boundary conditioning (CBC), learning the dynamics in observation space, to enable sub-image-based global observation forecasting. Comprehensive experiments demonstrate that AIDA initialization significantly improves the roll out and efficiency of AIWP. Additionally, we show that DAWP holds promising potential to be applied in global precipitation forecasting.

[615] Cog-Rethinker: Hierarchical Metacognitive Reinforcement Learning for LLM Reasoning

Zexu Sun, Yongcheng Zeng, Erxue Min, Heyang Gao, Bokai Ji, Xu Chen

Main category: cs.LG

TL;DR: Cog-Rethinker is a hierarchical metacognitive RL framework that improves sample efficiency in LLM reasoning by decomposing zero-accuracy problems and refining wrong solutions, outperforming baseline methods on mathematical reasoning benchmarks.

DetailsMotivation: Previous RL approaches for LLM reasoning rely on fixed prompt templates, causing substantial sampling inefficiencies as most problems generate invalid outputs during accuracy-driven filtration, wasting samples.

Method: Proposes Cog-Rethinker with a two-stage hierarchical metacognitive framework: 1) decomposes zero-accuracy problems into subproblems, 2) refines wrong answers by referencing previous solutions. Uses supervised fine-tuning for cold-start and maintains train-test consistency.

Result: Superior performance on various mathematical reasoning benchmarks with improved sample efficiency that accelerates convergence compared to baseline methods.

Conclusion: Cog-Rethinker effectively addresses sampling inefficiency in RL-based LLM reasoning through hierarchical metacognitive processing, demonstrating both performance improvements and faster convergence.

Abstract: Contemporary progress in large language models (LLMs) has revealed notable inferential capacities via reinforcement learning (RL) employing verifiable reward, facilitating the development of O1 and R1-like reasoning models. Directly training from base models with RL is called zero-RL. However, previous works rely upon activating LLMs’ inherent capacities through fixed prompt templates. This strategy introduces substantial sampling inefficiencies for weak LLMs, as the majority of problems generate invalid outputs during accuracy-driven filtration in reasoning tasks, which causes a waste of samples. To solve this issue, we propose Cog-Rethinker, a novel hierarchical metacognitive RL framework for LLM reasoning. Our Cog-Rethinker mainly focuses on the rollout procedure in RL training. After the direct rollout, our Cog-Rethinker improves sample utilization in a hierarchical metacognitive two-stage framework. By leveraging human cognition during solving problems, firstly, it prompts policy to decompose zero-accuracy problems into subproblems to produce final reasoning results. Secondly, with zero-accuracy problems in previous rollout stage, it further prompts policy to refine these answers by referencing previous wrong solutions. Moreover, to enable cold-start of the two new reasoning patterns and maintain train-test consistency across prompt templates, our Cog-Rethinker applies supervised fine-tuning on the policy using correct samples of the two stages with direct rollout template. Experimental results demonstrate Cog-Rethinker’s superior performance on various mathematical reasoning benchmarks, we also analyzed its improved sample efficiency that accelerates convergence compared to baseline methods.

[616] AMiD: Knowledge Distillation for LLMs with $α$-mixture Assistant Distribution

Donghyeok Shin, Yeongmin Kim, Suhyeon Jo, Byeonghu Na, Il-Chul Moon

Main category: cs.LG

TL;DR: AMiD proposes a generalized framework for knowledge distillation using α-mixture assistant distributions to address training instability and capacity gaps in LLM distillation.

DetailsMotivation: Autoregressive LLMs have high computational costs, and existing knowledge distillation methods suffer from capacity gaps and training instability due to near-zero probabilities in high-dimensional outputs.

Method: Introduces α-mixture assistant distribution with variable α parameter, and AMiD framework that generalizes divergences used with assistant distributions based on optimality.

Result: AMiD demonstrates superior performance and training stability by leveraging a broader, theoretically grounded assistant distribution space.

Conclusion: The proposed α-mixture assistant distribution and AMiD framework provide a unified, systematic approach that outperforms previous fragmented methods for LLM knowledge distillation.

Abstract: Autoregressive large language models (LLMs) have achieved remarkable improvement across many tasks but incur high computational and memory costs. Knowledge distillation (KD) mitigates this issue by transferring knowledge from a large teacher to a smaller student through distributional alignment. Previous studies have proposed various discrepancy metrics, but the capacity gap and training instability caused by near-zero probabilities, stemming from the high-dimensional output of LLMs, remain fundamental limitations. To overcome these challenges, several approaches implicitly or explicitly incorporating assistant distribution have recently been proposed. However, the past proposals of assistant distributions have been a fragmented approach without a systematic investigation of the interpolation path and the divergence. This paper proposes $\alpha$-mixture assistant distribution, a novel generalized family of assistant distributions, and $\alpha$-mixture distillation, coined AMiD, a unified framework for KD using the assistant distribution. The $\alpha$-mixture assistant distribution provides a continuous extension of the assistant distribution by introducing a new distribution design variable $\alpha$, which has been fixed in all previous approaches. Furthermore, AMiD generalizes the family of divergences used with the assistant distributions based on optimality, which has also been restricted in previous works. Through extensive experiments, we demonstrate that AMiD offers superior performance and training stability by leveraging a broader and theoretically grounded assistant distribution space.

[617] MEET-Sepsis: Multi-Endogenous-View Enhanced Time-Series Representation Learning for Early Sepsis Prediction Representation Learning for Early Sepsis Prediction

Zexi Tan, Tao Xie, Binbin Sun, Xiang Zhang, Yiqun Zhang, Yiu-Ming Cheung

Main category: cs.LG

TL;DR: MEET-Sepsis framework uses multi-view feature enhancement and multi-scale temporal attention for early sepsis prediction, achieving competitive accuracy with only 20% of ICU monitoring time compared to SOTA methods.

DetailsMotivation: Early sepsis prediction is critical but challenging due to subtle early manifestations and rapidly escalating mortality. Existing AI methods struggle to capture weak early temporal signals.

Method: Proposes MERE mechanism for enriched feature views and CDTA module for multi-scale temporal representation learning in the MEET-Sepsis framework.

Result: Achieves competitive prediction accuracy using only 20% of the ICU monitoring time required by state-of-the-art methods.

Conclusion: MEET-Sepsis significantly advances early sepsis prediction, with extensive validation confirming its efficacy.

Abstract: Sepsis is a life-threatening infectious syndrome associated with high mortality in intensive care units (ICUs). Early and accurate sepsis prediction (SP) is critical for timely intervention, yet remains challenging due to subtle early manifestations and rapidly escalating mortality. While AI has improved SP efficiency, existing methods struggle to capture weak early temporal signals. This paper introduces a Multi-Endogenous-view Representation Enhancement (MERE) mechanism to construct enriched feature views, coupled with a Cascaded Dual-convolution Time-series Attention (CDTA) module for multi-scale temporal representation learning. The proposed MEET-Sepsis framework achieves competitive prediction accuracy using only 20% of the ICU monitoring time required by SOTA methods, significantly advancing early SP. Extensive validation confirms its efficacy. Code is available at: https://github.com/yueliangy/MEET-Sepsis.

[618] User Profiles of Sleep Disorder Sufferers: Towards Explainable Clustering and Differential Variable Analysis

Sifeddine Sellami, Juba Agoun, Lamia Yessad, Louenas Bounia

Main category: cs.LG

TL;DR: Proposes a clustering-based method using explainable AI to group patients by sleep disorder profiles and identify key influencing factors.

DetailsMotivation: Sleep disorders significantly impact health and quality of life, but diagnosis is complex due to symptom diversity. Technological advances and medical data analysis offer new opportunities to better understand these disorders.

Method: Clustering-based approach integrated with explainable artificial intelligence (XAI) to group patients according to different sleep disorder profiles and identify key influencing factors.

Result: Experiment on anonymized real data demonstrates the effectiveness and relevance of the proposed approach.

Conclusion: The XAI-based clustering method successfully identifies sleep disorder profiles and key factors, providing an effective tool for understanding and diagnosing sleep disorders.

Abstract: Sleep disorders have a major impact on patients’ health and quality of life, but their diagnosis remains complex due to the diversity of symptoms. Today, technological advances, combined with medical data analysis, are opening new perspectives for a better understanding of these disorders. In particular, explainable artificial intelligence (XAI) aims to make AI model decisions understandable and interpretable for users. In this study, we propose a clustering-based method to group patients according to different sleep disorder profiles. By integrating an explainable approach, we identify the key factors influencing these pathologies. An experiment on anonymized real data illustrates the effectiveness and relevance of our approach.

[619] Algorithmic Primitives and Compositional Geometry of Reasoning in Language Models

Samuel Lippl, Thomas McGee, Kimberly Lopez, Ziwen Pan, Pierce Zhang, Salma Ziadi, Oliver Eberle, Ida Momennejad

Main category: cs.LG

TL;DR: The paper introduces a framework for tracing and steering algorithmic primitives in LLMs’ reasoning processes, showing that reasoning is supported by compositional geometric primitives that can be transferred across tasks and models.

DetailsMotivation: To understand how latent and inference time computations enable LLMs to solve multi-step reasoning tasks by identifying and manipulating the underlying algorithmic primitives.

Method: Links reasoning traces to internal activation patterns, operationalizes primitives through clustering neural activations, and applies function vector methods to derive reusable primitive vectors that can be combined through arithmetic operations.

Result: Identified shared and task-specific primitives across benchmarks (TSP, 3SAT, AIME, graph navigation) and models (Phi-4, Phi-4-Reasoning, Llama-3-8B). Reasoning finetuning enhances compositional generalization and systematic use of verification/path-generation primitives.

Conclusion: LLM reasoning is supported by a compositional geometry of algorithmic primitives that transfer cross-task and cross-model, and reasoning finetuning strengthens algorithmic generalization across domains.

Abstract: How do latent and inference time computations enable large language models (LLMs) to solve multi-step reasoning? We introduce a framework for tracing and steering algorithmic primitives that underlie model reasoning. Our approach links reasoning traces to internal activation patterns and evaluates algorithmic primitives by injecting them into residual streams and measuring their effect on reasoning steps and task performance. We consider four benchmarks: Traveling Salesperson Problem (TSP), 3SAT, AIME, and graph navigation. We operationalize primitives by clustering neural activations and labeling their matched reasoning traces. We then apply function vector methods to derive primitive vectors as reusable compositional building blocks of reasoning. Primitive vectors can be combined through addition, subtraction, and scalar operations, revealing a geometric logic in activation space. Cross-task and cross-model evaluations (Phi-4, Phi-4-Reasoning, Llama-3-8B) show both shared and task-specific primitives. Notably, comparing Phi-4 with its reasoning-finetuned variant highlights compositional generalization after finetuning: Phi-4-Reasoning exhibits more systematic use of verification and path-generation primitives. Injecting the associated primitive vectors in Phi-4-Base induces behavioral hallmarks associated with Phi-4-Reasoning. Together, these findings demonstrate that reasoning in LLMs may be supported by a compositional geometry of algorithmic primitives, that primitives transfer cross-task and cross-model, and that reasoning finetuning strengthens algorithmic generalization across domains.

[620] Can GRPO Help LLMs Transcend Their Pretraining Origin?

Kangqi Ni, Zhen Tan, Zijie Liu, Pingzhi Li, Tianlong Chen

Main category: cs.LG

TL;DR: GRPO/RLVR improves reasoning in LLMs but only when tasks align with pretraining biases, failing to discover truly novel solutions due to its conservative reweighting nature.

DetailsMotivation: To understand why GRPO shows inconsistent reasoning improvements across domains and identify conditions for out-of-distribution generalization.

Method: Theoretical analysis proving GRPO is a conservative reweighting scheme, plus controlled experiments training transformers from scratch to evaluate generalization across reasoning depth, length, tokens, and compositionality.

Result: OOD improvement occurs only when target tasks align with pretrained biases; ID gains diminish as performance saturates. GRPO cannot discover completely novel solutions.

Conclusion: GRPO is not a universal reasoning enhancer but rather sharpens pretraining biases, motivating development of algorithms that can expand beyond pretraining origins.

Abstract: Reinforcement Learning with Verifiable Rewards (RLVR), primarily driven by the Group Relative Policy Optimization (GRPO) algorithm, is a leading approach for enhancing the reasoning abilities of Large Language Models (LLMs). Despite its wide adoption, GRPO’s gains are often inconsistent; for instance, a model may show significant improvement in one reasoning domain, like mathematics, yet remain stagnant in another, such as medicine. This inconsistency raises a critical question: under what conditions does GRPO improve reasoning and generalize out-of-distribution (OOD)? We investigate this from a data distribution perspective. We first prove theoretically that GRPO is a conservative reweighting scheme, bounded by the base model’s distribution and thus unable to discover completely novel solutions. We further validate this in carefully designed controlled studies by training transformers from scratch, evaluating generalization across reasoning depth, input length, token representation, and compositionality. Our results provide a principled explanation for GRPO’s boundaries: OOD improvement emerges only when the target task aligns with the model’s pretrained biases, while gains on in-distribution (ID) tasks diminish as performance saturates. This reframes GRPO not as a universal reasoning enhancer but as a tool that sharpens pretraining biases. Our findings motivate future development of algorithms that can expand a model’s capabilities beyond its pretraining origin.

[621] Stratos: An End-to-End Distillation Pipeline for Customized LLMs under Distributed Cloud Environments

Ziming Dai, Tuo Zhang, Fei Gao, Xingyi Cai, Xiaofei Wang, Cheng Zhang, Wenyu Wang, Chengjie Zang

Main category: cs.LG

TL;DR: Stratos is an automated LLM distillation pipeline that selects optimal servers, matches teacher-student pairs, and adapts distillation strategies to meet user-defined performance and budget constraints in cloud environments.

DetailsMotivation: Growing industrial demand for customized, cost-efficient LLMs for domain-specific tasks, with existing distillation frameworks requiring manual intervention and struggling with complex user requirements.

Method: End-to-end automated pipeline that: 1) Automatically selects Pareto-optimal servers, 2) Dynamically matches teacher-student pairs, 3) Adapts distillation strategies based on task complexity, 4) Optimizes cloud hosting deployment.

Result: Achieved 4x accuracy improvement over GPT-4o baseline on rare Mahjong reasoning task using reverse synthetic data and knowledge injection. Reduced latency and cost without compromising accuracy.

Conclusion: Stratos shows promise for efficient vertical-domain LLM deployment by automating the entire distillation process while meeting complex user constraints.

Abstract: The growing industrial demand for customized and cost-efficient large language models (LLMs) is fueled by the rise of vertical, domain-specific tasks and the need to optimize performance under constraints such as latency and budget. Knowledge distillation, as an efficient model compression and transfer technique, offers a feasible solution. However, existing distillation frameworks often require manual intervention and struggle to meet such complex user-defined distillation requirements. To bridge this gap, we propose Stratos, an end-to-end LLM distillation pipeline that automates server and model selection, knowledge distillation, and deployment in distributed cloud environments. Given user-defined constraints on model performance and system budget, Stratos automatically selects Pareto-optimal servers, dynamically matches teacher-student pairs, and adapts distillation strategies based on task complexity to optimize cloud hosting. Experiments show that Stratos produces a student model that achieves four times the accuracy of its GPT-4o teacher baseline on a rare, domain-specific Mahjong reasoning task with reverse synthetic data and knowledge injection. Moreover, it achieves reduced latency and cost without compromising accuracy. These results highlight its promise for vertical-domain LLM deployment.

[622] Using Kolmogorov-Smirnov Distance for Measuring Distribution Shift in Machine Learning

Ozan K. Tonguz, Federico Taschin

Main category: cs.LG

TL;DR: The paper proposes using Kolmogorov-Smirnov (KS) Test to measure distribution shift between training and test data in AI systems, showing that even small KS distances (0.02) can cause significant performance degradation (50% travel time increase) in reinforcement learning agents for transportation.

DetailsMotivation: Distribution shift between training and test data is a critical problem in ML/AI that can cause large prediction errors, especially in safety-critical applications like transportation where such errors could compromise system reliability.

Method: The authors propose using Kolmogorov-Smirnov (KS) Test to measure the deviation in probability distribution between training and test data, and use KS distance to quantify the distribution shift and its impact on AI agent performance.

Result: Results show that even a small KS distance of 0.02 can lead to about 50% increase in travel time at a single intersection when using a Reinforcement Learning agent, demonstrating significant performance degradation.

Conclusion: KS Test and KS distance can serve as valuable statistical tools for real-time monitoring of distribution shift in AI systems, particularly in smart transportation applications, helping AI agents better cope with distribution shifts.

Abstract: One of the major problems in Machine Learning (ML) and Artificial Intelligence (AI) is the fact that the probability distribution of the test data in the real world could deviate substantially from the probability distribution of the training data set. When this happens, the predictions of an ML system or an AI agent could involve large errors which is very troublesome and undesirable. While this is a well-known hard problem plaguing the AI and ML systems’ accuracy and reliability, in certain applications such errors could be critical for safety and reliability of AI and ML systems. One approach to deal with this problem is to monitor and measure the deviation in the probability distribution of the test data in real time and to compensate for this deviation. In this paper, we propose and explore the use of Kolmogorov-Smirnov (KS) Test for measuring the distribution shift and we show how the KS distance can be used to quantify the distribution shift and its impact on an AI agent’s performance. Our results suggest that KS distance could be used as a valuable statistical tool for monitoring and measuring the distribution shift. More specifically, it is shown that even a distance of KS=0.02 could lead to about 50% increase in the travel time at a single intersection using a Reinforcement Learning agent which is quite significant. It is hoped that the use of KS Test and KS distance in AI-based smart transportation could be an important step forward for gauging the performance degradation of an AI agent in real time and this, in turn, could help the AI agent to cope with the distribution shift in a more informed manner.

[623] AMStraMGRAM: Adaptive Multi-cutoff Strategy Modification for ANaGRAM

Nilo Schwencke, Cyriaque Rousselot, Alena Shilova, Cyril Furtlehner

Main category: cs.LG

TL;DR: Analysis of PINNs training dynamics with ANaGRAM natural gradient method, proposing multi-cutoff adaptation strategy that achieves machine precision on benchmark PDEs.

DetailsMotivation: Recent works show natural gradient methods outperform standard optimizers for PINNs, but training dynamics analysis and regularization strategies are needed.

Method: Analyze PINNs training with ANaGRAM (natural-gradient approach using SVD with cutoff regularization), propose multi-cutoff adaptation strategy, develop spectral theory framework.

Result: Multi-cutoff strategy enhances ANaGRAM performance, reaching machine precision on benchmark PDE experiments.

Conclusion: Regularization is necessary for PINNs training with natural gradient methods, multi-cutoff adaptation improves performance, spectral theory provides theoretical grounding connecting to Green’s functions.

Abstract: Recent works have shown that natural gradient methods can significantly outperform standard optimizers when training physics-informed neural networks (PINNs). In this paper, we analyze the training dynamics of PINNs optimized with ANaGRAM, a natural-gradient-inspired approach employing singular value decomposition with cutoff regularization. Building on this analysis, we propose a multi-cutoff adaptation strategy that further enhances ANaGRAM’s performance. Experiments on benchmark PDEs validate the effectiveness of our method, which allows to reach machine precision on some experiments. To provide theoretical grounding, we develop a framework based on spectral theory that explains the necessity of regularization and extend previous shown connections with Green’s functions theory.

[624] Layer-Aware Influence for Online Data Valuation Estimation

Ziao Yang, Longbo Huang, Hongfu Liu

Main category: cs.LG

TL;DR: A dynamic data valuation method that tracks training sample influence during optimization using layer-aware online estimation with only loss-to-output gradients, enabling efficient and scalable data curation.

DetailsMotivation: Prior data valuation methods focus on static influence measured after model convergence, ignoring how sample importance changes dynamically during training, especially in deep models. This limits practical data curation.

Method: Developed a layer-aware online estimator that requires only loss-to-output gradients, avoiding parameter-level and full-network gradients while maintaining ranking fidelity of sample influence.

Result: Extensive experiments across LLM pretraining, fine-tuning, and image classification show improved accuracy with significantly lower time and memory costs compared to existing methods.

Conclusion: The proposed method makes dynamic data curation efficient and scalable in practice by providing real-time influence estimation during training without computational overhead.

Abstract: Data-centric learning emphasizes curating high-quality training samples to boost performance rather than designing new architectures. A central problem is to estimate the influence of training sample efficiently. Prior studies largely focus on static influence measured on a converged model, overlooking how data valuation dynamically changes during optimization. This omission neglects the dynamic nature of sample influence during optimization, especially in deep models. To address the computational burden of frequent influence estimation, we develop a layer-aware online estimator that requires only loss-to-output gradients. This design avoids parameter-level and full-network gradients while preserving ranking fidelity. Extensive experiments across LLM pretraining, fine-tuning, and image classification show our method improves accuracy with substantially lower time and memory cost, making dynamic data curation efficient and scalable in practice.

[625] STAR: Boosting Time Series Foundation Models for Anomaly Detection through State-aware Adapter

Hanyin Cheng, Ruitong Zhang, Yuning Lu, Peng Chen, Meng Wang, Yang Shu, Bin Yang, Chenjuan Guo

Main category: cs.LG

TL;DR: STAR is a plug-and-play module that enhances Time Series Foundation Models’ ability to handle state variables in anomaly detection by properly modeling categorical state information through state-aware encoding and conditional adaptation.

DetailsMotivation: Existing Time Series Foundation Models fail to properly handle discrete state variables (like valve on/off, day of week) alongside numerical variables, treating them uniformly which degrades detection performance when state variables are integrated.

Method: STAR consists of three components: Identity-guided State Encoder with learnable State Memory for categorical semantics, Conditional Bottleneck Adapter that generates low-rank parameters based on current state, and Numeral-State Matching module for detecting state variable anomalies.

Result: Extensive experiments on real-world datasets show that STAR improves the performance of existing TSFMs on Multivariate Time Series Anomaly Detection.

Conclusion: STAR effectively addresses the limitation of existing TSFMs in handling state variables and enhances their anomaly detection capabilities through state-aware modeling.

Abstract: While Time Series Foundation Models (TSFMs) have demonstrated remarkable success in Multivariate Time Series Anomaly Detection (MTSAD), however, in real-world industrial scenarios, many time series comprise not only numerical variables such as temperature and flow, but also numerous discrete state variables that describe the system status, such as valve on/off or day of the week. Existing TSFMs often overlook the distinct categorical nature of state variables and their critical role as conditions, typically treating them uniformly with numerical variables. This inappropriate modeling approach prevents the model from fully leveraging state information and even leads to a significant degradation in detection performance after state variables are integrated. To address this critical limitation, this paper proposes a novel STate-aware AdapteR (STAR). STAR is a plug-and-play module designed to enhance the capability of TSFMs in modeling and leveraging state variables during the fine-tuning stage. Specifically, STAR comprisesthree core components: (1) We design an Identity-guided State Encoder, whicheffectively captures the complex categorical semantics of state variables through a learnable State Memory. (2) We propose a Conditional Bottleneck Adapter, which dynamically generates low-rank adaptation parameters conditioned on the current state, thereby flexibly injecting the influence of state variables into the backbone model. (3) We also introduce a Numeral-State Matching module to more effectively detect anomalies inherent to the state variables themselves. Extensive experiments conducted on real-world datasets demonstrate that STAR can improve the performance of existing TSFMs on MTSAD.

[626] Decision-focused Sensing and Forecasting for Adaptive and Rapid Flood Response: An Implicit Learning Approach

Qian Sun, Graham Hults, Susu Xu

Main category: cs.LG

TL;DR: A decision-focused framework for flood emergency response that optimizes sensor placement and flood forecasting models to minimize downstream decision regrets, using differentiable learning over discrete sensor configurations.

DetailsMotivation: Traditional flood management systems use task-agnostic strategies for sensor placement and model training, overlooking that systems with same sensing gain and forecasting errors may lead to different decisions, hindering timely and reliable emergency response.

Method: End-to-end pipeline with four components: contextual scoring network, differentiable sensor selection under budget constraints, spatio-temporal flood reconstruction/forecasting model, and differentiable decision layer. Uses Implicit Maximum Likelihood Estimation (I-MLE) for gradient-based learning over discrete sensor configurations and probabilistic decision heads.

Result: The framework enables strategic sensor placement and optimized flood forecasting models specifically tailored to minimize flood response decision regrets.

Conclusion: The proposed decision-focused approach addresses limitations of traditional methods by directly optimizing for downstream decision quality rather than intermediate metrics like information gain or forecasting errors.

Abstract: Timely and reliable decision-making is vital for flood emergency response, yet it remains severely hindered by limited and imprecise situational awareness due to various budget and data accessibility constraints. Traditional flood management systems often rely on in-situ sensors to calibrate remote sensing-based large-scale flood depth forecasting models, and further take flood depth estimates to optimize flood response decisions. However, these approaches often take fixed, decision task-agnostic strategies to decide where to put in-situ sensors (e.g., maximize overall information gain) and train flood forecasting models (e.g., minimize average forecasting errors), but overlook that systems with the same sensing gain and average forecasting errors may lead to distinct decisions. To address this, we introduce a novel decision-focused framework that strategically selects locations for in-situ sensor placement and optimize spatio-temporal flood forecasting models to optimize downstream flood response decision regrets. Our end-to-end pipeline integrates four components: a contextual scoring network, a differentiable sensor selection module under hard budget constraints, a spatio-temporal flood reconstruction and forecasting model, and a differentiable decision layer tailored to task-specific objectives. Central to our approach is the incorporation of Implicit Maximum Likelihood Estimation (I-MLE) to enable gradient-based learning over discrete sensor configurations, and probabilistic decision heads to enable differentiable approximation to various constrained disaster response tasks.

[627] Transfer learning strategies for accelerating reinforcement-learning-based flow control

Saeed Salehi

Main category: cs.LG

TL;DR: This paper investigates transfer learning for accelerating deep reinforcement learning in chaotic fluid flow control, comparing progressive neural networks (PNNs) against conventional fine-tuning methods.

DetailsMotivation: To accelerate deep reinforcement learning for multifidelity control of chaotic fluid flows by developing effective transfer learning strategies that can reuse knowledge across different fidelity environments.

Method: Employ progressive neural networks (PNNs) for the first time in DRL-based flow control and conduct comprehensive benchmarking of conventional fine-tuning strategies using the Kuramoto-Sivashinsky system as a benchmark.

Result: PNNs enable stable and efficient knowledge transfer with consistent performance gains, while fine-tuning is sensitive to pretraining duration and prone to catastrophic forgetting. PNNs remain effective even with substantial differences between source and target environments.

Conclusion: PNNs demonstrate superior robustness and scalability for transfer learning in flow control, highlighting their potential for computationally efficient control of complex flow configurations.

Abstract: This work investigates transfer learning strategies to accelerate deep reinforcement learning (DRL) for multifidelity control of chaotic fluid flows. Progressive neural networks (PNNs), a modular architecture designed to preserve and reuse knowledge across tasks, are employed for the first time in the context of DRL-based flow control. In addition, a comprehensive benchmarking of conventional fine-tuning strategies is conducted, evaluating their performance, convergence behavior, and ability to retain transferred knowledge. The Kuramoto-Sivashinsky (KS) system is employed as a benchmark to examine how knowledge encoded in control policies, trained in low-fidelity environments, can be effectively transferred to high-fidelity settings. Systematic evaluations show that while fine-tuning can accelerate convergence, it is highly sensitive to pretraining duration and prone to catastrophic forgetting. In contrast, PNNs enable stable and efficient transfer by preserving prior knowledge and providing consistent performance gains, and are notably robust to overfitting during the pretraining phase. Layer-wise sensitivity analysis further reveals how PNNs dynamically reuse intermediate representations from the source policy while progressively adapting deeper layers to the target task. Moreover, PNNs remain effective even when the source and target environments differ substantially, such as in cases with mismatched physical regimes or control objectives, where fine-tuning strategies often result in suboptimal adaptation or complete failure of knowledge transfer. The results highlight the potential of novel transfer learning frameworks for robust, scalable, and computationally efficient flow control that can potentially be applied to more complex flow configurations.

[628] Airfoil optimization using Design-by-Morphing with minimized design-space dimensionality

Sangjoon Lee, Haris Moazam Sheikh

Main category: cs.LG

TL;DR: AirDbM is a Design-by-Morphing approach that uses only 12 optimally selected baseline airfoils to achieve efficient airfoil optimization with reduced dimensionality while maintaining high reconstruction accuracy and superior optimization performance.

DetailsMotivation: To enable effective airfoil geometry optimization by exploring diverse designs with minimal design variables, addressing the need for dimensionality reduction in design space.

Method: Selects optimal set of 12 baseline airfoils from UIUC database (1600+ shapes) by sequentially adding baselines that maximize design capacity, then uses Design-by-Morphing approach for reconstruction and optimization.

Result: Reconstructs 99% of database with mean absolute error below 0.005, matches performance of previous approaches using more baselines. Achieves rapid convergence in multi-objective optimization with greater hypervolume and discovers new Pareto-optimal solutions with enhanced lift-to-drag ratios.

Conclusion: AirDbM demonstrates superior efficiency and adaptability, particularly for reinforcement learning applications, indicating broader potential of Design-by-Morphing in machine learning-driven design.

Abstract: Effective airfoil geometry optimization requires exploring a diverse range of designs using as few design variables as possible. This study introduces AirDbM, a Design-by-Morphing (DbM) approach specialized for airfoil optimization that systematically reduces design-space dimensionality. AirDbM selects an optimal set of 12 baseline airfoils from the UIUC airfoil database, which contains over 1,600 shapes, by sequentially adding the baseline that most increases the design capacity. With these baselines, AirDbM reconstructs 99 % of the database with a mean absolute error below 0.005, which matches the performance of a previous DbM approach that used more baselines. In multi-objective aerodynamic optimization, AirDbM demonstrates rapid convergence and achieves a Pareto front with a greater hypervolume than that of the previous larger-baseline study, where new Pareto-optimal solutions are discovered with enhanced lift-to-drag ratios at moderate stall tolerances. Furthermore, AirDbM demonstrates outstanding adaptability for reinforcement learning (RL) agents in generating airfoil geometry when compared to conventional airfoil parameterization methods, implying the broader potential of DbM in machine learning-driven design.

[629] Feature-driven reinforcement learning for photovoltaic in continuous intraday trading

Arega Getaneh Abate, Xiufeng Liu, Ruyu Liu, Xiaobing Zhang

Main category: cs.LG

TL;DR: A reinforcement learning approach using PPO for PV intraday trading that outperforms benchmarks by balancing trading profit and imbalance penalties through interpretable policies.

DetailsMotivation: PV operators face uncertainty in generation and electricity prices, needing real-time position adjustments in continuous intraday markets to improve revenues and reduce imbalance costs.

Method: Feature-driven RL approach using Markov Decision Process with PPO, integrating data-driven features into state representation with predominantly linear, interpretable policy.

Result: Strategy consistently outperforms benchmark baselines across diverse scenarios, shows rapid convergence, real-time inference, and transparent decision rules with learned weights highlighting market microstructure importance.

Conclusion: Feature-driven RL offers a practical, data-efficient, and operationally deployable pathway for active intraday participation by PV producers.

Abstract: Photovoltaic (PV) operators face substantial uncertainty in generation and short-term electricity prices. Continuous intraday markets enable producers to adjust their positions in real time, potentially improving revenues and reducing imbalance costs. We propose a feature-driven reinforcement learning (RL) approach for PV intraday trading that integrates data-driven features into the state and learns bidding policies in a sequential decision framework. The problem is cast as a Markov Decision Process with a reward that balances trading profit and imbalance penalties and is solved with Proximal Policy Optimization (PPO) using a predominantly linear, interpretable policy. Trained on historical market data and evaluated out-of-sample, the strategy consistently outperforms benchmark baselines across diverse scenarios. Extensive validation shows rapid convergence, real-time inference, and transparent decision rules. Learned weights highlight the central role of market microstructure and historical features. Taken together, these results indicate that feature-driven RL offers a practical, data-efficient, and operationally deployable pathway for active intraday participation by PV producers.

[630] Breaking Memorization Barriers in LLM Code Fine-Tuning via Information Bottleneck for Improved Generalization

Changsheng Wang, Xin Chen, Sijia Liu, Ke Ding

Main category: cs.LG

TL;DR: The paper identifies a memorization barrier in fine-tuning LLMs for code generation and proposes IB-FT, an information bottleneck-guided approach that compresses memorized features while preserving task-relevant information, leading to better performance and stability.

DetailsMotivation: Standard fine-tuning of pretrained LLMs for code generation suffers from a memorization barrier where strong memorization of downstream code data prevents effective acquisition of new, generalizable knowledge.

Method: Proposes IB-FT (Information Bottleneck-guided Fine-Tuning) that applies an IB penalty on hidden representations to compress spurious memorized features while preserving task-relevant information.

Result: IB-FT substantially alleviates the memorization barrier, improves top-1 performance (Pass@1), and yields more stable gains under stricter multi-sample metrics (Pass@k(m)) compared to conventional fine-tuning on two code benchmarks.

Conclusion: The information bottleneck approach effectively overcomes the memorization barrier in code generation fine-tuning, leading to better generalization and more stable performance improvements.

Abstract: Adapting pretrained large language models (LLMs) to code domains via supervised fine-tuning (FT) has been commonly used for code generation. However, we identify a previously underappreciated failure mode, the memorization barrier, where strong memorization of downstream code data in the base model could trap optimization and prevent the standard FT from effectively acquiring new, generalizable code knowledge. To overcome this barrier, we propose the information bottleneck (IB)-guided fine-tuning, termed IB-FT, which applies an IB penalty on hidden representations of the code data to compress spurious, memorized features while preserving task-relevant information. Extensive experiments on two code benchmarks (OriGen and Evol-CodeAlpaca-V1) show that IB-FT substantially alleviates the memorization barrier, improves top-1 performance (Pass@$1$), and yields far more stable gains under the stricter multi-sample metric Pass@$k^{(m)}$ (a problem counts as solved only if at least $m$ of $k$ samples pass unit tests) compared with conventional FT.

[631] Robust Federated Inference

Akash Dhasade, Sadegh Farhadkhani, Rachid Guerraoui, Nirupam Gupta, Maxime Jacovella, Anne-Marie Kermarrec, Rafael Pinot

Main category: cs.LG

TL;DR: This paper provides the first robustness analysis of federated inference methods and introduces a novel adversarial training approach to defend against attacks on federated ensembles.

DetailsMotivation: Federated inference methods (one-shot FL, edge ensembles, federated ensembles) are vulnerable to attacks, but their robustness has been largely neglected in previous research.

Method: Analyzed averaging-based aggregators, cast robust federated inference with non-linear aggregators as adversarial ML problem, and proposed DeepSet aggregation model with adversarial training and test-time robust aggregation.

Result: The composition of adversarial training and test-time robust aggregation achieved significant improvements, surpassing existing robust aggregation methods by 4.7-22.2% accuracy across diverse benchmarks.

Conclusion: The proposed approach effectively robustifies federated inference against attacks, addressing a critical gap in the security of federated learning systems.

Abstract: Federated inference, in the form of one-shot federated learning, edge ensembles, or federated ensembles, has emerged as an attractive solution to combine predictions from multiple models. This paradigm enables each model to remain local and proprietary while a central server queries them and aggregates predictions. Yet, the robustness of federated inference has been largely neglected, leaving them vulnerable to even simple attacks. To address this critical gap, we formalize the problem of robust federated inference and provide the first robustness analysis of this class of methods. Our analysis of averaging-based aggregators shows that the error of the aggregator is small either when the dissimilarity between honest responses is small or the margin between the two most probable classes is large. Moving beyond linear averaging, we show that problem of robust federated inference with non-linear aggregators can be cast as an adversarial machine learning problem. We then introduce an advanced technique using the DeepSet aggregation model, proposing a novel composition of adversarial training and test-time robust aggregation to robustify non-linear aggregators. Our composition yields significant improvements, surpassing existing robust aggregation methods by 4.7 - 22.2% in accuracy points across diverse benchmarks.

[632] Unifying Polymer Modeling and Design via a Conformation-Centric Generative Foundation Model

Fanmeng Wang, Shan Mei, Wentao Guo, Hongshuai Wang, Qi Ou, Zhifeng Gao, Hongteng Xu

Main category: cs.LG

TL;DR: PolyConFM is the first polymer foundation model that uses conformation-centric generative pretraining to address limitations of existing methods that overlook global structural information in polymer modeling.

DetailsMotivation: Existing deep learning methods for polymers only use monomer-level descriptors and lack global structural information from polymer conformations, limiting performance. The field also lacks a universal foundation model for diverse downstream tasks.

Method: PolyConFM uses conformation-centric generative pretraining by decomposing polymer conformations into sequences of local conformations. It employs masked autoregressive modeling to reconstruct local conformations and generate orientation transformations to recover full polymer conformations.

Result: PolyConFM consistently outperforms representative task-specific methods on diverse downstream tasks, demonstrating superior performance across various polymer science applications.

Conclusion: PolyConFM provides polymer science with a universal and powerful foundation model that effectively supports diverse downstream tasks through conformation-centric generative pretraining.

Abstract: Polymers, macromolecules formed from covalently bonded monomers, underpin countless technologies and are indispensable to modern life. While deep learning is advancing polymer science, existing methods typically represent the whole polymer solely through monomer-level descriptors, overlooking the global structural information inherent in polymer conformations, which ultimately limits their practical performance. Moreover, this field still lacks a universal foundation model that can effectively support diverse downstream tasks, thereby severely constraining progress. To address these challenges, we introduce PolyConFM, the first polymer foundation model that unifies polymer modeling and design through conformation-centric generative pretraining. Recognizing that each polymer conformation can be decomposed into a sequence of local conformations (i.e., those of its repeating units), we pretrain PolyConFM under the conditional generation paradigm, reconstructing these local conformations via masked autoregressive (MAR) modeling and further generating their orientation transformations to recover the corresponding polymer conformation. Besides, we construct the first high-quality polymer conformation dataset via molecular dynamics simulations to mitigate data sparsity, thereby enabling conformation-centric pretraining. Experiments demonstrate that PolyConFM consistently outperforms representative task-specific methods on diverse downstream tasks, equipping polymer science with a universal and powerful tool.

[633] A tutorial on discovering and quantifying the effect of latent causal sources of multimodal EHR data

Marco Barbero-Mota, Eric V. Strobl, John M. Still, William W. Stead, Thomas A. Lasko

Main category: cs.LG

TL;DR: A causal machine learning pipeline for discovering latent causal sources in electronic health records and quantifying their effects on clinical outcomes.

DetailsMotivation: To address the challenge of analyzing imperfect multimodal clinical data and discovering causal relationships in large-scale electronic health records for medical discovery.

Method: Process imperfect multimodal clinical data, decompose into probabilistic independent latent sources, and train task-specific causal models to estimate individual causal effects.

Result: Successfully applied the approach in two real-world applications, demonstrating its versatility and utility for medical discovery at scale.

Conclusion: The proposed pipeline provides an accessible and generalizable method for causal discovery and effect quantification in electronic health records, showing promise for scalable medical research.

Abstract: We provide an accessible description of a peer-reviewed generalizable causal machine learning pipeline to (i) discover latent causal sources of large-scale electronic health records observations, and (ii) quantify the source causal effects on clinical outcomes. We illustrate how imperfect multimodal clinical data can be processed, decomposed into probabilistic independent latent sources, and used to train taskspecific causal models from which individual causal effects can be estimated. We summarize the findings of the two real-world applications of the approach to date as a demonstration of its versatility and utility for medical discovery at scale.

[634] RoBCtrl: Attacking GNN-Based Social Bot Detectors via Reinforced Manipulation of Bots Control Interaction

Yingguang Yang, Xianghua Zeng, Qi Wu, Hao Peng, Yutong Xia, Hao Liu, Bin Chong, Philip S. Yu

Main category: cs.LG

TL;DR: This paper proposes RoBCtrl, the first adversarial multi-agent reinforcement learning framework for social bot control attacks targeting GNN-based bot detectors, using diffusion models to generate realistic bot accounts and MARL to optimize evasion strategies.

DetailsMotivation: Social networks are crucial information sources, but current GNN-based bot detection methods have underexplored vulnerabilities due to limited control over social agents, black-box detectors, and bot heterogeneity.

Method: Uses diffusion models to generate high-fidelity bot accounts by reconstructing existing data with minor modifications, then employs Multi-Agent Reinforcement Learning (MARL) to simulate adversarial bot behavior across different account categories with hierarchical state abstraction for acceleration.

Result: Extensive experiments show the framework effectively undermines GNN-based bot detector performance.

Conclusion: RoBCtrl successfully demonstrates vulnerabilities in current GNN-based social bot detection systems through adversarial attacks combining diffusion models and multi-agent reinforcement learning.

Abstract: Social networks have become a crucial source of real-time information for individuals. The influence of social bots within these platforms has garnered considerable attention from researchers, leading to the development of numerous detection technologies. However, the vulnerability and robustness of these detection methods is still underexplored. Existing Graph Neural Network (GNN)-based methods cannot be directly applied due to the issues of limited control over social agents, the black-box nature of bot detectors, and the heterogeneity of bots. To address these challenges, this paper proposes the first adversarial multi-agent Reinforcement learning framework for social Bot control attacks (RoBCtrl) targeting GNN-based social bot detectors. Specifically, we use a diffusion model to generate high-fidelity bot accounts by reconstructing existing account data with minor modifications, thereby evading detection on social platforms. To the best of our knowledge, this is the first application of diffusion models to mimic the behavior of evolving social bots effectively. We then employ a Multi-Agent Reinforcement Learning (MARL) method to simulate bots adversarial behavior. We categorize social accounts based on their influence and budget. Different agents are then employed to control bot accounts across various categories, optimizing the attachment strategy through reinforcement learning. Additionally, a hierarchical state abstraction based on structural entropy is designed to accelerate the reinforcement learning. Extensive experiments on social bot detection datasets demonstrate that our framework can effectively undermine the performance of GNN-based detectors.

[635] Vector Quantization in the Brain: Grid-like Codes in World Models

Xiangyuan Peng, Xingsi Dong, Si Wu

Main category: cs.LG

TL;DR: GCQ is a brain-inspired method that compresses observation-action sequences into discrete representations using grid-like patterns and attractor dynamics, enabling spatiotemporal compression and serving as a unified world model.

DetailsMotivation: To develop a more efficient method for compressing observation-action sequences that can handle both spatial and temporal information simultaneously, unlike conventional vector quantization approaches that only work on static inputs.

Method: Uses action-conditioned codebook with codewords derived from continuous attractor neural networks, dynamically selecting codewords based on actions to perform spatiotemporal compression.

Result: GCQ effectively compresses sequences into compact representations that support long-horizon prediction, goal-directed planning, and inverse modeling across diverse tasks.

Conclusion: GCQ provides both a practical computational tool for efficient sequence modeling and theoretical insights into how grid-like codes form in neural systems.

Abstract: We propose Grid-like Code Quantization (GCQ), a brain-inspired method for compressing observation-action sequences into discrete representations using grid-like patterns in attractor dynamics. Unlike conventional vector quantization approaches that operate on static inputs, GCQ performs spatiotemporal compression through an action-conditioned codebook, where codewords are derived from continuous attractor neural networks and dynamically selected based on actions. This enables GCQ to jointly compress space and time, serving as a unified world model. The resulting representation supports long-horizon prediction, goal-directed planning, and inverse modeling. Experiments across diverse tasks demonstrate GCQ’s effectiveness in compact encoding and downstream performance. Our work offers both a computational tool for efficient sequence modeling and a theoretical perspective on the formation of grid-like codes in neural systems.

[636] AMS-QUANT: Adaptive Mantissa Sharing for Floating-point Quantization

Mengtao Lv, Ruiqi Zhu, Xinyu Wang, Yun Li

Main category: cs.LG

TL;DR: AMS-Quant introduces non-integer floating-point quantization for LLMs using mantissa-bit sharing and adaptive searching, achieving 2.8-3.2x speedup with FP5.33 and FP4.25 formats while maintaining accuracy.

DetailsMotivation: Large language models have storage and efficiency bottlenecks due to their massive parameters. Floating-point quantization can speed up inference but existing methods use integer bitwidths, leaving room for optimization.

Method: Proposes AMS-Quant with two techniques: (1) Mantissa-bit Sharing - groups weights to share least significant mantissa bits, enabling non-integer bitwidths; (2) Adaptive Searching - offline optimization to minimize accuracy loss from sharing. Also implements efficient CUDA kernels.

Result: Successfully quantizes models to FP5.33-e2m3 and FP4.25-e2m2 formats, achieving 2.8x and 3.2x speedup over FP16 inference respectively, with negligible accuracy loss on large-scale datasets and models.

Conclusion: AMS-Quant demonstrates that non-integer floating-point quantization through mantissa-bit sharing and adaptive optimization can significantly accelerate LLM inference while preserving model accuracy, advancing beyond traditional integer bitwidth quantization approaches.

Abstract: Large language models (LLMs) have demonstrated remarkable capabilities in various kinds of tasks, while the billion or even trillion parameters bring storage and efficiency bottlenecks for inference. Quantization, particularly floating-point quantization, is known to be capable of speeding up LLM inference by reducing memory footprint and data movement during the inference process. For the first time, we advance the floating-point quantization exploration from integer bitwidths to non-integer bit-widths, namely AMS-Quant, to further approach the quantization sweet spot. AMS-Quant incorporates two novel techniques to put it into effect: (1) it proposes Mantissa-bit Sharing, which groups k quantized weights and lets them share the least significant mantissa bit, allowing us to further approach the minimum quantization bit-width without accuracy loss. (2) It introduces Adaptive Searching, which employs an offline optimization strategy to minimize the accuracy degradation introduced by sharing. Moreover, AMS-Quant is also prototyped as efficient CUDA Linear kernels, which translates memory savings into wall-clock latency reduction by reducing memory access. Extensive experiments on large-scale datasets and models show that AMS-Quant can quantize the model to FP-5.33-e2m3 and FP4.25-e2m2, and significantly speed up the LLM decoding over FP16 inference (2.8x and 3.2x), with negligible accuracy loss.

[637] GUIrilla: A Scalable Framework for Automated Desktop UI Exploration

Sofiya Garkot, Maksym Shamrai, Ivan Synytsia, Mariya Hirna

Main category: cs.LG

TL;DR: GUIrilla is an automated framework for collecting GUI interaction data on macOS using accessibility APIs, creating a large dataset (GUIrilla-Task) that improves LLM-based UI automation performance with significantly less data.

DetailsMotivation: Current UI automation faces data limitations due to costly manual annotation, closed-source datasets, and superficial synthetic pipelines, especially for complex desktop environments like macOS.

Method: Systematic exploration of applications via native accessibility APIs, organizing elements into hierarchical GUI graphs, using specialized interaction handlers for comprehensive coverage, and constructing a large-scale dataset with screenshots and action traces.

Result: Created GUIrilla-Task dataset with 27,171 tasks across 1,108 macOS apps; tuning LLM agents on this data significantly improves UI task performance, outperforming synthetic baselines while using 97% less data.

Conclusion: GUIrilla addresses critical data collection challenges in GUI automation and enables more effective desktop autonomy through systematic data collection and open-source tools.

Abstract: Autonomous agents capable of operating complex graphical user interfaces (GUIs) have the potential to transform desktop automation. While recent advances in large language models (LLMs) have significantly improved UI understanding, navigating full-window, multi-application desktop environments remains a major challenge. Data availability is limited by costly manual annotation, closed-source datasets and surface-level synthetic pipelines. We introduce GUIrilla, an automated scalable framework that systematically explores applications via native accessibility APIs to address the critical data collection challenge in GUI automation. Our framework focuses on macOS - an ecosystem with limited representation in current UI datasets - though many of its components are designed for broader cross-platform applicability. GUIrilla organizes discovered interface elements and crawler actions into hierarchical GUI graphs and employs specialized interaction handlers to achieve comprehensive application coverage. Using the application graphs from GUIrilla crawler, we construct and release GUIrilla-Task, a large-scale dataset of 27,171 functionally grounded tasks across 1,108 macOS applications, each annotated with full-desktop and window-level screenshots, accessibility metadata, and semantic action traces. Empirical results show that tuning LLM-based agents on GUIrilla-Task significantly improves performance on downstream UI tasks, outperforming synthetic baselines on the ScreenSpot Pro benchmark while using 97% less data. We also release macapptree, an open-source library for reproducible collection of structured accessibility metadata, along with the full GUIrilla-Task dataset, the manually verified GUIrilla-Gold benchmark, and the framework code to support open research in desktop autonomy.

[638] FUSE-Traffic: Fusion of Unstructured and Structured Data for Event-aware Traffic Forecasting

Chenyang Yu, Xinpeng Xie, Yan Huang, Chenxi Qiu

Main category: cs.LG

TL;DR: This paper discusses traffic forecasting using Graph Neural Networks (GNNs) to capture spatial-temporal dependencies in traffic data, highlighting challenges in handling unexpected events and limitations of manual feature engineering approaches.

DetailsMotivation: Accurate traffic forecasting is crucial for Intelligent Transportation Systems to address growing urban congestion and improve travel experiences. The need for reliable models that can handle both regular traffic patterns and unexpected events motivates this research.

Method: The paper reviews GNN-based approaches including STGCN, GraphWaveNet, STWave, and D2STGNN that capture spatial dependencies through graph convolutions and temporal patterns. It also discusses early attempts using manually engineered event features and incident effect scores.

Result: GNN models have achieved impressive performance on standard traffic datasets, particularly in capturing periodic regularities. However, manual event feature engineering approaches suffer from heavy reliance on domain expertise, poor generalization to unknown events, and loss of semantic details.

Conclusion: While GNNs are effective for regular traffic patterns, current event-handling approaches using manual feature engineering are insufficient. There is a need for more robust methods that can better incorporate event information without heavy reliance on domain knowledge.

Abstract: Accurate traffic forecasting is a core technology for building Intelligent Transportation Systems (ITS), enabling better urban resource allocation and improved travel experiences. With growing urbanization, traffic congestion has intensified, highlighting the need for reliable and responsive forecasting models. In recent years, deep learning, particularly Graph Neural Networks (GNNs), has emerged as the mainstream paradigm in traffic forecasting. GNNs can effectively capture complex spatial dependencies in road network topology and dynamic temporal evolution patterns in traffic flow data. Foundational models such as STGCN and GraphWaveNet, along with more recent developments including STWave and D2STGNN, have achieved impressive performance on standard traffic datasets. These approaches incorporate sophisticated graph convolutional structures and temporal modeling mechanisms, demonstrating particular effectiveness in capturing and forecasting traffic patterns characterized by periodic regularities. To address this challenge, researchers have explored various ways to incorporate event information. Early attempts primarily relied on manually engineered event features. For instance, some approaches introduced manually defined incident effect scores or constructed specific subgraphs for different event-induced traffic conditions. While these methods somewhat enhance responsiveness to specific events, their core drawback lies in a heavy reliance on domain experts’ prior knowledge, making generalization to diverse and complex unknown events difficult, and low-dimensional manual features often lead to the loss of rich semantic details.

[639] Beyond Accuracy: Are Time Series Foundation Models Well-Calibrated?

Coen Adler, Yuxin Chang, Felix Draxler, Samar Abdi, Padhraic Smyth

Main category: cs.LG

TL;DR: Time series foundation models show better calibration than baselines and avoid systematic overconfidence, unlike typical deep learning models.

DetailsMotivation: Foundation models for time series achieve state-of-the-art performance but their calibration properties remain underexplored, despite calibration being critical for practical applications.

Method: Systematic evaluation of five recent time series foundation models and two competitive baselines, assessing model calibration, effects of varying prediction heads, and calibration under long-term autoregressive forecasting.

Result: Time series foundation models are consistently better calibrated than baseline models and tend not to be systematically over- or under-confident.

Conclusion: Time series foundation models demonstrate superior calibration properties compared to traditional deep learning approaches, avoiding the overconfidence issues commonly seen in other deep learning models.

Abstract: The recent development of foundation models for time series data has generated considerable interest in using such models across a variety of applications. Although foundation models achieve state-of-the-art predictive performance, their calibration properties remain relatively underexplored, despite the fact that calibration can be critical for many practical applications. In this paper, we investigate the calibration-related properties of five recent time series foundation models and two competitive baselines. We perform a series of systematic evaluations assessing model calibration (i.e., over- or under-confidence), effects of varying prediction heads, and calibration under long-term autoregressive forecasting. We find that time series foundation models are consistently better calibrated than baseline models and tend not to be either systematically over- or under-confident, in contrast to the overconfidence often seen in other deep learning models.

[640] Learning a Generalized Model for Substation Level Voltage Estimation in Distribution Networks

Muhy Eddin Za’ter, Bri-Mathias Hodge

Main category: cs.LG

TL;DR: Hierarchical graph neural network for substation-level voltage estimation that outperforms alternative models with 2x lower RMSE and works with only 1% measurement coverage.

DetailsMotivation: Traditional distribution system state estimation (DSSE) struggles with sparse measurements and scalability in modern distribution networks with high DER penetration and voltage variability.

Method: Proposes a hierarchical graph neural network that exploits electrical topology and physical features, trained and evaluated on SMART-DS datasets across multiple substations and DER penetration scenarios.

Result: Achieves up to 2 times lower RMSE than alternative data-driven models and maintains high accuracy with only 1% measurement coverage.

Conclusion: GNNs have potential to enable scalable, reproducible, and data-driven voltage monitoring for distribution systems.

Abstract: Accurate voltage estimation in distribution networks is critical for real-time monitoring and increasing the reliability of the grid. As DER penetration and distribution level voltage variability increase, robust distribution system state estimation (DSSE) has become more essential to maintain safe and efficient operations. Traditional DSSE techniques, however, struggle with sparse measurements and the scale of modern feeders, limiting their scalability to large networks. This paper presents a hierarchical graph neural network for substation-level voltage estimation that exploits both electrical topology and physical features, while remaining robust to the low observability levels common to real-world distribution networks. Leveraging the public SMART-DS datasets, the model is trained and evaluated on thousands of buses across multiple substations and DER penetration scenarios. Comprehensive experiments demonstrate that the proposed method achieves up to 2 times lower RMSE than alternative data-driven models, and maintains high accuracy with as little as 1% measurement coverage. The results highlight the potential of GNNs to enable scalable, reproducible, and data-driven voltage monitoring for distribution systems.

[641] Residual Correction Models for AC Optimal Power Flow Using DC Optimal Power Flow Solutions

Muhy Eddin Za’ter, Bri-Mathias Hodge, Kyri Baker

Main category: cs.LG

TL;DR: A residual learning approach using DC OPF solutions as baseline and learning nonlinear corrections to achieve AC-OPF solutions with improved accuracy and speed.

DetailsMotivation: To overcome the computational bottleneck of solving nonlinear AC optimal power flow problems for real-time grid operations.

Method: Uses topology-aware Graph Neural Network with local attention and two-level DC feature integration, trained with physics-informed loss to enforce AC power-flow feasibility and operational limits.

Result: 25% lower MSE, up to 3X reduction in feasibility error, and up to 13X runtime speedup compared to conventional AC OPF solvers, with maintained accuracy under N-1 contingencies.

Conclusion: Residual learning provides a practical and scalable bridge between linear approximations and AC-feasible OPF, enabling near real-time operational decision making.

Abstract: Solving the nonlinear AC optimal power flow (AC OPF) problem remains a major computational bottleneck for real-time grid operations. In this paper, we propose a residual learning paradigm that uses fast DC optimal power flow (DC OPF) solutions as a baseline, and learns only the nonlinear corrections required to provide the full AC-OPF solution. The method utilizes a topology-aware Graph Neural Network with local attention and two-level DC feature integration, trained using a physics-informed loss that enforces AC power-flow feasibility and operational limits. Evaluations on OPFData for 57-, 118-, and 2000-bus systems show around 25% lower MSE, up to 3X reduction in feasibility error, and up to 13X runtime speedup compared to conventional AC OPF solvers. The model maintains accuracy under N-1 contingencies and scales efficiently to large networks. These results demonstrate that residual learning is a practical and scalable bridge between linear approximations and AC-feasible OPF, enabling near real-time operational decision making.

[642] FedPURIN: Programmed Update and Reduced INformation for Sparse Personalized Federated Learning

Lunchen Xie, Zehua He, Qingjiang Shi

Main category: cs.LG

TL;DR: FedPURIN is a communication-efficient personalized federated learning framework that uses integer programming to identify critical parameters for transmission, achieving significant communication reduction while maintaining competitive performance.

DetailsMotivation: To address the suboptimal communication efficiency in existing PFL methods that sustain substantial communication burdens, which impedes practical deployment, especially for edge intelligence systems with heterogeneous data.

Method: Uses an integer programming formulation to strategically identify critical parameters for transmission and integrates this into a sparse aggregation scheme for communication-efficient PFL.

Result: Comprehensive evaluations on standard image classification benchmarks under varied non-IID conditions demonstrate competitive performance relative to state-of-the-art methods with quantifiable communication reduction through sparse aggregation.

Conclusion: FedPURIN establishes a new paradigm for communication-efficient PFL that is particularly advantageous for edge intelligence systems operating with heterogeneous data sources.

Abstract: Personalized Federated Learning (PFL) has emerged as a critical research frontier addressing data heterogeneity issue across distributed clients. Novel model architectures and collaboration mechanisms are engineered to accommodate statistical disparities while producing client-specific models. Parameter decoupling represents a promising paradigm for maintaining model performance in PFL frameworks. However, the communication efficiency of many existing methods remains suboptimal, sustaining substantial communication burdens that impede practical deployment. To bridge this gap, we propose Federated Learning with Programmed Update and Reduced INformation (FedPURIN), a novel framework that strategically identifies critical parameters for transmission through an integer programming formulation. This mathematically grounded strategy is seamlessly integrated into a sparse aggregation scheme, achieving a significant communication reduction while preserving the efficacy. Comprehensive evaluations on standard image classification benchmarks under varied non-IID conditions demonstrate competitive performance relative to state-of-the-art methods, coupled with quantifiable communication reduction through sparse aggregation. The framework establishes a new paradigm for communication-efficient PFL, particularly advantageous for edge intelligence systems operating with heterogeneous data sources.

[643] MNO: Multiscale Neural Operator for Computational Fluid Dynamics with 3D Point Cloud Data

Qinxuan Wang, Chuang Wang, Mingyu Zhang, Jingwei Sun, Peipei Yang, Shuo Tang, Shiming Xiang

Main category: cs.LG

TL;DR: MNO is a multiscale neural operator for 3D CFD on unstructured point clouds that decomposes information across global, local, and micro scales using attention modules, achieving significant error reduction over state-of-the-art methods.

DetailsMotivation: Existing neural operators for PDEs suffer from limited accuracy and scalability on irregular domains where fluid flows exhibit rich multiscale structures.

Method: MNO explicitly decomposes information across three scales: global dimension-shrinkage attention for long-range dependencies, local graph attention for neighborhood interactions, and micro point-wise attention for fine details.

Result: MNO consistently outperforms state-of-the-art baselines across four diverse benchmarks, reducing prediction errors by 5% to 40% and demonstrating improved robustness in challenging 3D CFD problems with up to 300K points.

Conclusion: The explicit multiscale design is crucial for neural operators, and MNO establishes a scalable framework for learning complex fluid dynamics on irregular domains.

Abstract: Neural operators have emerged as a powerful data-driven paradigm for solving Partial Differential Equations (PDEs), offering orders-of-magnitude acceleration over traditional solvers. However, existing approaches still suffer from limited accuracy and scalability, particularly on irregular domains where fluid flows exhibit rich multiscale structures. In this work, we introduce the Multiscale Neural Operator (MNO), a new architecture for Computational Fluid Dynamics (CFD) on three-dimensional (3D) unstructured point clouds. MNO explicitly decomposes information across three scales: a global dimension-shrinkage attention module for long-range dependencies, a local graph attention module for neighborhood-level interactions, and a micro point-wise attention module for fine-grained details. This design preserves multiscale inductive biases while remaining computationally efficient. We evaluate MNO on four diverse benchmarks, covering both steady-state and unsteady flow scenarios with up to 300K points. Across all tasks, MNO consistently outperforms state-of-the-art baselines, reducing prediction errors by 5% to 40% and demonstrating improved robustness in challenging 3D CFD problems. Our results highlight the importance of explicit multiscale design for neural operators and establish MNO as a scalable framework for learning complex fluid dynamics on irregular domains.

[644] Early-stopping for Transformer model training

Jing He, Hua Jiang, Cheng Li, Siqian Xin, Shuzhen Yang

Main category: cs.LG

TL;DR: A Random Matrix Theory framework for analyzing Transformer training dynamics, identifying three training stages and proposing validation-free early-stopping criteria based on spectral properties.

DetailsMotivation: To understand the underlying mechanisms driving Transformer performance improvements and derive principled early-stopping criteria without relying on validation data.

Method: Using Random Matrix Theory to analyze the spectral density evolution of the self-attention matrix V, applying PL (Power Law) fit to identify training stages, and developing spectral-based metrics for convergence detection.

Result: The spectral density of the shallow self-attention matrix consistently evolves into heavy-tailed distribution, revealing three distinct training stages: structural exploration, heavy-tailed structure stabilization, and convergence saturation.

Conclusion: Random Matrix Theory provides effective tools for monitoring Transformer training progression, with spectral signatures offering consistent validation-free criteria for early stopping decisions.

Abstract: This work introduces a novel theoretical framework grounded in Random Matrix Theory (RMT) for analyzing Transformer training dynamics. We focus on the underlying mechanisms that drive performance improvements and derive principled early-stopping criteria. Empirically, we observe that the spectral density of the shallow self-attention matrix V consistently evolves into a heavy-tailed distribution. Utilizing the PL (Power Law) fit to this matrix as a probe, we demarcate training into three stages: structural exploration, heavy-tailed structure stabilization, and convergence saturation. This staging provides guidance for preliminary stopping decisions. Crucially, we propose two consistent and validation-free criteria: a quantitative metric for heavy-tailed dynamics and a novel spectral signature indicative of convergence. The strong alignment between these criteria highlights the utility of RMT for monitoring and diagnosing the progression of Transformer model training.

[645] Optimization of the quantization of dense neural networks from an exact QUBO formulation

Sergio Muñiz Subiñas, Manuel L. González, Jorge Ruiz Gómez, Alejandro Mata Ali, Jorge Martínez Martín, Miguel Franco Hernando, Ángel Miguel García-Vico

Main category: cs.LG

TL;DR: A novel PTQ method using ADAROUND-based QUBO formulation for neural network quantization, with efficient decomposition into independent subproblems solvable via simulated annealing.

DetailsMotivation: To develop an efficient post-training quantization method that minimizes accuracy loss while reducing computational complexity through mathematical optimization.

Method: Uses Frobenius distance between theoretical and dequantized outputs as objective, formulates explicit QUBO with binary variables for rounding choices, and decomposes into independent subproblems of size f+1.

Result: Evaluated on MNIST, Fashion-MNIST, EMNIST, and CIFAR-10 across int8 to int1 precisions, showing improved performance over traditional round-to-nearest quantization.

Conclusion: The proposed ADAROUND-based QUBO formulation with problem decomposition enables efficient neural network quantization with better accuracy preservation than conventional methods.

Abstract: This work introduces a post-training quantization (PTQ) method for dense neural networks via a novel ADAROUND-based QUBO formulation. Using the Frobenius distance between the theoretical output and the dequantized output (before the activation function) as the objective, an explicit QUBO whose binary variables represent the rounding choice for each weight and bias is obtained. Additionally, by exploiting the structure of the coefficient QUBO matrix, the global problem can be exactly decomposed into $n$ independent subproblems of size $f+1$, which can be efficiently solved using some heuristics such as simulated annealing. The approach is evaluated on MNIST, Fashion-MNIST, EMNIST, and CIFAR-10 across integer precisions from int8 to int1 and compared with a round-to-nearest traditional quantization methodology.

[646] BPL: Bias-adaptive Preference Distillation Learning for Recommender System

SeongKu Kang, Jianxun Lian, Dongha Lee, Wonbin Kweon, Sanghwan Jang, Jaehyun Lee, Jindong Wang, Xing Xie, Hwanjo Yu

Main category: cs.LG

TL;DR: BPL is a bias-adaptive preference distillation learning framework that uses dual distillation strategies to achieve high performance in both factual and counterfactual test environments for recommender systems.

DetailsMotivation: Recommender systems suffer from biases in collected feedback, and existing debiasing methods focus on counterfactual test environments but degrade accuracy in factual test environments. Both test environments are important - counterfactual for long-term user satisfaction and factual for predicting user behaviors.

Method: Uses dual distillation strategies: teacher-student distillation from a biased model to retain accurate preference knowledge for factual test performance, and self-distillation with reliability filtering to iteratively refine knowledge for counterfactual test performance.

Result: Comprehensive experiments validate BPL’s effectiveness in both factual and counterfactual tests, showing improved performance across both test environments.

Conclusion: BPL successfully addresses the dual challenge of performing well in both factual and counterfactual test environments through its bias-adaptive preference distillation approach with dual distillation strategies.

Abstract: Recommender systems suffer from biases that cause the collected feedback to incompletely reveal user preference. While debiasing learning has been extensively studied, they mostly focused on the specialized (called counterfactual) test environment simulated by random exposure of items, significantly degrading accuracy in the typical (called factual) test environment based on actual user-item interactions. In fact, each test environment highlights the benefit of a different aspect: the counterfactual test emphasizes user satisfaction in the long-terms, while the factual test focuses on predicting subsequent user behaviors on platforms. Therefore, it is desirable to have a model that performs well on both tests rather than only one. In this work, we introduce a new learning framework, called Bias-adaptive Preference distillation Learning (BPL), to gradually uncover user preferences with dual distillation strategies. These distillation strategies are designed to drive high performance in both factual and counterfactual test environments. Employing a specialized form of teacher-student distillation from a biased model, BPL retains accurate preference knowledge aligned with the collected feedback, leading to high performance in the factual test. Furthermore, through self-distillation with reliability filtering, BPL iteratively refines its knowledge throughout the training process. This enables the model to produce more accurate predictions across a broader range of user-item combinations, thereby improving performance in the counterfactual test. Comprehensive experiments validate the effectiveness of BPL in both factual and counterfactual tests. Our implementation is accessible via: https://github.com/SeongKu-Kang/BPL.

[647] Continual Knowledge Consolidation LORA for Domain Incremental Learning

Naeem Paeedeh, Mahardhika Pratama, Weiping Ding, Jimmy Cao, Wolfgang Mayer, Ryszard Kowalczyk

Main category: cs.LG

TL;DR: CONEC-LoRA is a novel approach for Domain Incremental Learning that combines task-shared and task-specific LoRA modules with a stochastic classifier and auxiliary network to prevent catastrophic forgetting and improve generalization.

DetailsMotivation: Existing DIL methods using parameter-efficient fine-tuning create task-specific LoRAs but overlook shared knowledge across tasks, leading to inaccurate LoRA selection during inference and suboptimal generalization with linear/prototype classifiers.

Method: Proposes CONEC-LoRA with: 1) Consolidation between task-shared and task-specific LoRAs, 2) Stochastic classifier with parameters sampled from distribution, 3) Auxiliary network for optimal LoRA prediction using different-depth structure with local classifiers, 4) Ball-generator loss and transformation module to address synthetic sample bias.

Result: CONEC-LoRA outperforms prior methods by over 5% margins across 4 popular benchmark problems, demonstrating significant improvements in domain incremental learning performance.

Conclusion: The proposed CONEC-LoRA framework effectively addresses DIL challenges by leveraging shared knowledge, stochastic classification, and intermediate representations, achieving state-of-the-art performance in preventing catastrophic forgetting while adapting to new domains.

Abstract: Domain Incremental Learning (DIL) is a continual learning sub-branch that aims to address never-ending arrivals of new domains without catastrophic forgetting problems. Despite the advent of parameter-efficient fine-tuning (PEFT) approaches, existing works create task-specific LoRAs overlooking shared knowledge across tasks. Inaccurate selection of task-specific LORAs during inference results in significant drops in accuracy, while existing works rely on linear or prototype-based classifiers, which have suboptimal generalization powers. Our paper proposes continual knowledge consolidation low rank adaptation (CONEC-LoRA) addressing the DIL problems. CONEC-LoRA is developed from consolidations between task-shared LORA to extract common knowledge and task-specific LORA to embrace domain-specific knowledge. Unlike existing approaches, CONEC-LoRA integrates the concept of a stochastic classifier whose parameters are sampled from a distribution, thus enhancing the likelihood of correct classifications. Last but not least, an auxiliary network is deployed to optimally predict the task-specific LoRAs for inferences and implements the concept of a different-depth network structure in which every layer is connected with a local classifier to take advantage of intermediate representations. This module integrates the ball-generator loss and transformation module to address the synthetic sample bias problem. Our rigorous experiments demonstrate the advantage of CONEC-LoRA over prior arts in 4 popular benchmark problems with over 5% margins.

[648] PassREfinder-FL: Privacy-Preserving Credential Stuffing Risk Prediction via Graph-Based Federated Learning for Representing Password Reuse between Websites

Jaehan Kim, Minkyoo Song, Minjae Seo, Youngjin Jin, Seungwon Shin, Jinwoo Kim

Main category: cs.LG

TL;DR: PassREfinder-FL is a federated learning framework that predicts credential stuffing risks by modeling password reuse relations as edges in a website graph using graph neural networks, achieving high accuracy while preserving user privacy.

DetailsMotivation: Credential stuffing attacks harm users who reuse passwords across websites. Existing detection methods compromise usability and face deployment challenges due to complex account-sharing mechanisms.

Method: Proposes password reuse relations as edges in a website graph and uses graph neural networks for link prediction. Incorporates federated learning to preserve user privacy by eliminating cross-administrator data sharing.

Result: Achieves F1-score of 0.9153 on real-world dataset of 360M breached accounts from 22,378 websites. FL-based GNN shows 4-11% performance improvement over other state-of-the-art GNN models.

Conclusion: The framework effectively predicts credential stuffing risks, provides actionable risk scores for password reuse likelihood, and maintains user privacy through federated learning approach.

Abstract: Credential stuffing attacks have caused significant harm to online users who frequently reuse passwords across multiple websites. While prior research has attempted to detect users with reused passwords or identify malicious login attempts, existing methods often compromise usability by restricting password creation or website access, and their reliance on complex account-sharing mechanisms hinders real-world deployment. To address these limitations, we propose PassREfinder-FL, a novel framework that predicts credential stuffing risks across websites. We introduce the concept of password reuse relations – defined as the likelihood of users reusing passwords between websites – and represent them as edges in a website graph. Using graph neural networks (GNNs), we perform a link prediction task to assess credential reuse risk between sites. Our approach scales to a large number of arbitrary websites by incorporating public website information and linking newly observed websites as nodes in the graph. To preserve user privacy, we extend PassREfinder-FL with a federated learning (FL) approach that eliminates the need to share user sensitive information across administrators. Evaluation on a real-world dataset of 360 million breached accounts from 22,378 websites shows that PassREfinder-FL achieves an F1-score of 0.9153 in the FL setting. We further validate that our FL-based GNN achieves a 4-11% performance improvement over other state-of-the-art GNN models through an ablation study. Finally, we demonstrate that the predicted results can be used to quantify password reuse likelihood as actionable risk scores.

[649] Near-Equilibrium Propagation training in nonlinear wave systems

Karol Sajnok, Michał Matuszewski

Main category: cs.LG

TL;DR: Extended Equilibrium Propagation to complex-valued wave systems, enabling in-situ training in weakly dissipative physical systems with local parameter control.

DetailsMotivation: Backpropagation is difficult to implement in physical neural networks, so Equilibrium Propagation offers an alternative for in-situ training in physical systems.

Method: Extended EP learning to discrete and continuous complex-valued wave systems, replacing trainable inter-node connections with trainable local potentials in systems without well-defined nodes.

Result: Tested in driven-dissipative exciton-polariton condensates; numerical studies on logical tasks and handwritten-digit recognition showed stable convergence.

Conclusion: Establishes practical route to in-situ learning in physical systems where system control is restricted to local parameters.

Abstract: Backpropagation learning algorithm, the workhorse of modern artificial intelligence, is notoriously difficult to implement in physical neural networks. Equilibrium Propagation (EP) is an alternative with comparable efficiency and strong potential for in-situ training. We extend EP learning to both discrete and continuous complex-valued wave systems. In contrast to previous EP implementations, our scheme is valid in the weakly dissipative regime, and readily applicable to a wide range of physical settings, even without well defined nodes, where trainable inter-node connections can be replaced by trainable local potential. We test the method in driven-dissipative exciton-polariton condensates governed by generalized Gross-Pitaevskii dynamics. Numerical studies on standard benchmarks, including a simple logical task and handwritten-digit recognition, demonstrate stable convergence, establishing a practical route to in-situ learning in physical systems in which system control is restricted to local parameters.

[650] FSRF: Factorization-guided Semantic Recovery for Incomplete Multimodal Sentiment Analysis

Ziyang Liu, Pengjunfei Chu, Shuming Dong, Chen Zhang, Mingcheng Li, Jin Wang

Main category: cs.LG

TL;DR: FSRF is a framework that addresses modality missing in multimodal sentiment analysis through factorization and self-distillation techniques.

DetailsMotivation: Previous MSA methods assume complete multimodal data, but real-world scenarios often have missing modalities due to occlusion, privacy, or device issues, leading to poor generalizability.

Method: Uses de-redundant homo-heterogeneous factorization to separate modality representations, and distribution-aligned self-distillation for semantic recovery through bidirectional knowledge transfer.

Result: Comprehensive experiments show FSRF significantly outperforms previous methods when dealing with uncertain missing modalities.

Conclusion: FSRF effectively mitigates the modality missing problem in MSA through factorization-guided semantic recovery, improving model robustness in real-world scenarios.

Abstract: In recent years, Multimodal Sentiment Analysis (MSA) has become a research hotspot that aims to utilize multimodal data for human sentiment understanding. Previous MSA studies have mainly focused on performing interaction and fusion on complete multimodal data, ignoring the problem of missing modalities in real-world applications due to occlusion, personal privacy constraints, and device malfunctions, resulting in low generalizability. To this end, we propose a Factorization-guided Semantic Recovery Framework (FSRF) to mitigate the modality missing problem in the MSA task. Specifically, we propose a de-redundant homo-heterogeneous factorization module that factorizes modality into modality-homogeneous, modality-heterogeneous, and noisy representations and design elaborate constraint paradigms for representation learning. Furthermore, we design a distribution-aligned self-distillation module that fully recovers the missing semantics by utilizing bidirectional knowledge transfer. Comprehensive experiments on two datasets indicate that FSRF has a significant performance advantage over previous methods with uncertain missing modalities.

[651] STABLE: Gated Continual Learning for Large Language Models

William Hoy, Nurcin Celik

Main category: cs.LG

TL;DR: STABLE is a gated continual self-editing framework that uses LoRA-based parameter efficient fine-tuning to prevent catastrophic forgetting during sequential updates to LLMs, with three gating metrics (EM drop, bits increase, KL divergence) to constrain updates.

DetailsMotivation: Large language models need continual adaptation mechanisms without full retraining, but sequential updates cause catastrophic forgetting where new edits degrade previously acquired knowledge.

Method: Uses parameter efficient fine-tuning via LoRA with gating mechanism. Each candidate edit is evaluated against stability budget using three metrics: Exact Match drop, bits increase, or KL divergence. Updates exceeding thresholds are rescaled through clipping or rejected.

Result: Experiments on Qwen-2.5-7B show gating effectively mitigates forgetting while preserving adaptability. EM-based gating achieved highest cumulative performance in short continual learning sequences. Different gating strategies achieve comparable distribution shift but different accuracy outcomes.

Conclusion: STABLE provides a principled method for continual model editing, enabling LLMs to integrate new knowledge while maintaining reliability, with gating design being crucial for continual adaptation.

Abstract: Large language models (LLMs) increasingly require mechanisms for continual adaptation without full retraining. However, sequential updates can lead to catastrophic forgetting, where new edits degrade previously acquired knowledge. This work presents STABLE, a gated continual self editing framework that constrains forgetting during sequential updates using parameter efficient fine tuning via Low Rank Adaptation (LoRA; see arXiv:2106.09685). Each candidate edit is evaluated against a stability budget using one of three metrics: (i) Exact Match (EM) drop, capturing factual accuracy loss; (ii) bits increase, reflecting reduced model confidence; and (iii) KL divergence, quantifying distributional drift between the base and adapted models. If a threshold is exceeded, the LoRA update is rescaled through a clipping procedure or rejected. Experiments on the Qwen-2.5-7B model show that gating effectively mitigates forgetting while preserving adaptability. EM based gating achieved the highest cumulative performance in short continual learning sequences. Our results show that different gating strategies can achieve comparable distribution shift (measured by KL divergence) while producing different accuracy outcomes, highlighting the importance of gating design in continual adaptation. This approach offers a principled method for continual model editing, enabling LLMs to integrate new knowledge while maintaining reliability. Code: https://github.com/Bhoy1/STABLE

[652] Compressing Many-Shots in In-Context Learning

Devvrit Khatri, Pranamya Kulkarni, Nilesh Gupta, Yerram Varun, Liqian Peng, Jay Yagnik, Praneeth Netrapalli, Cho-Jui Hsieh, Alec Go, Inderjit S Dhillon, Aditya Kusupati, Prateek Jain

Main category: cs.LG

TL;DR: MemComp is a layer-wise compression method that improves memory and computational efficiency of many-shot in-context learning by compressing prompts into soft-token summaries while maintaining high accuracy.

DetailsMotivation: Large Language Models (LLMs) benefit from more input-output examples (shots) in In-Context Learning (ICL), but this increases memory and computational costs. Existing prompt compression methods are ineffective for many-shot compression, and simply using fewer shots is surprisingly strong but limited.

Method: Proposed MemComp, a layer-wise compression method that uses a stronger compressor model with more trainable parameters and compresses many-shot representations at each transformer layer to enable fine-grained compression.

Result: MemComp outperforms strong baselines across all compression ratios (3x to 8x) on multiple classification tasks with large label sets. While baseline performance degrades sharply (20-30%) at higher compression ratios, MemComp maintains high accuracy with minimal degradation (<10%).

Conclusion: Layer-wise compression with stronger compressor models is effective for many-shot ICL compression, enabling significant memory and computational savings while preserving performance.

Abstract: Large Language Models (LLMs) have been shown to be able to learn different tasks without explicit finetuning when given many input-output examples / demonstrations through In-Context Learning (ICL). Increasing the number of examples, called ``shots’’, improves downstream task performance but incurs higher memory and computational costs. In this work, we study an approach to improve the memory and computational efficiency of ICL inference by compressing the many-shot prompts. Given many shots comprising t tokens, our goal is to generate a m soft-token summary, where m < t. We first show that existing prompt compression methods are ineffective for many-shot compression, and simply using fewer shots as a baseline is surprisingly strong. To achieve effective compression, we find that: (a) a stronger compressor model with more trainable parameters is necessary, and (b) compressing many-shot representations at each transformer layer enables more fine-grained compression by providing each layer with its own compressed representation. Based on these insights, we propose MemCom, a layer-wise compression method. We systematically evaluate various compressor models and training approaches across different model sizes (2B and 7B), architectures (Gemma and Mistral), many-shot sequence lengths (3k-6k tokens), and compression ratios (3x to 8x). MemCom outperforms strong baselines across all compression ratios on multiple classification tasks with large label sets. Notably, while baseline performance degrades sharply at higher compression ratios, often by over 20-30%, MemCom maintains high accuracy with minimal degradation, typically dropping by less than 10%.

[653] Narrowing Action Choices with AI Improves Human Sequential Decisions

Eleni Straitouri, Stratis Tsirtsis, Ander Artola Velasco, Manuel Gomez-Rodriguez

Main category: cs.LG

TL;DR: A decision support system that adaptively controls human agency by narrowing action choices achieves complementarity in sequential decision making, outperforming both humans alone and AI agents alone.

DetailsMotivation: To extend the principle of adaptive human agency control from classification tasks to sequential decision making tasks, achieving complementarity between humans and AI agents.

Method: Developed a decision support system that uses a pre-trained AI agent to narrow down the action set for humans, combined with a bandit algorithm that optimizes the level of human agency by leveraging smoothness properties of action sets.

Result: In a large-scale human study (n=1,600) using a wildfire mitigation game, participants using the system outperformed those playing alone by ~30% and the AI agent by >2%, despite the AI agent outperforming unsupported humans.

Conclusion: The same principle of adaptively controlling human agency can successfully achieve complementarity in sequential decision making tasks, with the proposed system enabling humans to outperform both their own unaided performance and the AI agent used in the system.

Abstract: Recent work has shown that, in classification tasks, it is possible to design decision support systems that do not require human experts to understand when to cede agency to a classifier or when to exercise their own agency to achieve complementarity$\unicode{x2014}$experts using these systems make more accurate predictions than those made by the experts or the classifier alone. The key principle underpinning these systems reduces to adaptively controlling the level of human agency, by design. Can we use the same principle to achieve complementarity in sequential decision making tasks? In this paper, we answer this question affirmatively. We develop a decision support system that uses a pre-trained AI agent to narrow down the set of actions a human can take to a subset, and then asks the human to take an action from this action set. Along the way, we also introduce a bandit algorithm that leverages the smoothness properties of the action sets provided by our system to efficiently optimize the level of human agency. To evaluate our decision support system, we conduct a large-scale human subject study ($n = 1{,}600$) where participants play a wildfire mitigation game. We find that participants who play the game supported by our system outperform those who play on their own by $\sim$$30$% and the AI agent used by our system by $>$$2$%, even though the AI agent largely outperforms participants playing without support. We have made available the data gathered in our human subject study as well as an open source implementation of our system at https://github.com/Networks-Learning/narrowing-action-choices .

[654] Zero-shot World Models via Search in Memory

Federico Malato, Ville Hautamäki

Main category: cs.LG

TL;DR: This paper proposes a search-based world model using similarity search and stochastic representations that approximates world models without training, showing comparable performance to trained models like PlaNet.

DetailsMotivation: To improve sample efficiency in reinforcement learning by developing world models that can model environment transition dynamics without requiring extensive training procedures.

Method: Leverage similarity search and stochastic representations to create a world model without training, comparing it with PlaNet (Dreamer family) on latent reconstruction quality and image similarity for next-step and long-horizon predictions.

Result: The search-based world model performs comparably to training-based models in both next-step and long-horizon predictions, with notably stronger performance in long-horizon prediction across visually diverse environments.

Conclusion: Search-based approaches can effectively approximate world models without training, offering competitive performance especially for long-term predictions in reinforcement learning environments.

Abstract: World Models have vastly permeated the field of Reinforcement Learning. Their ability to model the transition dynamics of an environment have greatly improved sample efficiency in online RL. Among them, the most notorious example is Dreamer, a model that learns to act in a diverse set of image-based environments. In this paper, we leverage similarity search and stochastic representations to approximate a world model without a training procedure. We establish a comparison with PlaNet, a well-established world model of the Dreamer family. We evaluate the models on the quality of latent reconstruction and on the perceived similarity of the reconstructed image, on both next-step and long horizon dynamics prediction. The results of our study demonstrate that a search-based world model is comparable to a training based one in both cases. Notably, our model show stronger performance in long-horizon prediction with respect to the baseline on a range of visually different environments.

[655] A Minimal-Assumption Analysis of Q-Learning with Time-Varying Policies

Phalguni Nanda, Zaiwei Chen

Main category: cs.LG

TL;DR: First finite-time analysis of Q-learning with time-varying learning policies, showing O(1/ε²) sample complexity matching off-policy Q-learning but with different exploration-exploitation trade-offs.

DetailsMotivation: To provide rigorous theoretical guarantees for on-policy Q-learning under minimal assumptions, addressing the analytical challenges of time-varying policies and rapidly time-inhomogeneous Markovian noise.

Method: Employed refined approach using Poisson equation to decompose Markovian noise into martingale-difference and residual terms, with sensitivity analysis of Poisson equation solution to handle time inhomogeneity.

Result: Established last-iterate convergence rates for Q-function estimates and derived explicit rate for policy performance, showing on-policy Q-learning has weaker exploration but better exploitation than off-policy counterpart.

Conclusion: On-policy Q-learning achieves same sample complexity as off-policy but with different exploration-exploitation characteristics, and the developed analytical tools can benefit analysis of other RL algorithms with time-varying policies.

Abstract: In this work, we present the first finite-time analysis of the Q-learning algorithm under time-varying learning policies (i.e., on-policy sampling) with minimal assumptions – specifically, assuming only the existence of a policy that induces an irreducible Markov chain over the state space. We establish a last-iterate convergence rate for $\mathbb{E}[|Q_k - Q^|_\infty^2]$, implying a sample complexity of order $O(1/\epsilon^2)$ for achieving $\mathbb{E}[|Q_k - Q^|\infty] \le \epsilon$, matching that of off-policy Q-learning but with a worse dependence on exploration-related parameters. We also derive an explicit rate for $\mathbb{E}[|Q^{\pi_k} - Q^*|\infty^2]$, where $\pi_k$ is the learning policy at iteration $k$. These results reveal that on-policy Q-learning exhibits weaker exploration than its off-policy counterpart but enjoys an exploitation advantage, as its policy converges to an optimal one rather than remaining fixed. Numerical simulations corroborate our theory. Technically, the combination of time-varying learning policies (which induce rapidly time-inhomogeneous Markovian noise) and the minimal assumption on exploration presents significant analytical challenges. To address these challenges, we employ a refined approach that leverages the Poisson equation to decompose the Markovian noise corresponding to the lazy transition matrix into a martingale-difference term and residual terms. To control the residual terms under time inhomogeneity, we perform a sensitivity analysis of the Poisson equation solution with respect to both the Q-function estimate and the learning policy. These tools may further facilitate the analysis of general reinforcement learning algorithms with rapidly time-varying learning policies – such as single-timescale actor–critic methods and learning-in-games algorithms – and are of independent interest.

[656] Expert Merging in Sparse Mixture of Experts with Nash Bargaining

Dung V. Nguyen, Anh T. Nguyen, Minh H. Nguyen, Luc Q. Nguyen, Shiqi Jiang, Ethan Fetaya, Linh Duy Tran, Gal Chechik, Tan M. Nguyen

Main category: cs.LG

TL;DR: NAMEx is a novel expert merging framework that uses Nash Bargaining from game theory to enable balanced collaboration among experts in Sparse Mixture of Experts models, outperforming existing methods across multiple tasks.

DetailsMotivation: Existing expert merging strategies for SMoE lack principled weighting mechanisms and fail to capture the cooperative and competitive dynamics among experts.

Method: NAMEx incorporates Nash Bargaining into the merging process and uses complex momentum to accelerate expert propagation with theoretical convergence guarantees.

Result: NAMEx consistently outperforms competing methods in language modeling, text classification, image classification, and zero-shot robustness under data corruption, and scales effectively to large models like Qwen1.5-MoE (14B) and DeepSeek-MoE (16B).

Conclusion: NAMEx provides a principled framework for expert merging that enables more balanced and efficient collaboration among experts, demonstrating strong performance across diverse tasks and model scales.

Abstract: Existing expert merging strategies for Sparse Mixture of Experts (SMoE) typically rely on input-dependent or input-independent averaging of expert parameters, but often lack a principled weighting mechanism. In this work, we reinterpret expert merging through the lens of game theory, revealing cooperative and competitive dynamics among experts. Based on this perspective, we introduce Nash Merging of Experts (NAMEx), a novel framework that incorporates Nash Bargaining into the merging process, enabling more balanced and efficient collaboration among experts. Additionally, we incorporate complex momentum into NAMEx to accelerate expert propagation with theoretical guarantees for convergence. Extensive experiments across language modelling, text classification, image classification, and zero-shot robustness under data corruption show that NAMEx consistently outperforms competing methods while integrating seamlessly with popular MoE architectures. Finally, we demonstrate NAMEx’s scalability by applying it to large-scale systems, including Qwen1.5-MoE (14B) and DeepSeek-MoE (16B), where it proves effective in both zero-shot and fine-tuning settings.

[657] Zeroth-Order Sharpness-Aware Learning with Exponential Tilting

Xuchen Gong, Tian Li

Main category: cs.LG

TL;DR: This paper connects zeroth-order optimization with sharpness-aware minimization (SAM) through an exponential tilting objective that bridges average-loss and max-loss formulations, enabling gradient-free flat minima optimization.

DetailsMotivation: To bridge the gap between classic zeroth-order optimization (which optimizes average loss) and SAM approaches (which focus on worst-case loss), providing a smooth transition between these objectives for better generalization.

Method: Proposes an exponential tilting objective parameterized by a tilting parameter t that smoothly transitions between average-loss and max-loss formulations. Develops new zeroth-order algorithms to solve this soft SAM objective, providing gradient-free and memory-efficient alternatives to SAM variants.

Result: Achieves better generalization compared to vanilla zeroth-order baselines across various downstream tasks including classification, multiple choice QA, and language generation. Provides precise characterizations of sharpness notions within the tilted SAM framework.

Conclusion: The proposed exponential tilting framework effectively connects zeroth-order optimization with SAM, offering a practical gradient-free alternative that achieves improved generalization performance while maintaining memory efficiency.

Abstract: Classic zeroth-order optimization approaches typically optimize for a smoothed version of the original function, i.e., the expected objective under randomly perturbed model parameters. This can be interpreted as encouraging the loss values in the perturbation set to be small on average. Popular sharpness-aware minimization (SAM) objectives, however, typically focus on the largest loss within the neighborhood to arrive at flat minima more effectively. In this work, we connect zeroth-order optimization (and its corresponding objectives) with SAM approaches explicitly, through an exponential tilting objective that provides a smooth transition between the average- and the max-loss formulations. We explore new zeroth-order algorithms to solve a soft SAM objective parameterized by a tilting parameter $t$. We provide precise characterizations of the sharpness notions of the tilted SAM framework. Practically, our approach can be used as a gradient-free and memory-efficient alternative to SAM variants, and it achieves better generalization compared to vanilla zeroth-order baselines on a wide range of downstream tasks, including classification, multiple choice QA, and language generation.

[658] Still Competitive: Revisiting Recurrent Models for Irregular Time Series Prediction

Ankitkumar Joshi, Milos Hauskrecht

Main category: cs.LG

TL;DR: GRUwE is a modified GRU model with exponential basis functions that handles irregularly sampled multivariate time series, achieving competitive performance with simpler architecture and lower computational overhead.

DetailsMotivation: To determine if simpler RNN-based approaches can compete with complex architectures for irregular time series prediction, and to provide an efficient solution for domains like healthcare and sensor networks.

Method: Extends GRU with two reset mechanisms: observation-triggered and time-triggered resets using learnable exponential decays, maintaining Markov state representation that updates with irregular observations.

Result: GRUwE achieves competitive to superior performance compared to state-of-the-art methods on next-observation and next-event prediction tasks across real-world benchmarks.

Conclusion: Simple RNN-based approaches with clever modifications like GRUwE remain competitive with complex architectures while offering implementation simplicity, minimal hyper-parameter tuning, and reduced computational overhead.

Abstract: Modeling irregularly sampled multivariate time series is a persistent challenge in domains like healthcare and sensor networks. While recent works have explored a variety of complex learning architectures to solve the prediction problems for irregularly sampled time series, it remains unclear what are the true benefits of some of these architectures, and whether clever modifications of simpler and more efficient RNN-based algorithms are still competitive, i.e. they are on par with or even superior to these methods. In this work, we propose and study GRUwE: Gated Recurrent Unit with Exponential basis functions, that builds upon RNN-based architectures for observations made at irregular times. GRUwE supports both regression-based and event-based predictions in continuous time. GRUwE works by maintaining a Markov state representation of the time series that updates with the arrival of irregular observations. The Markov state update relies on two reset mechanisms: (i) observation-triggered reset, and (ii) time-triggered reset of the GRU state using learnable exponential decays, to support the predictions in continuous time. Our empirical evaluations across several real-world benchmarks on next-observation and next-event prediction tasks demonstrate that GRUwE can indeed achieve competitive to superior performance compared to the recent state-of-the-art (SOTA) methods. Thanks to its simplicity, GRUwE offers compelling advantages: it is easy to implement, requires minimal hyper-parameter tuning efforts, and significantly reduces the computational overhead in the online deployment.

[659] AtomBench: A Benchmark for Generative Atomic Structure Models using GPT, Diffusion, and Flow Architectures

Charles Rhys Campbell, Aldo H. Romero, Kamal Choudhary

Main category: cs.LG

TL;DR: Systematic benchmark of three generative models (AtomGPT, CDVAE, FlowMM) for crystal structure generation using superconductivity datasets, with CDVAE performing best.

DetailsMotivation: Lack of rigorous comparative evaluation of diverse generative model architectures for materials discovery, despite their increasing adoption.

Method: Trained three models (transformer-based AtomGPT, CDVAE, FlowMM) on superconductivity datasets and evaluated using KL divergence and MAE of lattice parameters.

Result: CDVAE performed most favorably, followed by AtomGPT, then FlowMM based on KLD and MAE scores.

Conclusion: CDVAE shows superior performance in crystal structure generation, and all benchmarking code will be publicly available.

Abstract: Generative models have become significant assets in the exploration and identification of new materials, enabling the rapid proposal of candidate crystal structures that satisfy target properties. Despite the increasing adoption of diverse architectures, a rigorous comparative evaluation of their performance on materials datasets is lacking. In this work, we present a systematic benchmark of three representative generative models- AtomGPT (a transformer-based model), Crystal Diffusion Variational Autoencoder (CDVAE), and FlowMM (a Riemannian flow matching model). These models were trained to reconstruct crystal structures from subsets of two publicly available superconductivity datasets- JARVIS Supercon 3D and DS A/B from the Alexandria database. Performance was assessed using the Kullback-Leibler (KL) divergence between predicted and reference distributions of lattice parameters, as well as the mean absolute error (MAE) of individual lattice constants. For the computed KLD and MAE scores, CDVAE performs most favorably, followed by AtomGPT, and then FlowMM. All benchmarking code and model configurations will be made publicly available at https://github.com/atomgptlab/atombench_inverse.

[660] Alignment is Localized: A Causal Probe into Preference Layers

Archie Chaudhury

Main category: cs.LG

TL;DR: This paper analyzes how reinforcement learning with human feedback (RLHF) aligns language models by examining layer activations, finding that alignment is spatially localized in mid-layers rather than diffusely distributed.

DetailsMotivation: To understand the opaque internal workings of how RLHF achieves language model alignment, as current methods remain largely unexplained despite their popularity for safety and intent alignment.

Method: Applied layer-wide causal patching between base and tuned models across human preference pairs using Llama-3.2-1B, and used LASSO regression to identify layers with non-zero coefficients linking activation distances to reward gains.

Result: Found that alignment is spatially localized: mid-layer activations encode a distinct subspace that causally determines reward-consistent behavior, while early and late layers remain largely unaffected. Only a small number of layers have non-zero coefficients.

Conclusion: Alignment from human-based preferential tuning is a directional, low-rank process rather than diffuse and parametric, at least for some language models.

Abstract: Reinforcement Learning frameworks, particularly those utilizing human annotations, have become an increasingly popular method for preference fine-tuning, where the outputs of a language model are tuned to match a certain set of behavioral policies or guidelines. Reinforcement Learning through Human Feedback (RLHF) is perhaps the most popular implementation of such a framework, particularly for aligning LMs toward safety and human intent. However, the internal workings of how such alignment is achieved remain largely opaque. In this work, we systematically analyze preference optimization for language model alignment by applying layer-wide causal patching between a base model and its tuned counterpart across human preference pairs. We implement our methodology on \textit{Llama-3.2-1B}, and find that alignment is spatially localized: mid-layer activations encode a distinct subspace that causally determines reward-consistent behavior, while early and late layers remain largely unaffected. Utilizing LASSO regression, we also find that only a small number of layers possess non-zero coefficients linking activation distances to reward gains. Overall, we show that, at least for some language models, alignment from human-based, preferential tuning is a directional, low rank process, rather than diffuse and parameteric.

[661] Bridging Symmetry and Robustness: On the Role of Equivariance in Enhancing Adversarial Robustness

Longwei Wang, Ifrat Ikhtear Uddin, KC Santosh, Chaowei Zhang, Xiao Qin, Yang Zhou

Main category: cs.LG

TL;DR: This paper proposes using group-equivariant convolutions (rotation- and scale-equivariant layers) in CNNs to improve adversarial robustness without adversarial training, showing both theoretical and empirical benefits.

DetailsMotivation: Adversarial training is computationally expensive and can reduce clean-data accuracy. The authors explore architectural solutions using symmetry priors to build more robust models.

Method: Two symmetry-aware architectures: parallel design (independent processing of standard and equivariant features with fusion) and cascaded design (sequential equivariant operations). Uses group-equivariant convolutions to encode rotation and scale symmetries.

Result: Models consistently improve adversarial robustness and generalization on CIFAR-10, CIFAR-100, and CIFAR-10C under FGSM and PGD attacks without adversarial training. Theoretically reduces hypothesis space complexity and provides tighter certified robustness bounds.

Conclusion: Symmetry-enforcing architectures offer efficient and principled alternatives to data augmentation-based defenses, demonstrating the potential of architectural approaches to adversarial robustness.

Abstract: Adversarial examples reveal critical vulnerabilities in deep neural networks by exploiting their sensitivity to imperceptible input perturbations. While adversarial training remains the predominant defense strategy, it often incurs significant computational cost and may compromise clean-data accuracy. In this work, we investigate an architectural approach to adversarial robustness by embedding group-equivariant convolutions-specifically, rotation- and scale-equivariant layers-into standard convolutional neural networks (CNNs). These layers encode symmetry priors that align model behavior with structured transformations in the input space, promoting smoother decision boundaries and greater resilience to adversarial attacks. We propose and evaluate two symmetry-aware architectures: a parallel design that processes standard and equivariant features independently before fusion, and a cascaded design that applies equivariant operations sequentially. Theoretically, we demonstrate that such models reduce hypothesis space complexity, regularize gradients, and yield tighter certified robustness bounds under the CLEVER (Cross Lipschitz Extreme Value for nEtwork Robustness) framework. Empirically, our models consistently improve adversarial robustness and generalization across CIFAR-10, CIFAR-100, and CIFAR-10C under both FGSM and PGD attacks, without requiring adversarial training. These findings underscore the potential of symmetry-enforcing architectures as efficient and principled alternatives to data augmentation-based defenses.

[662] The Formalism-Implementation Gap in Reinforcement Learning Research

Pablo Samuel Castro

Main category: cs.LG

TL;DR: The paper argues that RL research should shift focus from demonstrating agent capabilities to understanding learning dynamics, and calls for more precise mapping between benchmarks and mathematical formalisms.

DetailsMotivation: Current RL research prioritizes performance over understanding learning dynamics, risking overfitting on benchmarks and making techniques harder to transfer to real-world problems.

Method: The paper uses the Arcade Learning Environment (ALE) as an example benchmark to demonstrate how it can be repurposed for understanding learning dynamics rather than just performance demonstration.

Result: The analysis shows that even “saturated” benchmarks like ALE can be effectively used for developing scientific understanding and facilitating real-world RL deployment.

Conclusion: RL research needs a paradigm shift toward advancing scientific understanding of learning dynamics and establishing clearer connections between benchmarks and mathematical foundations.

Abstract: The last decade has seen an upswing in interest and adoption of reinforcement learning (RL) techniques, in large part due to its demonstrated capabilities at performing certain tasks at “super-human levels”. This has incentivized the community to prioritize research that demonstrates RL agent performance, often at the expense of research aimed at understanding their learning dynamics. Performance-focused research runs the risk of overfitting on academic benchmarks – thereby rendering them less useful – which can make it difficult to transfer proposed techniques to novel problems. Further, it implicitly diminishes work that does not push the performance-frontier, but aims at improving our understanding of these techniques. This paper argues two points: (i) RL research should stop focusing solely on demonstrating agent capabilities, and focus more on advancing the science and understanding of reinforcement learning; and (ii) we need to be more precise on how our benchmarks map to the underlying mathematical formalisms. We use the popular Arcade Learning Environment (ALE; Bellemare et al., 2013) as an example of a benchmark that, despite being increasingly considered “saturated”, can be effectively used for developing this understanding, and facilitating the deployment of RL techniques in impactful real-world problems.

[663] Expressive Reward Synthesis with the Runtime Monitoring Language

Daniel Donnelly, Angelo Ferrando, Francesco Belardinelli

Main category: cs.LG

TL;DR: This paper introduces language-based Reward Machines using Runtime Monitoring Language (RML) to overcome limitations of traditional Reward Machines, enabling specification of non-regular, non-Markovian reward functions with built-in memory capabilities.

DetailsMotivation: Address reward misspecification in RL by overcoming limitations of traditional black-box reward functions and regular-language-bounded Reward Machines, which cannot capture complex behaviors like counting or parameterized conditions.

Method: Develop language-based Reward Machines using Runtime Monitoring Language (RML), leveraging RML’s built-in memory to specify non-regular, non-Markovian reward functions.

Result: The approach demonstrates expressiveness for non-regular tasks and shows advantages in flexible event-handling and task specification compared to existing Reward Machine methods.

Conclusion: Language-based Reward Machines using RML provide a more expressive framework for specifying complex reward functions in RL, addressing limitations of traditional approaches while maintaining interpretability.

Abstract: A key challenge in reinforcement learning (RL) is reward (mis)specification, whereby imprecisely defined reward functions can result in unintended, possibly harmful, behaviours. Indeed, reward functions in RL are typically treated as black-box mappings from state-action pairs to scalar values. While effective in many settings, this approach provides no information about why rewards are given, which can hinder learning and interpretability. Reward Machines address this issue by representing reward functions as finite state automata, enabling the specification of structured, non-Markovian reward functions. However, their expressivity is typically bounded by regular languages, leaving them unable to capture more complex behaviours such as counting or parametrised conditions. In this work, we build on the Runtime Monitoring Language (RML) to develop a novel class of language-based Reward Machines. By leveraging the built-in memory of RML, our approach can specify reward functions for non-regular, non-Markovian tasks. We demonstrate the expressiveness of our approach through experiments, highlighting additional advantages in flexible event-handling and task specification over existing Reward Machine-based methods.

[664] Human-Allied Relational Reinforcement Learning

Fateme Golivand Darvishvand, Hikaru Shindo, Sahil Sidheekh, Kristian Kersting, Sriraam Natarajan

Main category: cs.LG

TL;DR: A novel framework combining relational reinforcement learning with object-centric representations and active human guidance to handle both structured and unstructured data effectively.

DetailsMotivation: Traditional RL systems ignore inherent problem structure, while relational RL (RRL) makes strong assumptions about structure. There's a need for approaches that can handle both structured and unstructured data while improving learning efficiency.

Method: Combines RRL with object-centric representation learning and incorporates active human guidance by explicitly modeling policy uncertainty to query experts when needed.

Result: Empirical evaluation shows the proposed approach is effective and efficient in learning performance.

Conclusion: The framework successfully bridges the gap between structured and unstructured data handling in RL while leveraging human expertise through active learning mechanisms.

Abstract: Reinforcement learning (RL) has experienced a second wind in the past decade. While incredibly successful in images and videos, these systems still operate within the realm of propositional tasks ignoring the inherent structure that exists in the problem. Consequently, relational extensions (RRL) have been developed for such structured problems that allow for effective generalization to arbitrary number of objects. However, they inherently make strong assumptions about the problem structure. We introduce a novel framework that combines RRL with object-centric representation to handle both structured and unstructured data. We enhance learning by allowing the system to actively query the human expert for guidance by explicitly modeling the uncertainty over the policy. Our empirical evaluation demonstrates the effectiveness and efficiency of our proposed approach.

[665] Explore-then-Commit for Nonstationary Linear Bandits with Latent Dynamics

Sunmook Choi, Yahya Sattar, Yassir Jedra, Maryam Fazel, Sarah Dean

Main category: cs.LG

TL;DR: The paper studies nonstationary bandits with action-dependent linear state dynamics, proposes an explore-then-commit algorithm achieving Õ(T^{2/3}) regret, and addresses challenges in learning from correlated rewards and optimizing long-term action sequences.

DetailsMotivation: To solve bandit problems where rewards depend on both actions and latent states with unknown linear dynamics, creating tension between short-term and long-term rewards due to action-dependent state transitions.

Method: Explore-then-commit algorithm with random Rademacher actions for system identification during exploration, followed by optimized action sequence design using estimated Markov parameters during commitment phase.

Result: The algorithm achieves Õ(T^{2/3}) regret, with analysis providing near-optimal sample complexity for system identification and sub-optimality guarantees for the NP-hard indefinite quadratic optimization problem.

Conclusion: The proposed approach successfully handles temporally correlated rewards and long-term optimization challenges in nonstationary bandits with linear dynamics, with practical implementation using semidefinite relaxation and Goemans-Williamson rounding.

Abstract: We study a nonstationary bandit problem where rewards depend on both actions and latent states, the latter governed by unknown linear dynamics. Crucially, the state dynamics also depend on the actions, resulting in tension between short-term and long-term rewards. We propose an explore-then-commit algorithm for a finite horizon $T$. During the exploration phase, random Rademacher actions enable estimation of the Markov parameters of the linear dynamics, which characterize the action-reward relationship. In the commit phase, the algorithm uses the estimated parameters to design an optimized action sequence for long-term reward. Our proposed algorithm achieves $\tilde{\mathcal{O}}(T^{2/3})$ regret. Our analysis handles two key challenges: learning from temporally correlated rewards, and designing action sequences with optimal long-term reward. We address the first challenge by providing near-optimal sample complexity and error bounds for system identification using bilinear rewards. We address the second challenge by proving an equivalence with indefinite quadratic optimization over a hypercube, a known NP-hard problem. We provide a sub-optimality guarantee for this problem, enabling our regret upper bound. Lastly, we propose a semidefinite relaxation with Goemans-Williamson rounding as a practical approach.

[666] Benchmarking noisy label detection methods

Henrique Pickler, Jorge K. S. Kamassury, Danilo Silva

Main category: cs.LG

TL;DR: Comprehensive benchmark of label noise detection methods decomposed into three components: label agreement function, aggregation method, and information gathering approach, with evaluation across vision and tabular datasets under synthetic and real-world noise conditions.

DetailsMotivation: Label noise is a common problem in real-world datasets affecting model training and validation, but there's no clear consensus on optimal detection approaches despite various proposed techniques.

Method: Decompose detection methods into three fundamental components (label agreement function, aggregation method, information gathering approach) and propose unified benchmark task of detecting training samples equal to dataset’s noise rate with novel false negative rate metric.

Result: In-sample information gathering using average probability aggregation combined with logit margin as label agreement function achieves best results across most scenarios in vision and tabular datasets under both synthetic and real-world noise conditions.

Conclusion: Provides practical guidance for designing new detection methods and selecting techniques for specific applications based on systematic comparison across diverse approaches.

Abstract: Label noise is a common problem in real-world datasets, affecting both model training and validation. Clean data are essential for achieving strong performance and ensuring reliable evaluation. While various techniques have been proposed to detect noisy labels, there is no clear consensus on optimal approaches. We perform a comprehensive benchmark of detection methods by decomposing them into three fundamental components: label agreement function, aggregation method, and information gathering approach (in-sample vs out-of-sample). This decomposition can be applied to many existing detection methods, and enables systematic comparison across diverse approaches. To fairly compare methods, we propose a unified benchmark task, detecting a fraction of training samples equal to the dataset’s noise rate. We also introduce a novel metric: the false negative rate at this fixed operating point. Our evaluation spans vision and tabular datasets under both synthetic and real-world noise conditions. We identify that in-sample information gathering using average probability aggregation combined with the logit margin as the label agreement function achieves the best results across most scenarios. Our findings provide practical guidance for designing new detection methods and selecting techniques for specific applications.

[667] Machine Learning for Climate Policy: Understanding Policy Progression in the European Green Deal

Patricia West, Michelle WL Wan, Alexander Hepburn, Edwin Simpson, Raul Santos-Rodriguez, Jeffrey N Clark

Main category: cs.LG

TL;DR: Machine learning models predict climate policy progression using text and metadata, with ClimateBERT performing best on text alone and BERT excelling with metadata.

DetailsMotivation: To understand how climate policies progress from announcement to adoption and support legislative decision-making through ML analysis.

Method: Used dataset of 165 European Green Deal policies with text and metadata; compared TF-IDF, BERT, and ClimateBERT text representations; added metadata features; applied explainable AI methods.

Result: ClimateBERT achieved best performance on text features alone (RMSE=0.17, R²=0.29); BERT performed best with metadata (RMSE=0.16, R²=0.38); policy wording, political party, and country representation were key factors.

Conclusion: ML tools show strong potential for supporting climate policy analysis and decision-making by predicting policy progression and identifying influential factors.

Abstract: Climate change demands effective legislative action to mitigate its impacts. This study explores the application of machine learning (ML) to understand the progression of climate policy from announcement to adoption, focusing on policies within the European Green Deal. We present a dataset of 165 policies, incorporating text and metadata. We aim to predict a policy’s progression status, and compare text representation methods, including TF-IDF, BERT, and ClimateBERT. Metadata features are included to evaluate the impact on predictive performance. On text features alone, ClimateBERT outperforms other approaches (RMSE = 0.17, R^2 = 0.29), while BERT achieves superior performance with the addition of metadata features (RMSE = 0.16, R^2 = 0.38). Using methods from explainable AI highlights the influence of factors such as policy wording and metadata including political party and country representation. These findings underscore the potential of ML tools in supporting climate policy analysis and decision-making.

[668] One-Bit Quantization for Random Features Models

Danil Akhtiamov, Reza Ghane, Babak Hassibi

Main category: cs.LG

TL;DR: One-bit weight quantization in neural networks incurs no asymptotic loss in generalization error for Random Features models, providing theoretical justification for efficient compression with practical speed benefits.

DetailsMotivation: Address the theoretical gap in understanding one-bit weight compression for neural networks, which is important for efficient inference on resource-constrained devices despite recent computational and memory demands.

Method: Analyze one-bit quantization in the Random Features model, a simplified framework corresponding to neural networks with random representations, and prove asymptotic properties of generalization error.

Result: Proved that quantizing weights of all layers except the last incurs no asymptotic loss in generalization error compared to full precision models. Demonstrated empirical speed improvements on laptop GPUs.

Conclusion: One-bit quantization is theoretically justified for neural network compression and provides practical inference speed benefits, with analysis yielding more general results than previous literature.

Abstract: Recent advances in neural networks have led to significant computational and memory demands, spurring interest in one-bit weight compression to enable efficient inference on resource-constrained devices. However, the theoretical underpinnings of such compression remain poorly understood. We address this gap by analyzing one-bit quantization in the Random Features model, a simplified framework that corresponds to neural networks with random representations. We prove that, asymptotically, quantizing weights of all layers except the last incurs no loss in generalization error, compared to the full precision random features model. Our findings offer theoretical insights into neural network compression. We also demonstrate empirically that one-bit quantization leads to significant inference speed ups for the Random Features models even on a laptop GPU, confirming the practical benefits of our work. Additionally, we provide an asymptotically precise characterization of the generalization error for Random Features with an arbitrary number of layers. To the best of our knowledge, our analysis yields more general results than all previous works in the related literature.

[669] WEBSERV: A Browser-Server Environment for Efficient Training of Reinforcement Learning-based Web Agents at Scale

Yuxuan Lu, Jing Huang, Hui Liu, Jiri Gesi, Yan Han, Shihan Fu, Tianqi Zheng, Dakuo Wang

Main category: cs.LG

TL;DR: WEBSERV is a scalable web agent environment that combines realistic browser interaction with efficient server management, achieving state-of-the-art performance while significantly reducing launch latency and storage requirements.

DetailsMotivation: Existing RL web agent environments have limitations: they provide excessive noisy context, perform actions non-deterministically without waiting for UI/network stability, or cannot scale isolated client-server containers effectively for parallel RL rollouts.

Method: WEBSERV includes: 1) a compact, site-agnostic browser environment that balances context and action complexity, and 2) a scalable RL environment via efficient launching and resetting web-servers to enable scalable training and evaluation.

Result: Achieved state-of-the-art single-prompt success rates on WebArena shopping CMS and Gitlab tasks, while cutting launch latency by ~5x, storage need by ~240x, with comparable memory footprint, enabling 200+ concurrent containers on a single host.

Conclusion: WEBSERV provides a scalable and efficient environment for RL web agent training and evaluation, overcoming limitations of existing approaches through its compact browser environment and efficient server management system.

Abstract: Training and evaluation of Reinforcement Learning (RL) web agents have gained increasing attention, yet a scalable and efficient environment that couples realistic and robust browser-side interaction with controllable server-side state at scale is still missing. Existing environments tend to have one or more of the following issues: they overwhelm policy models with excessive and noisy context; they perform actions non-deterministically without waiting for the UI or network to stabilize; or they cannot scale isolated client-server containers effectively for parallel RL rollouts. We propose WEBSERV, an environment that includes 1) a compact, site-agnostic browser environment that balances context and action complexity, and 2) a scalable RL environment via efficient launching and resetting web-servers to enable scalable RL training and evaluation. We evaluate WEBSERV on the shopping CMS and Gitlab tasks in WebArena, achieving state-of-the-art single-prompt success rates while cutting launch latency by ~5x and storage need by ~240x, with a comparable memory footprint, enabling 200+ concurrent containers on a single host.

[670] Protein Folding with Neural Ordinary Differential Equations

Arielle Sanford, Shuo Sun, Christian B. Mendl

Main category: cs.LG

TL;DR: A continuous-depth Evoformer using Neural ODEs replaces AlphaFold’s 48 discrete blocks, achieving constant memory cost and faster training while maintaining structural prediction capability.

DetailsMotivation: To address the high computational costs and rigid layerwise discretization of AlphaFold's 48-block Evoformer architecture.

Method: Replace discrete Evoformer blocks with Neural ODE parameterization that preserves attention-based operations, using adjoint method for constant memory and adaptive ODE solvers for runtime-accuracy trade-off.

Result: Produces structurally plausible predictions and captures secondary structure elements like alpha-helices, though with slightly lower accuracy than original AlphaFold, but achieves this with dramatically fewer resources (17.5 hours on single GPU).

Conclusion: Continuous-depth models offer a promising lightweight and interpretable alternative for efficient and adaptive protein structure prediction.

Abstract: Recent advances in protein structure prediction, such as AlphaFold, have demonstrated the power of deep neural architectures like the Evoformer for capturing complex spatial and evolutionary constraints on protein conformation. However, the depth of the Evoformer, comprising 48 stacked blocks, introduces high computational costs and rigid layerwise discretization. Inspired by Neural Ordinary Differential Equations (Neural ODEs), we propose a continuous-depth formulation of the Evoformer, replacing its 48 discrete blocks with a Neural ODE parameterization that preserves its core attention-based operations. This continuous-time Evoformer achieves constant memory cost (in depth) via the adjoint method, while allowing a principled trade-off between runtime and accuracy through adaptive ODE solvers. Benchmarking on protein structure prediction tasks, we find that the Neural ODE-based Evoformer produces structurally plausible predictions and reliably captures certain secondary structure elements, such as alpha-helices, though it does not fully replicate the accuracy of the original architecture. However, our model achieves this performance using dramatically fewer resources, just 17.5 hours of training on a single GPU, highlighting the promise of continuous-depth models as a lightweight and interpretable alternative for biomolecular modeling. This work opens new directions for efficient and adaptive protein structure prediction frameworks.

[671] Disentangling Hyperedges through the Lens of Category Theory

Yoonho Lee, Junseok Lee, Sangwoo Seo, Sungwon Kim, Yeongmin Kim, Chanyoung Park

Main category: cs.LG

TL;DR: This paper proposes a novel criterion for hyperedge disentanglement in hypergraph neural networks using category theory, and demonstrates its effectiveness in capturing gene functional relations in genetic pathways.

DetailsMotivation: Few studies have explored disentangled representation learning for hypergraph-structured data, despite its potential to reveal hidden hyperedge semantics and unannotated node relations associated with labels.

Method: The paper analyzes hyperedge disentanglement from a category-theoretical perspective and proposes a novel disentanglement criterion derived from the naturality condition. A proof-of-concept model was developed to implement this approach.

Result: The experimental results showed that the proposed criterion successfully captured functional relations of genes (nodes) in genetic pathways (hyperedges), demonstrating the potential of the approach.

Conclusion: The category-theoretical framework provides a novel and effective criterion for hyperedge disentanglement, enabling hypergraph neural networks to leverage hidden hyperedge semantics for improved representation learning.

Abstract: Despite the promising results of disentangled representation learning in discovering latent patterns in graph-structured data, few studies have explored disentanglement for hypergraph-structured data. Integrating hyperedge disentanglement into hypergraph neural networks enables models to leverage hidden hyperedge semantics, such as unannotated relations between nodes, that are associated with labels. This paper presents an analysis of hyperedge disentanglement from a category-theoretical perspective and proposes a novel criterion for disentanglement derived from the naturality condition. Our proof-of-concept model experimentally showed the potential of the proposed criterion by successfully capturing functional relations of genes (nodes) in genetic pathways (hyperedges).

[672] QSVD: Efficient Low-rank Approximation for Unified Query-Key-Value Weight Compression in Low-Precision Vision-Language Models

Yutong Wang, Haiyu Wang, Sai Qian Zhang

Main category: cs.LG

TL;DR: Proposes QSVD method combining SVD compression and quantization to reduce Vision-Language Models’ computational cost while maintaining accuracy.

DetailsMotivation: Vision-Language Models have high computational costs that limit scalability and real-time deployment on resource-constrained devices.

Method: Uses SVD on QKV weight matrices with dynamic rank allocation, plus quantization of weights and activations.

Result: Achieves >10% accuracy improvement over quantization-only or SVD-only methods while consuming less hardware cost.

Conclusion: QSVD enables efficient VLM deployment on resource-constrained devices for real-time applications.

Abstract: Vision-Language Models (VLMs) are integral to tasks such as image captioning and visual question answering, but their high computational cost, driven by large memory footprints and processing time, limits their scalability and real-time applicability. In this work, we propose leveraging Singular-Value Decomposition (SVD) over the joint query (Q), key (K), and value (V) weight matrices to reduce KV cache size and computational overhead. We in addition introduce an efficient rank allocation strategy that dynamically adjusts the SVD rank based on its impact on VLM accuracy, achieving a significant reduction in both memory usage and computational cost. Finally, we extend this approach by applying quantization to both VLM weights and activations, resulting in a highly efficient VLM. Our method outperforms previous approaches that rely solely on quantization or SVD by achieving more than $10%$ accuracy improvement while consuming less hardware cost, making it better for real-time deployment on resource-constrained devices. We open source our code at \href{https://github.com/SAI-Lab-NYU/QSVD}{\texttt{https://github.com/SAI-Lab-NYU/QSVD}}.

[673] Scaffold-Aware Generative Augmentation and Reranking for Enhanced Virtual Screening

Xin Wang, Yu Wang, Yunchao Liu, Jens Meiler, Tyler Derr

Main category: cs.LG

TL;DR: ScaffAug is a scaffold-aware virtual screening framework that addresses class imbalance, structural imbalance, and diversity needs in drug discovery through generative data augmentation, self-training, and scaffold-diversity reranking.

DetailsMotivation: Virtual screening faces three major challenges: class imbalance (low active rate), structural imbalance (certain scaffolds dominate), and the need to identify structurally diverse active compounds for novel drug development.

Method: Three-module framework: 1) Augmentation module using graph diffusion model to generate synthetic data conditioned on scaffolds, 2) Model-agnostic self-training to integrate synthetic and original data, 3) Reranking module to enhance scaffold diversity in top recommendations.

Result: Comprehensive experiments across five target classes show ScaffAug outperforms baseline methods on multiple evaluation metrics while maintaining and enhancing performance in identifying novel active compounds.

Conclusion: ScaffAug introduces novel perspectives for enhancing virtual screening through generative augmentations, reranking, and scaffold-awareness, effectively addressing key challenges in drug discovery.

Abstract: Ligand-based virtual screening (VS) is an essential step in drug discovery that evaluates large chemical libraries to identify compounds that potentially bind to a therapeutic target. However, VS faces three major challenges: class imbalance due to the low active rate, structural imbalance among active molecules where certain scaffolds dominate, and the need to identify structurally diverse active compounds for novel drug development. We introduce ScaffAug, a scaffold-aware VS framework that addresses these challenges through three modules. The augmentation module first generates synthetic data conditioned on scaffolds of actual hits using generative AI, specifically a graph diffusion model. This helps mitigate the class imbalance and furthermore the structural imbalance, due to our proposed scaffold-aware sampling algorithm, designed to produce more samples for active molecules with underrepresented scaffolds. A model-agnostic self-training module is then used to safely integrate the generated synthetic data from our augmentation module with the original labeled data. Lastly, we introduce a reranking module that improves VS by enhancing scaffold diversity in the top recommended set of molecules, while still maintaining and even enhancing the overall general performance of identifying novel, active compounds. We conduct comprehensive computational experiments across five target classes, comparing ScaffAug against existing baseline methods by reporting the performance of multiple evaluation metrics and performing ablation studies on ScaffAug. Overall, this work introduces novel perspectives on effectively enhancing VS by leveraging generative augmentations, reranking, and general scaffold-awareness.

[674] Toward General Digraph Contrastive Learning: A Dual Spatial Perspective

Daohan Su, Yang Zhang, Xunkai Li, Rong-Hua Li, Guoren Wang

Main category: cs.LG

TL;DR: S2-DiGCL is a novel directed graph contrastive learning framework that leverages complex-domain magnetic Laplacian perturbations and real-domain path-based subgraph augmentation to capture directional information, achieving state-of-the-art performance.

DetailsMotivation: Existing graph contrastive learning methods focus on undirected graphs and ignore directional information, which is crucial in real-world networks like social networks and recommendation systems.

Method: Uses two complementary spatial views: complex-domain perspective with personalized perturbations in magnetic Laplacian to modulate edge phases, and real-domain perspective with path-based subgraph augmentation to capture local asymmetries.

Result: Achieves 4.41% improvement in node classification and 4.34% in link prediction on 7 real-world digraph datasets, demonstrating SOTA performance in both supervised and unsupervised settings.

Conclusion: S2-DiGCL effectively captures directional semantics through complementary spatial views, providing more general and robust digraph contrastive learning.

Abstract: Graph Contrastive Learning (GCL) has emerged as a powerful tool for extracting consistent representations from graphs, independent of labeled information. However, existing methods predominantly focus on undirected graphs, disregarding the pivotal directional information that is fundamental and indispensable in real-world networks (e.g., social networks and recommendations).In this paper, we introduce S2-DiGCL, a novel framework that emphasizes spatial insights from complex and real domain perspectives for directed graph (digraph) contrastive learning. From the complex-domain perspective, S2-DiGCL introduces personalized perturbations into the magnetic Laplacian to adaptively modulate edge phases and directional semantics. From the real-domain perspective, it employs a path-based subgraph augmentation strategy to capture fine-grained local asymmetries and topological dependencies. By jointly leveraging these two complementary spatial views, S2-DiGCL constructs high-quality positive and negative samples, leading to more general and robust digraph contrastive learning. Extensive experiments on 7 real-world digraph datasets demonstrate the superiority of our approach, achieving SOTA performance with 4.41% improvement in node classification and 4.34% in link prediction under both supervised and unsupervised settings.

[675] Memorizing Long-tail Data Can Help Generalization Through Composition

Mo Zhou, Haoyang Ma, Rong Ge

Main category: cs.LG

TL;DR: Memorization combined with composition helps models make correct predictions on rare test examples with unseen combinations of long-tailed features.

DetailsMotivation: To explore the synergy between memorization and composition in deep learning, particularly how they help generalization on rare examples with unseen feature combinations.

Method: Theoretical analysis in linear settings and experiments with neural networks on simple data to test composition capabilities across different architectures.

Result: Memorization with composition enables correct predictions on rare test examples with unseen long-tailed feature combinations, and composition capability varies with model architecture.

Conclusion: The combination of memorization and composition enhances generalization on rare examples, with architecture playing a key role in composition effectiveness.

Abstract: Deep learning has led researchers to rethink the relationship between memorization and generalization. In many settings, memorization does not hurt generalization due to implicit regularization and may help by memorizing long-tailed examples. In this paper, we consider the synergy between memorization and simple composition – the ability to make correct prediction on a combination of long-tailed features. Theoretically, we show that for a linear setting, memorization together with composition can help the model make correct predictions on rare test examples that require a combination of long-tailed features, even if such combinations were never observed in the training data. Experiments on neural network architecture on simple data show that the theoretical insight extends beyond the linear setting, and we further observe that the composition capability of the model depends on its architecture.

[676] MGTS-Net: Exploring Graph-Enhanced Multimodal Fusion for Augmented Time Series Forecasting

Shule Hao, Junpeng Bao, Wenli Li

Main category: cs.LG

TL;DR: MGTS-Net is a multimodal graph-enhanced network for time series forecasting that addresses challenges in fine-grained temporal pattern extraction, multimodal integration, and multi-scale feature adaptability through three core components: multimodal feature extraction, feature fusion via heterogeneous graphs, and multi-scale prediction.

DetailsMotivation: Current multimodal time series forecasting methods face limitations in extracting fine-grained temporal patterns, optimally integrating multimodal information, and adapting to dynamic multi-scale features, which constrains their accuracy.

Method: MGTS-Net uses three components: (1) Multimodal Feature Extraction layer that optimizes encoders for temporal, visual, and textual modalities; (2) Multimodal Feature Fusion layer that constructs heterogeneous graphs to model temporal dependencies and cross-modal alignment; (3) Multi-Scale Prediction layer that dynamically weights short-term, medium-term, and long-term predictors.

Result: Extensive experiments show MGTS-Net achieves excellent performance with lightweight and high efficiency, outperforming other state-of-the-art baseline models.

Conclusion: The proposed MGTS-Net methodology demonstrates superior performance in multimodal time series forecasting, validating the effectiveness of its three-component architecture for addressing key challenges in the field.

Abstract: Recent research in time series forecasting has explored integrating multimodal features into models to improve accuracy. However, the accuracy of such methods is constrained by three key challenges: inadequate extraction of fine-grained temporal patterns, suboptimal integration of multimodal information, and limited adaptability to dynamic multi-scale features. To address these problems, we propose MGTS-Net, a Multimodal Graph-enhanced Network for Time Series forecasting. The model consists of three core components: (1) a Multimodal Feature Extraction layer (MFE), which optimizes feature encoders according to the characteristics of temporal, visual, and textual modalities to extract temporal features of fine-grained patterns; (2) a Multimodal Feature Fusion layer (MFF), which constructs a heterogeneous graph to model intra-modal temporal dependencies and cross-modal alignment relationships and dynamically aggregates multimodal knowledge; (3) a Multi-Scale Prediction layer (MSP), which adapts to multi-scale features by dynamically weighting and fusing the outputs of short-term, medium-term, and long-term predictors. Extensive experiments demonstrate that MGTS-Net exhibits excellent performance with light weight and high efficiency. Compared with other state-of-the-art baseline models, our method achieves superior performance, validating the superiority of the proposed methodology.

[677] Sparse Transformer Architectures via Regularized Wasserstein Proximal Operator with $L_1$ Prior

Fuqun Han, Stanley Osher, Wuchen Li

Main category: cs.LG

TL;DR: A sparse transformer architecture that incorporates prior data distribution information through optimal transport theory, improving convexity and sparsity compared to classical flow-based models.

DetailsMotivation: To improve upon classical flow-based models by incorporating prior information about data distributions directly into transformer architecture, addressing limitations in optimization convexity and sample sparsity.

Method: Proposes a sparse transformer architecture motivated by regularized Wasserstein proximal operator from optimal transport theory, which has closed-form solution and represents transformer architectures.

Result: The sparse transformer achieves higher accuracy and faster convergence to target distributions than classical neural ODE-based methods in generative modeling and Bayesian inverse problems.

Conclusion: The proposed sparse transformer architecture effectively incorporates prior distribution information, improves optimization properties, and outperforms traditional neural ODE methods in both theoretical analysis and numerical experiments.

Abstract: In this work, we propose a sparse transformer architecture that incorporates prior information about the underlying data distribution directly into the transformer structure of the neural network. The design of the model is motivated by a special optimal transport problem, namely the regularized Wasserstein proximal operator, which admits a closed-form solution and turns out to be a special representation of transformer architectures. Compared with classical flow-based models, the proposed approach improves the convexity properties of the optimization problem and promotes sparsity in the generated samples. Through both theoretical analysis and numerical experiments, including applications in generative modeling and Bayesian inverse problems, we demonstrate that the sparse transformer achieves higher accuracy and faster convergence to the target distribution than classical neural ODE-based methods.

[678] Modeling Expert Interactions in Sparse Mixture of Experts via Graph Structures

Minh-Khoi Nguyen-Nhat, Rachel S. Y. Teo, Laziz Abdullaev, Maurice Mok, Viet-Hoang Tran, Tan Minh Nguyen

Main category: cs.LG

TL;DR: SymphonySMoE introduces a social graph structure to enhance sparse mixture of experts (SMoE) models, improving robustness against distributional shifts while maintaining efficiency.

DetailsMotivation: Conventional SMoE models struggle with distributional shifts and reduced robustness under data contamination, despite their scalability advantages.

Method: Introduces a social graph to model interactions among experts, enhancing token routing process. The approach is lightweight, modular, and integrates with existing SMoE models like XMoE and Generalist Language Model.

Result: Extensive experiments on language modeling and visual instruction tuning validate effectiveness. Successfully scales to 4.2B and 7.4B parameter models, showing applicability in large-scale fine-tuning tasks.

Conclusion: SymphonySMoE addresses robustness challenges in SMoE while maintaining scalability advantages, with both theoretical analysis and empirical evidence supporting its superiority over baseline SMoE.

Abstract: Sparse Mixture of Experts (SMoE) has emerged as a promising solution to achieving unparalleled scalability in deep learning by decoupling model parameter count from computational cost. By activating only a small subset of parameters per sample, SMoE enables significant growth in model capacity while maintaining efficiency. However, SMoE struggles to adapt to distributional shifts, leading to reduced robustness under data contamination. In this work, we introduce SymphonySMoE, a novel family of SMoE that introduces a social graph to model interactions among experts. This graph-based structure enhances the token routing process, addressing the robustness challenges that are inherent in conventional SMoE designs. SymphonySMoE is lightweight, modular, and integrates seamlessly with existing SMoE-based models such as the XMoE and the Generalist Language Model. We provide both theoretical analysis and empirical evidence demonstrating SymphonySMoE’s advantages over baseline SMoE. Extensive experiments on language modeling and visual instruction tuning validate our method’s effectiveness. We further highlight the scalability of SymphonySMoE to models with 4.2 and 7.4 billion parameters, showcasing its applicability in fine-tuning tasks for large-scale systems.

[679] Colliding with Adversaries at ECML-PKDD 2025 Adversarial Attack Competition 1st Prize Solution

Dimitris Stefanopoulos, Andreas Voskou

Main category: cs.LG

TL;DR: Winning solution for adversarial attack competition using multi-round gradient-based strategy with random initialization and sample-mixing techniques.

DetailsMotivation: To design an adversarial attack that maximizes misclassification while minimizing perturbations for a high energy physics classification model in the ECML-PKDD 2025 competition.

Method: Multi-round gradient-based strategy leveraging the model’s differentiable structure, augmented with random initialization and sample-mixing techniques to enhance attack effectiveness.

Result: Achieved best results in both perturbation size and fooling success rate, securing first place in the competition.

Conclusion: The proposed gradient-based adversarial attack with enhanced techniques proved highly effective for the competition task, demonstrating superior performance in minimizing perturbations while maximizing misclassification.

Abstract: This report presents the winning solution for Task 1 of Colliding with Adversaries: A Challenge on Robust Learning in High Energy Physics Discovery at ECML-PKDD 2025. The task required designing an adversarial attack against a provided classification model that maximizes misclassification while minimizing perturbations. Our approach employs a multi-round gradient-based strategy that leverages the differentiable structure of the model, augmented with random initialization and sample-mixing techniques to enhance effectiveness. The resulting attack achieved the best results in perturbation size and fooling success rate, securing first place in the competition.

[680] Colliding with Adversaries at ECML-PKDD 2025 Model Robustness Competition 1st Prize Solution

Dimitris Stefanopoulos, Andreas Voskou

Main category: cs.LG

TL;DR: Winning solution for ECML-PKDD 2025 challenge that combines adversarial data generation with a robust neural network architecture to achieve 80% mixed accuracy on both clean and adversarial data.

DetailsMotivation: To design a robust ANN-based model that maintains high accuracy on both clean and adversarial data generated by Random Distribution Shuffle Attack (RDSA) in high energy physics discovery.

Method: Two-phase approach: 1) Generate 15 million adversarial training samples using custom RDSA methodology, 2) Train robust architecture with Feature Embedding Block (shared weights for same feature types) and Dense Fusion Tail for final prediction.

Result: Achieved 80% mixed accuracy score, outperforming second-place solution by 2 percentage points.

Conclusion: The combination of adversarial data generation and specialized robust architecture successfully addresses the challenge of maintaining performance on both clean and adversarial data in high energy physics applications.

Abstract: This report presents the winning solution for Task 2 of Colliding with Adversaries: A Challenge on Robust Learning in High Energy Physics Discovery at ECML-PKDD 2025. The goal of the challenge was to design and train a robust ANN-based model capable of achieving high accuracy in a binary classification task on both clean and adversarial data generated with the Random Distribution Shuffle Attack (RDSA). Our solution consists of two components: a data generation phase and a robust model training phase. In the first phase, we produced 15 million artificial training samples using a custom methodology derived from Random Distribution Shuffle Attack (RDSA). In the second phase, we introduced a robust architecture comprising (i)a Feature Embedding Block with shared weights among features of the same type and (ii)a Dense Fusion Tail responsible for the final prediction. Training this architecture on our adversarial dataset achieved a mixed accuracy score of 80%, exceeding the second-place solution by two percentage points.

[681] Input Domain Aware MoE: Decoupling Routing Decisions from Task Optimization in Mixture of Experts

Yongxiang Hua, Haoyu Cao, Zhou Tao, Bocheng Li, Zihao Wu, Chaohu Liu, Linli Xu

Main category: cs.LG

TL;DR: The paper proposes Input Domain Aware MoE, a novel routing framework for sparse Mixture of Experts that uses probabilistic mixture modeling to better partition input space, enabling expert specialization while maintaining balanced utilization.

DetailsMotivation: Existing routing mechanisms in sparse Mixture of Experts struggle to capture input structure effectively, creating a trade-off between expert specialization and balanced computation that hinders scalability and performance.

Method: Proposes a routing framework that leverages probabilistic mixture model to partition input space, modeling routing probabilities as mixture of distributions. The routing mechanism is trained independently of task-specific objectives for stable optimization.

Result: Empirical results on vision-language tasks show the method consistently outperforms existing sMoE approaches, achieving higher task performance and improved expert utilization balance.

Conclusion: The Input Domain Aware MoE framework successfully addresses limitations of conventional routing mechanisms by enabling clear expert specialization boundaries while maintaining balanced utilization through probabilistic mixture modeling.

Abstract: Sparse Mixture of Experts (sMoE) has become a pivotal approach for scaling large vision-language models, offering substantial capacity while maintaining computational efficiency through dynamic, sparse activation of experts. However, existing routing mechanisms, typically based on similarity scoring, struggle to effectively capture the underlying input structure. This limitation leads to a trade-off between expert specialization and balanced computation, hindering both scalability and performance. We propose Input Domain Aware MoE, a novel routing framework that leverages a probabilistic mixture model to better partition the input space. By modeling routing probabilities as a mixture of distributions, our method enables experts to develop clear specialization boundaries while achieving balanced utilization. Unlike conventional approaches, our routing mechanism is trained independently of task-specific objectives, allowing for stable optimization and decisive expert assignments. Empirical results on vision-language tasks demonstrate that our method consistently outperforms existing sMoE approaches, achieving higher task performance and improved expert utilization balance.

[682] Buzz, Choose, Forget: A Meta-Bandit Framework for Bee-Like Decision Making

Emmanuelle Claeys, Elena Kerjean, Jean-Michel Loubes

Main category: cs.LG

TL;DR: A sequential reinforcement learning framework for imitation learning that models heterogeneous cognitive strategies in pollinators, particularly honeybees, addressing limitations of existing methods in capturing diverse learning behaviors and providing interpretability.

DetailsMotivation: To overcome the limitations of state-of-the-art imitation learning methods that fail when expert policies shift across memory windows or deviate from optimality, and to provide interpretable models for biological insight into pollinator decision-making.

Method: Introduces a model that minimizes predictive loss while identifying effective memory horizons consistent with behavioral data, ensures full interpretability, and provides a mathematical framework linking bee policy search with bandit formulations under varying exploration-exploitation dynamics.

Result: The approach successfully captures and forecasts behavior across individuals using distinct cognitive strategies (numerical cues, memory, environmental factors), and provides a novel dataset of 80 tracked bees observed under diverse weather conditions.

Conclusion: The framework sheds new light on learning strategies and memory interplay shaping pollinator decision-making, supports ecological governance by improving insect behavior simulations in agroecosystems, and facilitates research on pollinator cognition.

Abstract: We introduce a sequential reinforcement learning framework for imitation learning designed to model heterogeneous cognitive strategies in pollinators. Focusing on honeybees, our approach leverages trajectory similarity to capture and forecast behavior across individuals that rely on distinct strategies: some exploiting numerical cues, others drawing on memory, or being influenced by environmental factors such as weather. Through empirical evaluation, we show that state-of-the-art imitation learning methods often fail in this setting: when expert policies shift across memory windows or deviate from optimality, these models overlook both fast and slow learning behaviors and cannot faithfully reproduce key decision patterns. Moreover, they offer limited interpretability, hindering biological insight. Our contribution addresses these challenges by (i) introducing a model that minimizes predictive loss while identifying the effective memory horizon most consistent with behavioral data, and (ii) ensuring full interpretability to enable biologists to analyze underlying decision-making strategies and finally (iii) providing a mathematical framework linking bee policy search with bandit formulations under varying exploration-exploitation dynamics, and releasing a novel dataset of 80 tracked bees observed under diverse weather conditions. This benchmark facilitates research on pollinator cognition and supports ecological governance by improving simulations of insect behavior in agroecosystems. Our findings shed new light on the learning strategies and memory interplay shaping pollinator decision-making.

[683] SCALAR: Self-Calibrating Adaptive Latent Attention Representation Learning

Farwa Abbas, Hussain Ahmad, Claudia Szabo

Main category: cs.LG

TL;DR: Proposes a novel adaptive kernel-based attention method to handle high-dimensional heterogeneous data with complex feature interactions, overcoming PLS limitations in modeling non-linear relationships and cross-group dependencies.

DetailsMotivation: Traditional PLS struggles with complex non-linear relationships in high-dimensional correlated data, fails to capture cross-group dependencies, and has static feature weighting that limits adaptability to contextual variations.

Method: Introduces an adaptive kernel-based attention mechanism that processes distinct feature groups separately before integration, enabling capture of both local patterns and global relationships.

Result: Experimental results show substantial improvements in performance metrics compared to state-of-the-art methods across diverse datasets.

Conclusion: The proposed method effectively addresses limitations of traditional approaches by enabling adaptive feature processing and capturing complex relationships in high-dimensional heterogeneous data.

Abstract: High-dimensional, heterogeneous data with complex feature interactions pose significant challenges for traditional predictive modeling approaches. While Projection to Latent Structures (PLS) remains a popular technique, it struggles to model complex non-linear relationships, especially in multivariate systems with high-dimensional correlation structures. This challenge is further compounded by simultaneous interactions across multiple scales, where local processing fails to capture crossgroup dependencies. Additionally, static feature weighting limits adaptability to contextual variations, as it ignores sample-specific relevance. To address these limitations, we propose a novel method that enhances predictive performance through novel architectural innovations. Our architecture introduces an adaptive kernel-based attention mechanism that processes distinct feature groups separately before integration, enabling capture of local patterns while preserving global relationships. Experimental results show substantial improvements in performance metrics, compared to the state-of-the-art methods across diverse datasets.

[684] Structured Temporal Causality for Interpretable Multivariate Time Series Anomaly Detection

Dongchan Cho, Jiho Han, Keumyeong Kang, Minsang Kim, Honggyu Ryu, Namsoon Jung

Main category: cs.LG

TL;DR: OracleAD is a simple unsupervised framework for multivariate time series anomaly detection that uses causal embeddings and a Stable Latent Structure (SLS) to identify anomalies through temporal dynamics modeling and spatial relationship capture.

DetailsMotivation: Real-world multivariate time series anomalies are rare and often unlabeled, and existing methods are complex, detect only fragments of anomalies, and overstate performance.

Method: Encodes each variable’s past sequence into causal embeddings for joint prediction and reconstruction, uses self-attention to project embeddings into shared latent space capturing spatial relationships, aligns embeddings to a Stable Latent Structure (SLS), and employs dual scoring based on prediction error and SLS deviation.

Result: Achieves state-of-the-art results across multiple real-world datasets and evaluation protocols.

Conclusion: OracleAD provides an interpretable framework that effectively detects anomalies while directly pinpointing root-cause variables through embedding-level analysis.

Abstract: Real-world multivariate time series anomalies are rare and often unlabeled. Additionally, prevailing methods rely on increasingly complex architectures tuned to benchmarks, detecting only fragments of anomalous segments and overstating performance. In this paper, we introduce OracleAD, a simple and interpretable unsupervised framework for multivariate time series anomaly detection. OracleAD encodes each variable’s past sequence into a single causal embedding to jointly predict the present time point and reconstruct the input window, effectively modeling temporal dynamics. These embeddings then undergo a self-attention mechanism to project them into a shared latent space and capture spatial relationships. These relationships are not static, since they are modeled by a property that emerges from each variable’s temporal dynamics. The projected embeddings are aligned to a Stable Latent Structure (SLS) representing normal-state relationships. Anomalies are identified using a dual scoring mechanism based on prediction error and deviation from the SLS, enabling fine-grained anomaly diagnosis at each time point and across individual variables. Since any noticeable SLS deviation originates from embeddings that violate the learned temporal causality of normal data, OracleAD directly pinpoints the root-cause variables at the embedding level. OracleAD achieves state-of-the-art results across multiple real-world datasets and evaluation protocols, while remaining interpretable through SLS.

[685] eDCF: Estimating Intrinsic Dimension using Local Connectivity

Dhruv Gupta, Aditya Nagarsekar, Vraj Shah, Sujith Thomas

Main category: cs.LG

TL;DR: The paper introduces eDCF, a scalable and parallelizable method for robust intrinsic dimension estimation that handles scale dependencies and noise better than existing approaches.

DetailsMotivation: Modern datasets have high-dimensional features with complex dependencies, making intrinsic dimension estimation challenging due to scale dependency issues where noise inflates estimates at fine scales and coarser scales give lower values.

Method: The method is called eDCF, based on Connectivity Factor (CF), a local connectivity-based metric that robustly estimates intrinsic dimension across varying scales in a scalable and parallelizable manner.

Result: eDCF consistently matches leading estimators with comparable MAE on synthetic benchmarks, achieves higher exact intrinsic dimension match rates (25.0% vs 16.7% for MLE and 12.5% for TWO-NN), especially under medium to high noise levels and large datasets, and accurately detects fractal geometries in decision boundaries.

Conclusion: The eDCF method provides robust intrinsic dimension estimation across scales, excels in noisy environments, and demonstrates utility for analyzing realistic structured data through fractal geometry detection.

Abstract: Modern datasets often contain high-dimensional features exhibiting complex dependencies. To effectively analyze such data, dimensionality reduction methods rely on estimating the dataset’s intrinsic dimension (id) as a measure of its underlying complexity. However, estimating id is challenging due to its dependence on scale: at very fine scales, noise inflates id estimates, while at coarser scales, estimates stabilize to lower, scale-invariant values. This paper introduces a novel, scalable, and parallelizable method called eDCF, which is based on Connectivity Factor (CF), a local connectivity-based metric, to robustly estimate intrinsic dimension across varying scales. Our method consistently matches leading estimators, achieving comparable values of mean absolute error (MAE) on synthetic benchmarks with noisy samples. Moreover, our approach also attains higher exact intrinsic dimension match rates, reaching up to 25.0% compared to 16.7% for MLE and 12.5% for TWO-NN, particularly excelling under medium to high noise levels and large datasets. Further, we showcase our method’s ability to accurately detect fractal geometries in decision boundaries, confirming its utility for analyzing realistic, structured data.

[686] Realizing LLMs’ Causal Potential Requires Science-Grounded, Novel Benchmarks

Ashutosh Srivastava, Lokesh Nagalapatti, Gautam Jajoo, Aniket Vashishtha, Parameswari Krishnamurthy, Amit Sharma

Main category: cs.LG

TL;DR: LLMs’ apparent success in causal discovery is undermined by dataset leakage in benchmarks. Real performance requires science-grounded evaluation and hybrid methods combining LLM knowledge with statistical approaches.

DetailsMotivation: To challenge claims of LLM superiority in causal discovery and address concerns about memorization from pretraining data, proposing robust evaluation and hybrid methods for real-world scientific use.

Method: Developed science-grounded evaluation using recent publications after LLM training cutoff, and created hybrid methods using LLM predictions as priors for classical PC algorithm.

Result: LLMs perform poorly on curated science-grounded graphs compared to benchmarks like BNLearn, but using LLM predictions as priors significantly improves PC algorithm accuracy over both LLM-only and statistical methods.

Conclusion: Community should adopt science-grounded benchmarks resistant to leakage and invest in hybrid causal discovery methods that combine LLM knowledge with statistical approaches for real-world scientific inquiry.

Abstract: Recent claims of strong performance by Large Language Models (LLMs) on causal discovery are undermined by a key flaw: many evaluations rely on benchmarks likely included in pretraining corpora. Thus, apparent success suggests that LLM-only methods, which ignore observational data, outperform classical statistical approaches. We challenge this narrative by asking: Do LLMs truly reason about causal structure, and how can we measure it without memorization concerns? Can they be trusted for real-world scientific discovery? We argue that realizing LLMs’ potential for causal analysis requires two shifts: (P.1) developing robust evaluation protocols based on recent scientific studies to guard against dataset leakage, and (P.2) designing hybrid methods that combine LLM-derived knowledge with data-driven statistics. To address P.1, we encourage evaluating discovery methods on novel, real-world scientific studies. We outline a practical recipe for extracting causal graphs from recent publications released after an LLM’s training cutoff, ensuring relevance and preventing memorization while capturing both established and novel relations. Compared to benchmarks like BNLearn, where LLMs achieve near-perfect accuracy, they perform far worse on our curated graphs, underscoring the need for statistical grounding. Supporting P.2, we show that using LLM predictions as priors for the classical PC algorithm significantly improves accuracy over both LLM-only and purely statistical methods. We call on the community to adopt science-grounded, leakage-resistant benchmarks and invest in hybrid causal discovery methods suited to real-world inquiry.

[687] Predicting life satisfaction using machine learning and explainable AI

Alif Elham Khan, Mohammad Junayed Hasan, Humayra Anjum, Nabeel Mohammed, Sifat Momen

Main category: cs.LG

TL;DR: Machine learning and LLMs can predict life satisfaction with high accuracy (93.80% and 93.74%) using Danish survey data, identifying health as the key determinant across age groups.

DetailsMotivation: Traditional life satisfaction measurement methods are analog, complicated, and error-prone, raising validation concerns. This study explores ML and LLMs as more reliable alternatives.

Method: Used Danish government survey data (19,000 people), applied feature learning to extract 27 key questions, explored clinical/biomedical LLMs by converting tabular data to natural language, and conducted ablation studies on resampling and feature selection.

Result: Achieved 93.80% accuracy with ML and 93.74% with LLMs, with health condition identified as the most important determinant across all age groups. Biomedical LLMs performed better than clinical ones.

Conclusion: ML, LLMs, and XAI can jointly build trust in using AI to investigate human behavior, with significant implications for quantifying and understanding subjective well-being.

Abstract: Life satisfaction is a crucial facet of human well-being. Hence, research on life satisfaction is incumbent for understanding how individuals experience their lives and influencing interventions targeted at enhancing mental health and well-being. Life satisfaction has traditionally been measured using analog, complicated, and frequently error-prone methods. These methods raise questions concerning validation and propagation. However, this study demonstrates the potential for machine learning algorithms to predict life satisfaction with a high accuracy of 93.80% and a 73.00% macro F1-score. The dataset comes from a government survey of 19000 people aged 16-64 years in Denmark. Using feature learning techniques, 27 significant questions for assessing contentment were extracted, making the study highly reproducible, simple, and easily interpretable. Furthermore, clinical and biomedical large language models (LLMs) were explored for predicting life satisfaction by converting tabular data into natural language sentences through mapping and adding meaningful counterparts, achieving an accuracy of 93.74% and macro F1-score of 73.21%. It was found that life satisfaction prediction is more closely related to the biomedical domain than the clinical domain. Ablation studies were also conducted to understand the impact of data resampling and feature selection techniques on model performance. Moreover, the correlation between primary determinants with different age brackets was analyzed, and it was found that health condition is the most important determinant across all ages. This study demonstrates how machine learning, large language models and XAI can jointly contribute to building trust and understanding in using AI to investigate human behavior, with significant ramifications for academics and professionals working to quantify and comprehend subjective well-being.

[688] NeurIPT: Foundation Model for Neural Interfaces

Zitao Fang, Chenxuan Li, Hongting Zhou, Shuyang Yu, Guodong Du, Ashwaq Qasem, Yang Lu, Jing Li, Junsong Zhang, Sim Kuan Goh

Main category: cs.LG

TL;DR: NeurIPT is a foundation model for EEG data that uses amplitude-aware masking and progressive MoE architecture to handle diverse EEG characteristics, achieving SOTA performance across multiple BCI datasets.

DetailsMotivation: EEG data faces challenges with inter-subject, inter-task, and inter-condition variability, plus diverse electrode configurations. Current foundation models struggle to generalize across these variations in EEG applications.

Method: Uses Amplitude-Aware Masked Pretraining (AAMP) for temporal modeling, Progressive Mixture-of-Experts (PMoE) architecture for diverse temporal characteristics, and 3D electrode coordinates with Intra-Inter Lobe Pooling (IILP) for spatial modeling.

Result: Achieved state-of-the-art performance across eight downstream BCI datasets through fine-tuning, demonstrating broad applicability and robust generalization.

Conclusion: NeurIPT advances foundation models for EEG by providing scalable and generalizable neural information processing that effectively handles EEG’s inherent variability.

Abstract: Electroencephalography (EEG) has wide-ranging applications, from clinical diagnosis to brain-computer interfaces (BCIs). With the increasing volume and variety of EEG data, there has been growing interest in establishing foundation models (FMs) to scale up and generalize neural decoding. Despite showing early potential, applying FMs to EEG remains challenging due to substantial inter-subject, inter-task, and inter-condition variability, as well as diverse electrode configurations across recording setups. To tackle these open challenges, we propose NeurIPT, a foundation model developed for diverse EEG-based Neural Interfaces with a Pre-trained Transformer by capturing both homogeneous and heterogeneous spatio-temporal characteristics inherent in EEG signals. Temporally, we introduce Amplitude-Aware Masked Pretraining (AAMP), masking based on signal amplitude rather than random intervals, to learn robust representations across varying signal intensities beyond local interpolation. Moreover, this temporal representation is enhanced by a Progressive Mixture-of-Experts (PMoE) architecture, where specialized expert subnetworks are progressively introduced at deeper layers, adapting effectively to the diverse temporal characteristics of EEG signals. Spatially, NeurIPT leverages the 3D physical coordinates of electrodes, enabling effective transfer of embedding across varying EEG settings, and develops Intra-Inter Lobe Pooling (IILP) during fine-tuning to efficiently exploit regional brain features. Empirical evaluations across eight downstream BCI datasets, via fine-tuning, demonstrated NeurIPT consistently achieved state-of-the-art performance, highlighting its broad applicability and robust generalization. Our work pushes forward the state of FMs in EEG and offers insights into scalable and generalizable neural information processing systems.

[689] LANPO: Bootstrapping Language and Numerical Feedback for Reinforcement Learning in LLMs

Ang Li, Yifei Wang, Zhihang Yuan, Stefanie Jegelka, Yisen Wang

Main category: cs.LG

TL;DR: LANPO is a reinforcement learning framework that separates language feedback for exploration from numerical rewards for optimization, using historical experiences to improve sample efficiency in LLMs.

DetailsMotivation: Traditional RL in LLMs uses scalar rewards that discard valuable textual rationale from rollouts, forcing models to explore from scratch each time and reducing sample efficiency. There's a tension between using feedback from the same problem (risking memorization) vs different problems (causing behavior collapse).

Method: LANPO builds a dynamic experience pool from past trials with two key principles: Reward-Agnostic Reflection for safe intra-sample self-correction, and Relevant Abstraction to distill generalizable lessons from inter-sample experiences. It cleanly separates language feedback (for exploration) from numerical rewards (for optimization).

Result: Across mathematical reasoning benchmarks, LANPO enables 7B and 14B models to significantly outperform strong baselines trained with GRPO in test accuracy.

Conclusion: LANPO provides a robust method for integrating historical experiences into the LLM RL loop, creating more effective and data-efficient learning agents by properly leveraging both language feedback and numerical rewards.

Abstract: Reinforcement learning in large language models (LLMs) often relies on scalar rewards, a practice that discards valuable textual rationale buried in the rollouts, forcing the model to explore \textit{de novo} with each attempt and hindering sample efficiency. While LLMs can uniquely learn from language feedback provided in-context, naively integrating on-line experiences into RL training presents a paradox: feedback from the same problem risks information leakage and memorization, while feedback from different problems often leads to behavior collapse due to irrelevant context. To resolve this tension, we propose \textbf{Language-And-Numerical Policy Optimization (LANPO)}, a framework that cleanly separates the roles of feedback: language guides exploration, while numerical rewards drive optimization. LANPO builds a dynamic experience pool from past trials and introduces two principles to ensure feedback is effective: \emph{Reward-Agnostic Reflection} for safe intra-sample self-correction and \emph{Relevant Abstraction} to distill generalizable lessons from inter-sample experiences. Across mathematical reasoning benchmarks, LANPO enables 7B and 14B models to significantly outperform strong baselines trained with GRPO in test accuracy. Our work provides a robust method for integrating historical experiences into the LLM RL loop, creating more effective and data-efficient learning agents.

[690] Copy-Augmented Representation for Structure Invariant Template-Free Retrosynthesis

Jiaxi Zhuang, Yu Zhang, Aimin Zhou, Ying Qian

Main category: cs.LG

TL;DR: C-SMILES introduces a novel molecular representation that decomposes SMILES into element-token pairs with special tokens to minimize editing distance between reactants and products, combined with copy-augmented mechanism and SMILES alignment guidance for improved retrosynthesis prediction.

DetailsMotivation: Current template-free methods for retrosynthesis prediction struggle to capture structural invariance in chemical reactions where substantial molecular scaffolds remain unchanged, leading to large search spaces and reduced accuracy.

Method: Proposes C-SMILES representation that decomposes traditional SMILES into element-token pairs with five special tokens, incorporates copy-augmented mechanism to preserve unchanged fragments, and uses SMILES alignment guidance for attention consistency with ground-truth atom mappings.

Result: Achieves 67.2% top-1 accuracy on USPTO-50K and 50.8% on USPTO-FULL datasets, with 99.9% validity in generated molecules.

Conclusion: Establishes a new paradigm for structure-aware molecular generation with direct applications in computational drug discovery, significantly improving retrosynthesis prediction accuracy.

Abstract: Retrosynthesis prediction is fundamental to drug discovery and chemical synthesis, requiring the identification of reactants that can produce a target molecule. Current template-free methods struggle to capture the structural invariance inherent in chemical reactions, where substantial molecular scaffolds remain unchanged, leading to unnecessarily large search spaces and reduced prediction accuracy. We introduce C-SMILES, a novel molecular representation that decomposes traditional SMILES into element-token pairs with five special tokens, effectively minimizing editing distance between reactants and products. Building upon this representation, we incorporate a copy-augmented mechanism that dynamically determines whether to generate new tokens or preserve unchanged molecular fragments from the product. Our approach integrates SMILES alignment guidance to enhance attention consistency with ground-truth atom mappings, enabling more chemically coherent predictions. Comprehensive evaluation on USPTO-50K and large-scale USPTO-FULL datasets demonstrates significant improvements: 67.2% top-1 accuracy on USPTO-50K and 50.8% on USPTO-FULL, with 99.9% validity in generated molecules. This work establishes a new paradigm for structure-aware molecular generation with direct applications in computational drug discovery.

[691] Atom-anchored LLMs speak Chemistry: A Retrosynthesis Demonstration

Alan Kai Hassen, Andrius Bernatavicius, Antonius P. A. Janssen, Mike Preuss, Gerard J. P. van Westen, Djork-Arné Clevert

Main category: cs.LG

TL;DR: A framework for molecular reasoning using LLMs without labeled data, achieving high success rates in retrosynthesis tasks by anchoring chain-of-thought reasoning to molecular structure with atomic identifiers.

DetailsMotivation: Addressing data scarcity in chemistry applications where traditional supervised methods are limited by expensive labeled data requirements.

Method: Uses LLMs with chain-of-thought reasoning anchored to molecular structure via unique atomic identifiers, performing one-shot fragment identification followed by optional few-shot transformation prediction.

Result: Achieved high success rates: ≥90% for reaction sites, ≥40% for named reaction classes, and ≥74% for final reactants across academic benchmarks and drug discovery molecules.

Conclusion: The framework enables LLMs to solve complex chemical tasks without labeled data and provides a method to generate synthetic datasets by mapping chemical knowledge to molecular structure.

Abstract: Applications of machine learning in chemistry are often limited by the scarcity and expense of labeled data, restricting traditional supervised methods. In this work, we introduce a framework for molecular reasoning using general-purpose Large Language Models (LLMs) that operates without requiring labeled training data. Our method anchors chain-of-thought reasoning to the molecular structure by using unique atomic identifiers. First, the LLM performs a one-shot task to identify relevant fragments and their associated chemical labels or transformation classes. In an optional second step, this position-aware information is used in a few-shot task with provided class examples to predict the chemical transformation. We apply our framework to single-step retrosynthesis, a task where LLMs have previously underperformed. Across academic benchmarks and expert-validated drug discovery molecules, our work enables LLMs to achieve high success rates in identifying chemically plausible reaction sites ($\geq90%$), named reaction classes ($\geq40%$), and final reactants ($\geq74%$). Beyond solving complex chemical tasks, our work also provides a method to generate theoretically grounded synthetic datasets by mapping chemical knowledge onto the molecular structure and thereby addressing data scarcity.

[692] Symmetry and Generalisation in Neural Approximations of Renormalisation Transformations

Cassidy Ashworth, Pietro Liò, Francesco Caso

Main category: cs.LG

TL;DR: The paper investigates how parameter symmetry and network expressivity affect neural networks’ generalization when learning real-space renormalization group transformations, using the central limit theorem as a test case.

DetailsMotivation: To understand the role of physical symmetries in deep learning models and evaluate the principle of parameter symmetry breaking and restoration in hierarchical learning dynamics.

Method: Used multilayer perceptrons (MLPs) and graph neural networks (GNNs) with varied weight symmetries and activation functions, and analytically analyzed constrained MLP architectures by recasting the central limit theorem as cumulant recursion relations.

Result: Found a competition between symmetry constraints and expressivity, with overly complex or overconstrained models generalizing poorly. Demonstrated poor generalization analytically for constrained MLPs and empirically validated extension to GNNs.

Conclusion: The findings provide new insights into symmetric networks’ learning dynamics and their limitations in modeling structured physical transformations, highlighting the trade-off between symmetry constraints and model expressivity.

Abstract: Deep learning models have proven enormously successful at using multiple layers of representation to learn relevant features of structured data. Encoding physical symmetries into these models can improve performance on difficult tasks, and recent work has motivated the principle of parameter symmetry breaking and restoration as a unifying mechanism underlying their hierarchical learning dynamics. We evaluate the role of parameter symmetry and network expressivity in the generalisation behaviour of neural networks when learning a real-space renormalisation group (RG) transformation, using the central limit theorem (CLT) as a test case map. We consider simple multilayer perceptrons (MLPs) and graph neural networks (GNNs), and vary weight symmetries and activation functions across architectures. Our results reveal a competition between symmetry constraints and expressivity, with overly complex or overconstrained models generalising poorly. We analytically demonstrate this poor generalisation behaviour for certain constrained MLP architectures by recasting the CLT as a cumulant recursion relation and making use of an established framework to propagate cumulants through MLPs. We also empirically validate an extension of this framework from MLPs to GNNs, elucidating the internal information processing performed by these more complex models. These findings offer new insight into the learning dynamics of symmetric networks and their limitations in modelling structured physical transformations.

[693] Asymptotically Stable Quaternion-valued Hopfield-structured Neural Network with Periodic Projection-based Supervised Learning Rules

Tianwei Wang, Xinhui Ma, Wei Pang

Main category: cs.LG

TL;DR: A quaternion-valued supervised learning Hopfield neural network (QSHNN) is proposed, leveraging quaternions’ geometric advantages for rotation representation. The model extends classic HNNs to quaternionic domain with proven stability and uses periodic projection to maintain quaternionic structure during training.

DetailsMotivation: To exploit the geometric advantages of quaternions in representing rotations and postures, particularly for applications like robotic control systems where joint postures are naturally parameterized by quaternions.

Method: Extends continuous-time Hopfield neural networks to quaternionic domain, establishes existence and uniqueness of fixed points with asymptotic stability, and introduces periodic projection strategy to maintain quaternionic structure during gradient descent training.

Result: The model achieves high accuracy, fast convergence, and strong reliability across randomly generated target sets. Evolution trajectories exhibit well-bounded curvature and sufficient smoothness.

Conclusion: QSHNN provides a practical implementation framework and general mathematical methodology for designing neural networks under hypercomplex or non-commutative algebraic structures, with particular relevance to robotic control and path planning applications.

Abstract: Motivated by the geometric advantages of quaternions in representing rotations and postures, we propose a quaternion-valued supervised learning Hopfield-structured neural network (QSHNN) with a fully connected structure inspired by the classic Hopfield neural network (HNN). Starting from a continuous-time dynamical model of HNNs, we extend the formulation to the quaternionic domain and establish the existence and uniqueness of fixed points with asymptotic stability. For the learning rules, we introduce a periodic projection strategy that modifies standard gradient descent by periodically projecting each 4*4 block of the weight matrix onto the closest quaternionic structure in the least-squares sense. This approach preserves both convergence and quaternionic consistency throughout training. Benefiting from this rigorous mathematical foundation, the experimental model implementation achieves high accuracy, fast convergence, and strong reliability across randomly generated target sets. Moreover, the evolution trajectories of the QSHNN exhibit well-bounded curvature, i.e., sufficient smoothness, which is crucial for applications such as control systems or path planning modules in robotic arms, where joint postures are parameterized by quaternion neurons. Beyond these application scenarios, the proposed model offers a practical implementation framework and a general mathematical methodology for designing neural networks under hypercomplex or non-commutative algebraic structures.

[694] Prior Makes It Possible: From Sublinear Graph Algorithms to LLM Test-Time Methods

Avrim Blum, Daniel Hsu, Cyrus Rashtchian, Donya Saless

Main category: cs.LG

TL;DR: The paper analyzes the theoretical relationship between a model’s parametric knowledge and test-time augmentation (like RAG), showing a phase transition where augmentation efficiency depends on whether the knowledge graph forms a giant connected component.

DetailsMotivation: To understand the theoretical foundations of how much pre-training knowledge is needed for efficient multi-step reasoning with augmentation, as current understanding of the interplay between parametric knowledge and external retrieval is limited.

Method: Formulates multi-step reasoning as an s-t connectivity problem on knowledge graphs, representing pre-training knowledge as a partial/noisy subgraph and augmentation as querying an oracle for true edges.

Result: Shows a phase transition: when prior knowledge is disconnected into small components, augmentation requires Ω(√n) queries, but once density surpasses a threshold forming a giant component, paths can be found with constant queries.

Conclusion: The efficiency of test-time augmentation critically depends on the connectivity structure of the model’s prior knowledge, with a sharp transition from inefficient to efficient augmentation based on knowledge graph density.

Abstract: Test-time augmentation, such as Retrieval-Augmented Generation (RAG) or tool use, critically depends on an interplay between a model’s parametric knowledge and externally retrieved information. However, the theoretical underpinnings of this relationship remain poorly understood. Specifically, it is not clear how much pre-training knowledge is required to answer queries with a small number of augmentation steps, which is a desirable property in practice. To address this question, we formulate multi-step reasoning as an $s$-$t$ connectivity problem on a knowledge graph. We represent a model’s pre-training parametric knowledge as a partial, potentially noisy subgraph. We view augmentation as querying an oracle for true edges that augment the model’s knowledge. Then, we characterize the necessary and sufficient number of augmentation steps for the model to generate an accurate answer given partial prior knowledge. One key result shows a phase transition: if the prior knowledge graph over $n$ vertices is disconnected into small components, then finding a path via augmentation is inefficient and requires $\Omega(\sqrt{n})$ queries. On the other hand, once the density of correct knowledge surpasses a threshold, forming a giant component, we can find paths with an expected constant number of queries.

[695] On the Impossibility of Retrain Equivalence in Machine Unlearning

Jiatong Yu, Yinghui He, Anirudh Goyal, Sanjeev Arora

Main category: cs.LG

TL;DR: Machine unlearning faces fundamental barriers in multi-stage training pipelines, as local unlearning methods cannot universally achieve Retrain Equivalence due to path dependency - the order of training stages affects unlearning outcomes.

DetailsMotivation: Modern ML pipelines involve multi-stage training (e.g., LLM fine-tuning for alignment, reasoning), but current machine unlearning theory was formulated for i.i.d. data batches. This creates a gap between theory and practice.

Method: Theoretical analysis and empirical experiments on LLMs (Llama and Qwen models 1B-14B) using gradient ascent, NPO, and SimNPO local unlearning algorithms across different training stage orderings.

Result: Models fine-tuned via different orderings of identical training stages diverge during unlearning, with GSM8K accuracy degradation varying by over 20% across paths. Some paths produce models that unlearn slowly, and probability mass distribution during unlearning is path-dependent.

Conclusion: Retrain Equivalence is ill-posed for local unlearning algorithms when models are trained in stages. When training histories are unavailable, the definition and desiderata of machine unlearning need rethinking.

Abstract: Machine unlearning seeks to selectively remove the “influence” of specific training data on a model’s outputs. The ideal goal is Retrain Equivalence–behavior identical to a model trained from scratch on only the retained data. This goal was formulated for models trained on i.i.d. data batches, but modern pipelines often involve multi-stage training, with each stage having a distinct data distribution and objective. Examples include LLM fine-tuning for alignment, reasoning ability, etc. Our study shows via theory and experiments that this shift to multi-stage training introduces a fundamental barrier for machine unlearning. The theory indicates that the outcome of local unlearning–methods that only use gradients computed on the forget set–is path-dependent. That is, a model’s behavior during unlearning is influenced by the order of its training stages during learning, making it impossible for path-oblivious algorithms to universally achieve Retrain Equivalence. We empirically demonstrate the same phenomenon in LLM post-training across Llama and Qwen models (1B to 14B) with gradient ascent, NPO, and SimNPO local unlearning algorithms. Models fine-tuned via different orderings of identical training stages diverge in behavior during unlearning, with the degradation in GSM8K accuracy after unlearning varying by over 20% across paths. We also observe that some learning paths consistently produce models that unlearn slowly. During unlearning, whether the probability mass gets squeezed into paraphrasing or alternative concepts is also path-dependent. These results consistently show that Retrain Equivalence is an ill-posed target for local unlearning algorithms, so long as the target models are trained in stages. In situations where access to models’ training histories is hard, the current work calls for rethinking the definition and desiderata of machine unlearning.

[696] Simulation-free Structure Learning for Stochastic Dynamics

Noah El Rimawi-Fine, Adam Stecklov, Lucas Nelson, Mathieu Blanchette, Alexander Tong, Stephen Y. Zhang, Lazar Atanackovic

Main category: cs.LG

TL;DR: StructureFlow is a simulation-free method that jointly learns network structure and stochastic population dynamics of high-dimensional physical systems, addressing both structure learning and dynamical modeling simultaneously.

DetailsMotivation: Many physical systems in natural sciences are high-dimensional, stochastic, and have partial noisy measurements, making it challenging to model dynamics and infer network structure. Existing methods typically address only one of these problems.

Method: StructureFlow is a principled simulation-free approach that jointly learns structure and stochastic population dynamics. It handles structure learning from interventions and dynamical inference of conditional population dynamics.

Result: The method was evaluated on high-dimensional synthetic systems, biologically plausible simulated systems, and experimental single-cell data. It successfully learned underlying system structures while modeling conditional population dynamics.

Conclusion: StructureFlow enables simultaneous learning of system structure and population dynamics, representing a key step toward mechanistic understanding of complex system behaviors.

Abstract: Modeling dynamical systems and unraveling their underlying causal relationships is central to many domains in the natural sciences. Various physical systems, such as those arising in cell biology, are inherently high-dimensional and stochastic in nature, and admit only partial, noisy state measurements. This poses a significant challenge for addressing the problems of modeling the underlying dynamics and inferring the network structure of these systems. Existing methods are typically tailored either for structure learning or modeling dynamics at the population level, but are limited in their ability to address both problems together. In this work, we address both problems simultaneously: we present StructureFlow, a novel and principled simulation-free approach for jointly learning the structure and stochastic population dynamics of physical systems. We showcase the utility of StructureFlow for the tasks of structure learning from interventions and dynamical (trajectory) inference of conditional population dynamics. We empirically evaluate our approach on high-dimensional synthetic systems, a set of biologically plausible simulated systems, and an experimental single-cell dataset. We show that StructureFlow can learn the structure of underlying systems while simultaneously modeling their conditional population dynamics – a key step toward the mechanistic understanding of systems behavior.

[697] Evaluating protein binding interfaces with PUMBA

Azam Shirali, Giri Narasimhan

Main category: cs.LG

TL;DR: PUMBA improves protein-protein docking by replacing Vision Transformers with Vision Mamba architecture, achieving better performance in scoring protein complexes.

DetailsMotivation: Current docking tools need robust scoring functions to differentiate native from non-native protein complexes. Vision Mamba has shown superior performance over Transformers in other domains, suggesting potential for improvement in protein interface evaluation.

Method: Replaced Vision Transformer backbone in PIsToN with Vision Mamba architecture to leverage efficient long-range sequence modeling for image patches, enhancing global and local pattern capture in protein-protein interfaces.

Result: PUMBA consistently outperforms its predecessor PIsToN across multiple large-scale public datasets, demonstrating improved scoring accuracy for protein-protein complexes.

Conclusion: Vision Mamba architecture successfully enhances protein-protein interface evaluation, providing a more effective alternative to Transformer-based approaches for docking scoring functions.

Abstract: Protein-protein docking tools help in studying interactions between proteins, and are essential for drug, vaccine, and therapeutic development. However, the accuracy of a docking tool depends on a robust scoring function that can reliably differentiate between native and non-native complexes. PIsToN is a state-of-the-art deep learning-based scoring function that uses Vision Transformers in its architecture. Recently, the Mamba architecture has demonstrated exceptional performance in both natural language processing and computer vision, often outperforming Transformer-based models in their domains. In this study, we introduce PUMBA (Protein-protein interface evaluation with Vision Mamba), which improves PIsToN by replacing its Vision Transformer backbone with Vision Mamba. This change allows us to leverage Mamba’s efficient long-range sequence modeling for sequences of image patches. As a result, the model’s ability to capture both global and local patterns in protein-protein interface features is significantly improved. Evaluation on several widely-used, large-scale public datasets demonstrates that PUMBA consistently outperforms its original Transformer-based predecessor, PIsToN.

[698] Active Target Discovery under Uninformative Prior: The Power of Permanent and Transient Memory

Anindya Sarkar, Binglin Ji, Yevgeniy Vorobeychik

Main category: cs.LG

TL;DR: A novel active target discovery approach for data-scarce domains that works with uninformative priors, featuring interpretable decision-making and guaranteed monotonic improvement in prior estimates.

DetailsMotivation: Existing active discovery methods relying on strong generative priors fail in domains with extremely limited data or high sampling costs (e.g., rare species discovery, emerging disease diagnostics).

Method: A theoretically principled framework inspired by neuroscience that enables robust exploration even with uninformative priors, featuring interpretable decision-making and guaranteed monotonic improvement in prior estimates.

Result: Comprehensive experiments across domains including species distribution modeling and remote sensing show substantial performance improvements over baseline approaches.

Conclusion: The proposed method provides reliable and adaptable active target discovery in data-scarce real-world scenarios where traditional generative approaches struggle.

Abstract: In many scientific and engineering fields, where acquiring high-quality data is expensive–such as medical imaging, environmental monitoring, and remote sensing–strategic sampling of unobserved regions based on prior observations is crucial for maximizing discovery rates within a constrained budget. The rise of powerful generative models, such as diffusion models, has enabled active target discovery in partially observable environments by leveraging learned priors–probabilistic representations that capture underlying structure from data. With guidance from sequentially gathered task-specific observations, these models can progressively refine exploration and efficiently direct queries toward promising regions. However, in domains where learning a strong prior is infeasible due to extremely limited data or high sampling cost (such as rare species discovery, diagnostics for emerging diseases, etc.), these methods struggle to generalize. To overcome this limitation, we propose a novel approach that enables effective active target discovery even in settings with uninformative priors, ensuring robust exploration and adaptability in complex real-world scenarios. Our framework is theoretically principled and draws inspiration from neuroscience to guide its design. Unlike black-box policies, our approach is inherently interpretable, providing clear insights into decision-making. Furthermore, it guarantees a strong, monotonic improvement in prior estimates with each new observation, leading to increasingly accurate sampling and reinforcing both reliability and adaptability in dynamic settings. Through comprehensive experiments and ablation studies across various domains, including species distribution modeling and remote sensing, we demonstrate that our method substantially outperforms baseline approaches.

[699] Renaissance of RNNs in Streaming Clinical Time Series: Compact Recurrence Remains Competitive with Transformers

Ran Tong, Jiaqi Liu, Su Liu, Xin Hu, Lanruo Wang

Main category: cs.LG

TL;DR: A compact, strictly causal benchmark for streaming clinical time series on MIT-BIH Arrhythmia Database comparing GRU-D and Transformer models for tachycardia risk prediction and heart rate forecasting.

DetailsMotivation: To establish a rigorous benchmark for streaming clinical time series analysis and compare the performance of RNNs vs Transformers in longitudinal monitoring tasks.

Method: Used MIT-BIH Arrhythmia Database with per-second heart rate data. Evaluated two tasks: near-term tachycardia risk (next 10 seconds) and one-step heart rate forecasting. Compared GRU-D (RNN) and Transformer models under matched training budgets against non-learned baselines. Used calibration-aware evaluation for classification and proper scoring for forecasting.

Result: GRU-D slightly outperformed Transformer for tachycardia risk prediction, while Transformer clearly reduced forecasting error compared to GRU-D and persistence baseline.

Conclusion: Model choice in longitudinal monitoring is task-dependent: compact RNNs remain competitive for short-horizon risk scoring, while compact Transformers provide clearer advantages for point forecasting tasks.

Abstract: We present a compact, strictly causal benchmark for streaming clinical time series on the MIT–BIH Arrhythmia Database using per-second heart rate. Two tasks are studied under record-level, non-overlapping splits: near-term tachycardia risk (next ten seconds) and one-step heart rate forecasting. We compare a GRU-D (RNN) and a Transformer under matched training budgets against strong non-learned baselines. Evaluation is calibration-aware for classification and proper for forecasting, with temperature scaling and grouped bootstrap confidence intervals. On MIT-BIH, GRU-D slightly surpasses the Transformer for tachycardia risk, while the Transformer clearly lowers forecasting error relative to GRU-D and persistence. Our results show that, in longitudinal monitoring, model choice is task-dependent: compact RNNs remain competitive for short-horizon risk scoring, whereas compact Transformers deliver clearer gains for point forecasting.

[700] High-Dimensional Privacy-Utility Dynamics of Noisy Stochastic Gradient Descent on Least Squares

Shurong Lin, Eric D. Kolaczyk, Adam Smith, Elliot Paquette

Main category: cs.LG

TL;DR: This paper provides a diffusion-based analysis of noisy SGD for privacy-preserving machine learning, offering exact characterization of statistical risk and privacy loss in high dimensions, with a variant that eliminates the need for gradient sensitivity knowledge.

DetailsMotivation: To address the unclear exact behavior of noisy SGD in high-dimensional settings and eliminate the requirement for explicit gradient sensitivity knowledge that existing methods rely on through gradient clipping.

Method: Uses a diffusion approach to analyze noisy SGD precisely in continuous-time, focusing on least squares problem with ℓ2 regularization, and introduces a variant that doesn’t require gradient sensitivity knowledge.

Result: Provides exact characterization of both statistical risk evolution and privacy loss dynamics in high dimensions for noisy SGD.

Conclusion: The diffusion approach enables precise analysis of noisy SGD’s behavior in high-dimensional privacy-preserving machine learning, with a practical variant that removes the gradient sensitivity requirement.

Abstract: The interplay between optimization and privacy has become a central theme in privacy-preserving machine learning. Noisy stochastic gradient descent (SGD) has emerged as a cornerstone algorithm, particularly in large-scale settings. These variants of gradient methods inject carefully calibrated noise into each update to achieve differential privacy, the gold standard notion of rigorous privacy guarantees. Prior work primarily provides various bounds on statistical risk and privacy loss for noisy SGD, yet the \textit{exact} behavior of the process remains unclear, particularly in high-dimensional settings. This work leverages a diffusion approach to analyze noisy SGD precisely, providing a continuous-time perspective that captures both statistical risk evolution and privacy loss dynamics in high dimensions. Moreover, we study a variant of noisy SGD that does not require explicit knowledge of gradient sensitivity, unlike existing work that assumes or enforces sensitivity through gradient clipping. Specifically, we focus on the least squares problem with $\ell_2$ regularization.

[701] CLIP: Client-Side Invariant Pruning for Mitigating Stragglers in Secure Federated Learning

Anthony DiMaggio, Raghav Sharma, Gururaj Saileshwar

Main category: cs.LG

TL;DR: CLIP introduces client-side invariant neuron pruning with network-aware pruning to mitigate straggler bottlenecks in secure federated learning, accelerating training by 13-34% with minimal accuracy impact.

DetailsMotivation: Secure federated learning preserves data privacy but suffers from performance bottlenecks due to straggler clients with limited computational or network capabilities, which slow down training for all participants.

Method: Proposes CLIP - a client-side invariant neuron pruning technique coupled with network-aware pruning to address compute and network bottlenecks caused by stragglers during secure FL training.

Result: Accelerates secure FL training by 13% to 34% across multiple datasets (CIFAR10, Shakespeare, FEMNIST) with accuracy impact ranging from 1.3% improvement to 2.6% reduction.

Conclusion: CLIP effectively mitigates straggler bottlenecks in secure federated learning with minimal accuracy loss, providing significant training speed improvements across diverse datasets.

Abstract: Secure federated learning (FL) preserves data privacy during distributed model training. However, deploying such frameworks across heterogeneous devices results in performance bottlenecks, due to straggler clients with limited computational or network capabilities, slowing training for all participating clients. This paper introduces the first straggler mitigation technique for secure aggregation with deep neural networks. We propose CLIP, a client-side invariant neuron pruning technique coupled with network-aware pruning, that addresses compute and network bottlenecks due to stragglers during training with minimal accuracy loss. Our technique accelerates secure FL training by 13% to 34% across multiple datasets (CIFAR10, Shakespeare, FEMNIST) with an accuracy impact of between 1.3% improvement to 2.6% reduction.

[702] Zero-Shot Performance Prediction for Probabilistic Scaling Laws

Viktoria Schram, Markus Hiller, Daniel Beck, Trevor Cohn

Main category: cs.LG

TL;DR: The paper presents a multitask learning approach using latent variable multi-output Gaussian Processes to predict learning curves for NLP models, enabling probabilistic scaling laws and reducing computational costs.

DetailsMotivation: To enable informed decision-making for NLP model performance objectives while reducing computational overhead and dataset acquisition costs through learning curve prediction.

Method: Formulates learning curve prediction as a multitask learning problem with two-layer hierarchical data organization, using latent variable multi-output Gaussian Processes to model task correlations and support zero-shot prediction.

Result: The approach facilitates probabilistic scaling laws at lower costs and, with active learning, provides predictions close to ground truth scaling laws. Validated on three small-scale NLP datasets with up to 30 learning curves.

Conclusion: The framework successfully predicts learning curves for various NLP models (nanoGPT, mBART, Transformer, M2M100) using multitask Gaussian Processes, enabling cost-effective scaling law estimation.

Abstract: The prediction of learning curves for Natural Language Processing (NLP) models enables informed decision-making to meet specific performance objectives, while reducing computational overhead and lowering the costs associated with dataset acquisition and curation. In this work, we formulate the prediction task as a multitask learning problem, where each task’s data is modelled as being organized within a two-layer hierarchy. To model the shared information and dependencies across tasks and hierarchical levels, we employ latent variable multi-output Gaussian Processes, enabling to account for task correlations and supporting zero-shot prediction of learning curves (LCs). We demonstrate that this approach facilitates the development of probabilistic scaling laws at lower costs. Applying an active learning strategy, LCs can be queried to reduce predictive uncertainty and provide predictions close to ground truth scaling laws. We validate our framework on three small-scale NLP datasets with up to $30$ LCs. These are obtained from nanoGPT models, from bilingual translation using mBART and Transformer models, and from multilingual translation using M2M100 models of varying sizes.

[703] Resolution-Aware Retrieval Augmented Zero-Shot Forecasting

Iman Deznabi, Peeyush Kumar, Madalina Fiterau

Main category: cs.LG

TL;DR: A Resolution-Aware Retrieval-Augmented Forecasting model that improves zero-shot forecasting by using spatial correlations and temporal frequency decomposition, achieving superior performance in microclimate prediction.

DetailsMotivation: Zero-shot forecasting for unseen conditions without direct historical data is challenging for traditional methods, requiring innovative approaches to handle spatial-temporal dependencies.

Method: Decomposes signals into frequency components and uses resolution-aware retrieval - lower frequencies use broader spatial context while higher frequencies focus on local influences, enabling dynamic data retrieval for new locations.

Result: Significantly outperforms traditional forecasting, numerical weather prediction, and modern time series models, achieving 71% lower MSE than HRRR and 34% lower MSE than Chronos on ERA5 dataset.

Conclusion: The retrieval-augmented and resolution-aware strategy provides an effective, scalable, and data-efficient solution for zero-shot forecasting in microclimate modeling and other domains.

Abstract: Zero-shot forecasting aims to predict outcomes for previously unseen conditions without direct historical data, posing a significant challenge for traditional forecasting methods. We introduce a Resolution-Aware Retrieval-Augmented Forecasting model that enhances predictive accuracy by leveraging spatial correlations and temporal frequency characteristics. By decomposing signals into different frequency components, our model employs resolution-aware retrieval, where lower-frequency components rely on broader spatial context, while higher-frequency components focus on local influences. This allows the model to dynamically retrieve relevant data and adapt to new locations with minimal historical context. Applied to microclimate forecasting, our model significantly outperforms traditional forecasting methods, numerical weather prediction models, and modern foundation time series models, achieving 71% lower MSE than HRRR and 34% lower MSE than Chronos on the ERA5 dataset. Our results highlight the effectiveness of retrieval-augmented and resolution-aware strategies, offering a scalable and data-efficient solution for zero-shot forecasting in microclimate modeling and beyond.

[704] On the Granularity of Causal Effect Identifiability

Yizuo Chen, Adnan Darwiche

Main category: cs.LG

TL;DR: State-based causal effects (interventions on specific states of treatment variables affecting specific states of outcome variables) can be identifiable even when variable-based causal effects are not, particularly when context-specific independencies and conditional functional dependencies are available.

DetailsMotivation: To explore whether causal effects defined at the state level (specific values of variables) rather than variable level can be identifiable from observational data when traditional variable-based approaches fail.

Method: Theoretical analysis of state-based causal effect identifiability, examining how additional knowledge like context-specific independencies and conditional functional dependencies enables identifiability even when variable-based effects are not identifiable.

Result: State-based causal effects may be identifiable when variable-based effects are not, but this separation only occurs with additional knowledge. Knowledge constraining variable states alone doesn’t improve identifiability but can enhance both variable-based and state-based identifiability when combined with other knowledge.

Conclusion: State-based causal analysis can reveal identifiable causal effects that would be missed by traditional variable-based frameworks, highlighting the importance of leveraging context-specific knowledge for causal inference from observational data.

Abstract: The classical notion of causal effect identifiability is defined in terms of treatment and outcome variables. In this note, we consider the identifiability of state-based causal effects: how an intervention on a particular state of treatment variables affects a particular state of outcome variables. We demonstrate that state-based causal effects may be identifiable even when variable-based causal effects may not. Moreover, we show that this separation occurs only when additional knowledge – such as context-specific independencies and conditional functional dependencies – is available. We further examine knowledge that constrains the states of variables, and show that such knowledge does not improve identifiability on its own but can improve both variable-based and state-based identifiability when combined with other knowledge such as context-specific independencies. Our findings highlight situations where causal effects of interest may be estimable from observational data and this identifiability may be missed by existing variable-based frameworks.

[705] LSTM-Based Forecasting and Analysis of EV Charging Demand in a Dense Urban Campus

Zak Ressler, Marcus Grijalva, Angelica Marie Ignacio, Melanie Torres, Abelardo Cuadra Rojas, Rohollah Moghadam, Mohammad Rasoul narimani

Main category: cs.LG

TL;DR: A framework using LSTM neural networks to forecast EV charging load across multiple time scales through data preprocessing and feature extraction.

DetailsMotivation: To enable accurate forecasting of EV charging demand for infrastructure planning, energy management, and grid integration of charging facilities.

Method: Processes raw EV charging data with normalization and feature extraction, then trains a Long Short-Term Memory (LSTM) model to capture both short-term fluctuations and long-term trends.

Result: The model accurately predicts charging demand across daily, weekly, and monthly time scales, with modular design allowing adaptation to different charging locations.

Conclusion: The proposed framework provides valuable insights for EV infrastructure planning and can be applied across diverse deployment scenarios with varying usage patterns.

Abstract: This paper presents a framework for processing EV charging load data in order to forecast future load predictions using a Recurrent Neural Network, specifically an LSTM. The framework processes a large set of raw data from multiple locations and transforms it with normalization and feature extraction to train the LSTM. The pre-processing stage corrects for missing or incomplete values by interpolating and normalizing the measurements. This information is then fed into a Long Short-Term Memory Model designed to capture the short-term fluctuations while also interpreting the long-term trends in the charging data. Experimental results demonstrate the model’s ability to accurately predict charging demand across multiple time scales (daily, weekly, and monthly), providing valuable insights for infrastructure planning, energy management, and grid integration of EV charging facilities. The system’s modular design allows for adaptation to different charging locations with varying usage patterns, making it applicable across diverse deployment scenarios.

[706] Utility-Diversity Aware Online Batch Selection for LLM Supervised Fine-tuning

Heming Zou, Yixiu Mao, Yun Qu, Qi Wang, Xiangyang Ji

Main category: cs.LG

TL;DR: UDS is an efficient online batch selection framework for supervised fine-tuning that selects valuable data samples by considering both utility and diversity, eliminating the need for external resources and reducing training time.

DetailsMotivation: Current supervised fine-tuning methods are computationally expensive and can suffer from overfitting or bias amplification. Existing online batch selection methods focus only on data utility, neglect diversity, require external resources, and add extra training time.

Method: UDS uses nuclear norm of logits matrix to capture data utility and intra-sample diversity, and estimates inter-sample diversity through low-dimensional embedding comparisons with a lightweight memory buffer of historical samples. This eliminates need for external resources and unnecessary backpropagation.

Result: Experiments show UDS consistently outperforms state-of-the-art online batch selection methods under varying data budgets and significantly reduces training time compared to full-dataset fine-tuning.

Conclusion: UDS provides an efficient and effective framework for data curation in supervised fine-tuning that balances utility and diversity while maintaining computational efficiency.

Abstract: Supervised fine-tuning (SFT) is a commonly used technique to adapt large language models (LLMs) to downstream tasks. In practice, SFT on a full dataset is computationally expensive and sometimes suffers from overfitting or bias amplification. This facilitates the rise of data curation in SFT, which prioritizes the most valuable data to optimze. This work studies the online batch selection family that dynamically scores and filters samples during the training process. However, existing popular methods often (i) rely merely on the utility of data to select a subset while neglecting other crucial factors like diversity, (ii) rely on external resources such as reference models or validation sets, and (iii) incur extra training time over full-dataset training. To address these limitations, this work develops \textbf{UDS (Utility-Diversity Sampling)}, a framework for efficient online batch selection in SFT. UDS leverages the nuclear norm of the logits matrix to capture both data utility and intra-sample diversity, while estimating inter-sample diversity through efficient low-dimensional embedding comparisons with a lightweight memory buffer of historical samples. Such a design eliminates the need for external resources and unnecessary backpropagation, securing computational efficiency. Experiments on multiple benchmarks demonstrate that UDS consistently outperforms state-of-the-art online batch selection methods under varying data budgets, and significantly reduces training time compared to full-dataset fine-tuning. Code is available at https://github.com/gfyddha/UDS.

[707] An Efficient Semantic Segmentation Decoder for In-Car or Distributed Applications

Danish Nazir, Gowtham Sai Inti, Timo Bartels, Jan Piewek, Thorsten Bagdonat, Tim Fingscheidt

Main category: cs.LG

TL;DR: Proposed joint feature and task decoding for SegDeformer transformer networks in automotive systems, enabling lower computational complexity for both in-car and distributed applications while maintaining performance.

DetailsMotivation: Modern automotive systems use DNNs for semantic segmentation but face computational complexity challenges, especially with transformer-based models like SegDeformer. Need to reduce complexity while maintaining performance for both in-car and distributed applications.

Method: Joint feature and task decoding approach for SegDeformer transformer networks, enabling computational complexity reduction in both in-car and distributed automotive applications.

Result: For in-car: Increased fps by up to 11.7x on Cityscapes (1.4 to 16.5 fps) and 3.5x on ADE20K (43.3 to 154.3 fps) while maintaining mIoU. For distributed: Achieved SOTA mIoU across bitrates using only 0.14% (ADE20K) and 0.04% (Cityscapes) of cloud DNN parameters compared to previous SOTA.

Conclusion: The proposed joint feature and task decoding enables efficient transformer-based semantic segmentation for automotive systems, significantly reducing computational complexity while maintaining or improving performance in both in-car and distributed applications.

Abstract: Modern automotive systems leverage deep neural networks (DNNs) for semantic segmentation and operate in two key application areas: (1) In-car, where the DNN solely operates in the vehicle without strict constraints on the data rate. (2) Distributed, where one DNN part operates in the vehicle and the other part typically on a large-scale cloud platform with a particular constraint on transmission bitrate efficiency. Typically, both applications share an image and source encoder, while each uses distinct (joint) source and task decoders. Prior work utilized convolutional neural networks for joint source and task decoding but did not investigate transformer-based alternatives such as SegDeformer, which offer superior performance at the cost of higher computational complexity. In this work, we propose joint feature and task decoding for SegDeformer, thereby enabling lower computational complexity in both in-car and distributed applications, despite SegDeformer’s computational demands. This improves scalability in the cloud while reducing in-car computational complexity. For the in-car application, we increased the frames per second (fps) by up to a factor of $11.7$ ($1.4$ fps to $16.5$ fps) on Cityscapes and by up to a factor of $3.5$ ($43.3$ fps to $154.3$ fps) on ADE20K, while being on-par w.r.t.\ the mean intersection over union (mIoU) of the transformer-based baseline that doesn’t compress by a source codec. For the distributed application, we achieve state-of-the-art (SOTA) over a wide range of bitrates on the mIoU metric, while using only $0.14$% ($0.04$%) of cloud DNN parameters used in previous SOTA, reported on ADE20K (Cityscapes).

[708] Peering Inside the Black Box: Uncovering LLM Errors in Optimization Modelling through Component-Level Evaluation

Dania Refai, Moataz Ahmed

Main category: cs.LG

TL;DR: A comprehensive evaluation framework for LLM-generated mathematical optimization formulations that assesses component-level metrics beyond conventional solution accuracy.

DetailsMotivation: Current evaluations treat formulations as a whole using coarse metrics like solution accuracy or runtime, which obscure structural or numerical errors in LLM-generated optimization formulations.

Method: Developed a component-level evaluation framework with metrics including precision/recall for variables/constraints, constraint/objective RMSE, and efficiency indicators. Evaluated GPT-5, LLaMA 3.1 Instruct, and DeepSeek Math across various optimization problems under six prompting strategies.

Result: GPT-5 consistently outperformed other models, with chain-of-thought, self-consistency, and modular prompting being most effective. Solver performance primarily depends on high constraint recall and low constraint RMSE for structural correctness and solution reliability.

Conclusion: Three key principles for NLP-to-optimization modeling: (i) complete constraint coverage prevents violations, (ii) minimizing constraint RMSE ensures solver-level accuracy, and (iii) concise outputs improve computational efficiency. The framework enables fine-grained diagnostic evaluation of LLMs in optimization.

Abstract: Large language models (LLMs) are increasingly used to convert natural language descriptions into mathematical optimization formulations. Current evaluations often treat formulations as a whole, relying on coarse metrics like solution accuracy or runtime, which obscure structural or numerical errors. In this study, we present a comprehensive, component-level evaluation framework for LLM-generated formulations. Beyond the conventional optimality gap, our framework introduces metrics such as precision and recall for decision variables and constraints, constraint and objective root mean squared error (RMSE), and efficiency indicators based on token usage and latency. We evaluate GPT-5, LLaMA 3.1 Instruct, and DeepSeek Math across optimization problems of varying complexity under six prompting strategies. Results show that GPT-5 consistently outperforms other models, with chain-of-thought, self-consistency, and modular prompting proving most effective. Analysis indicates that solver performance depends primarily on high constraint recall and low constraint RMSE, which together ensure structural correctness and solution reliability. Constraint precision and decision variable metrics play secondary roles, while concise outputs enhance computational efficiency. These findings highlight three principles for NLP-to-optimization modeling: (i) Complete constraint coverage prevents violations, (ii) minimizing constraint RMSE ensures solver-level accuracy, and (iii) concise outputs improve computational efficiency. The proposed framework establishes a foundation for fine-grained, diagnostic evaluation of LLMs in optimization modeling.

[709] SAMOSA: Sharpness Aware Minimization for Open Set Active learning

Young In Kim, Andrea Agiollo, Rajiv Khanna

Main category: cs.LG

TL;DR: SAMOSA is a novel open set active learning method that uses sharpness-aware minimization to select informative samples from unlabeled data containing unknown classes, achieving improved accuracy without computational overhead.

DetailsMotivation: To reduce the high cost of data labeling in machine learning by developing an active learning approach that can effectively select informative samples from unlabeled data that includes irrelevant or unknown classes.

Method: Proposes SAMOSA (Sharpness Aware Minimization for Open Set Active Learning) which builds on theoretical findings about data typicality’s impact on generalization. It actively queries samples based on their typicality, identifying atypical samples near model decision boundaries to prioritize highly informative samples for targeted classes and distinguish between targeted and unwanted classes.

Result: Extensive experiments show SAMOSA achieves up to 3% accuracy improvement over state-of-the-art methods across several datasets, while not introducing computational overhead.

Conclusion: SAMOSA is an effective open set active learning approach that successfully reduces labeling burden by selecting informative samples from unlabeled data containing unknown classes, outperforming existing methods in accuracy without additional computational costs.

Abstract: Modern machine learning solutions require extensive data collection where labeling remains costly. To reduce this burden, open set active learning approaches aim to select informative samples from a large pool of unlabeled data that includes irrelevant or unknown classes. In this context, we propose Sharpness Aware Minimization for Open Set Active Learning (SAMOSA) as an effective querying algorithm. Building on theoretical findings concerning the impact of data typicality on the generalization properties of traditional stochastic gradient descent (SGD) and sharpness-aware minimization (SAM), SAMOSA actively queries samples based on their typicality. SAMOSA effectively identifies atypical samples that belong to regions of the embedding manifold close to the model decision boundaries. Therefore, SAMOSA prioritizes the samples that are (i) highly informative for the targeted classes, and (ii) useful for distinguishing between targeted and unwanted classes. Extensive experiments show that SAMOSA achieves up to 3% accuracy improvement over the state of the art across several datasets, while not introducing computational overhead. The source code of our experiments is available at: https://anonymous.4open.science/r/samosa-DAF4

[710] Learning to play: A Multimodal Agent for 3D Game-Play

Yuguang Yue, Irakli Salia, Samuel Hunt, Christopher Green, Wenzhe Shi, Jonathan J Hunt

Main category: cs.LG

TL;DR: The paper presents a method for training text-conditioned agents to play 3D first-person games using a large dataset of human gameplay with inverse dynamics modeling and behavior cloning.

DetailsMotivation: 3D first-person video games present a challenging environment for real-time multi-modal reasoning, requiring agents to process visual input and respond to text instructions while navigating complex 3D environments.

Method: Collected a large diverse dataset of human gameplay with text instructions, learned an inverse dynamics model to impute actions on public videos, and trained text-conditioned agents using behavior cloning with a custom real-time inference architecture.

Result: The resulting model can play various 3D games and respond to text input, demonstrating capability in real-time game playing with multimodal reasoning.

Conclusion: While successful in creating functional game-playing agents, challenges remain including long-horizon tasks and quantitative evaluation across multiple games.

Abstract: We argue that 3-D first-person video games are a challenging environment for real-time multi-modal reasoning. We first describe our dataset of human game-play, collected across a large variety of 3-D first-person games, which is both substantially larger and more diverse compared to prior publicly disclosed datasets, and contains text instructions. We demonstrate that we can learn an inverse dynamics model from this dataset, which allows us to impute actions on a much larger dataset of publicly available videos of human game play that lack recorded actions. We then train a text-conditioned agent for game playing using behavior cloning, with a custom architecture capable of realtime inference on a consumer GPU. We show the resulting model is capable of playing a variety of 3-D games and responding to text input. Finally, we outline some of the remaining challenges such as long-horizon tasks and quantitative evaluation across a large set of games.

[711] Leave It to the Experts: Detecting Knowledge Distillation via MoE Expert Signatures

Pingzhi Li, Morris Yu-Chao Huang, Zhen Tan, Qingquan Song, Jie Peng, Kai Zou, Yu Cheng, Kaidi Xu, Tianlong Chen

Main category: cs.LG

TL;DR: A novel KD detection framework that identifies knowledge distillation by analyzing MoE structural habits and routing patterns, achieving >94% accuracy and robustness against prompt-based evasion.

DetailsMotivation: Existing KD detection methods based on self-identity or output similarity are easily evaded through prompt engineering, posing risks to intellectual property protection and LLM diversity.

Method: Exploits transfer of MoE structural habits, especially internal routing patterns, and proposes Shadow-MoE for black-box detection by constructing proxy MoE representations via auxiliary distillation.

Result: Achieves >94% detection accuracy across various scenarios with strong robustness to prompt-based evasion, outperforming existing baselines.

Conclusion: The framework effectively detects knowledge distillation by analyzing structural habits transfer in LLMs, providing a comprehensive benchmark for future research.

Abstract: Knowledge Distillation (KD) accelerates training of large language models (LLMs) but poses intellectual property protection and LLM diversity risks. Existing KD detection methods based on self-identity or output similarity can be easily evaded through prompt engineering. We present a KD detection framework effective in both white-box and black-box settings by exploiting an overlooked signal: the transfer of MoE “structural habits”, especially internal routing patterns. Our approach analyzes how different experts specialize and collaborate across various inputs, creating distinctive fingerprints that persist through the distillation process. To extend beyond the white-box setup and MoE architectures, we further propose Shadow-MoE, a black-box method that constructs proxy MoE representations via auxiliary distillation to compare these patterns between arbitrary model pairs. We establish a comprehensive, reproducible benchmark that offers diverse distilled checkpoints and an extensible framework to facilitate future research. Extensive experiments demonstrate >94% detection accuracy across various scenarios and strong robustness to prompt-based evasion, outperforming existing baselines while highlighting the structural habits transfer in LLMs.

[712] 3D-GSRD: 3D Molecular Graph Auto-Encoder with Selective Re-mask Decoding

Chang Wu, Zhiyuan Liu, Wen Shu, Liang Wang, Yanchen Luo, Wenqiang Lei, Yatao Bian, Junfeng Fang, Xiang Wang

Main category: cs.LG

TL;DR: 3D-GSRD is a novel molecular representation learning method that uses selective re-mask decoding to effectively learn 3D molecular structures while preventing 2D structure leakage.

DetailsMotivation: Extending masked graph modeling from 2D to 3D is challenging due to conflicting requirements: avoiding 2D structure leakage to the decoder while providing sufficient 2D context for reconstructing re-masked atoms.

Method: Proposes 3D-GSRD with Selective Re-mask Decoding (SRD) that re-masks only 3D-relevant information while preserving 2D graph structures, combined with a 3D Relational-Transformer encoder and structure-independent decoder.

Result: Achieves state-of-the-art performance on 7 out of 8 targets in the MD17 molecular property prediction benchmark.

Conclusion: The selective re-mask decoding approach effectively enhances encoder’s role in molecular representation learning and demonstrates strong downstream performance for 3D molecular structures.

Abstract: Masked graph modeling (MGM) is a promising approach for molecular representation learning (MRL).However, extending the success of re-mask decoding from 2D to 3D MGM is non-trivial, primarily due to two conflicting challenges: avoiding 2D structure leakage to the decoder, while still providing sufficient 2D context for reconstructing re-masked atoms.To address these challenges, we propose 3D-GSRD: a 3D Molecular Graph Auto-Encoder with Selective Re-mask Decoding. The core innovation of 3D-GSRD lies in its Selective Re-mask Decoding(SRD), which re-masks only 3D-relevant information from encoder representations while preserving the 2D graph structures.This SRD is synergistically integrated with a 3D Relational-Transformer(3D-ReTrans) encoder alongside a structure-independent decoder. We analyze that SRD, combined with the structure-independent decoder, enhances the encoder’s role in MRL. Extensive experiments show that 3D-GSRD achieves strong downstream performance, setting a new state-of-the-art on 7 out of 8 targets in the widely used MD17 molecular property prediction benchmark. The code is released at https://github.com/WuChang0124/3D-GSRD.

[713] Mixed-Precision Quantization for Language Models: Techniques and Prospects

Mariam Rakka, Marios Fournarakis, Olga Krestinskaya, Jinane Bazzi, Khaled N. Salama, Fadi Kurdahi, Ahmed M. Eltawil, Mohammed E. Fouda

Main category: cs.LG

TL;DR: This survey provides a comprehensive overview of Mixed-Precision Quantization frameworks for Large Language Models (MXPLMs), analyzing bit allocation strategies, comparing performance metrics, and identifying future research directions.

DetailsMotivation: The rapid scaling of language models has led to unsustainable computational, memory, and energy requirements, creating a need for efficient compression techniques like mixed-precision quantization to balance efficiency and accuracy.

Method: The survey categorizes and compares MXPLM frameworks based on bit allocation strategies and precision configurations across weights, activations, and key-value caches, while contrasting them with earlier mixed-precision methods for deep neural networks.

Result: The comparative analysis reveals differences in perplexity, zero-shot task performance, and deployment trade-offs, identifying which strategies transfer well to LM settings and which face challenges.

Conclusion: The work consolidates recent advances in mixed-precision quantization for large-scale language models and identifies key future directions including hardware-aware design, activation quantization, and scalable optimization methods for billion-parameter models.

Abstract: The rapid scaling of language models (LMs) has resulted in unprecedented computational, memory, and energy requirements, making their training and deployment increasingly unsustainable. Quantization has emerged as an essential compression technique to reduce model size, alleviate memory bottlenecks, and accelerate inference. However, while uniform low-bit quantization (e.g., INT8, INT4) provides significant efficiency gains, it can degrade accuracy in sensitive components of transformer-based LMs. Mixed-precision quantization offers a promising alternative by selectively allocating precision across layers or within tensors to balance efficiency and accuracy. This survey provides a comprehensive overview of Mixed-Precision quantization frameworks for LMs (MXPLMs). We first review quantization fundamentals, including uniform and non-uniform quantizers, quantization granularity, and methods widely used in post-training quantization. We then categorize and compare recent MXPLM frameworks according to their bit allocation strategies and precision configurations across weights, activations, and key-value caches. A comparative analysis highlights differences in perplexity, zero-shot task performance, and deployment trade-offs. Furthermore, we contrast MXPLMs with earlier mixed-precision quantization methods for deep neural networks, identifying strategies that transfer and those that face challenges in the LM setting. Finally, we summarize open issues and future directions, including hardware-aware design, activation quantization, and scalable optimization methods for billion-parameter models. By consolidating recent advances, this work serves as a reference for understanding the current landscape and research prospects of mixed-precision quantization for large-scale language models.

[714] Forgetting to Forget: Attention Sink as A Gateway for Backdooring LLM Unlearning

Bingqi Shang, Yiwei Chen, Yihua Zhang, Bingquan Shen, Sijia Liu

Main category: cs.LG

TL;DR: This paper introduces ‘backdoor unlearning’ - a vulnerability where LLM unlearning appears successful but secretly embeds triggers that can restore forgotten knowledge when activated, exploiting the attention sink phenomenon.

DetailsMotivation: To investigate whether LLM unlearning can be backdoored to appear successful while secretly maintaining the ability to recover unlearned content through hidden triggers.

Method: The authors design backdoor attacks by placing triggers at attention sink positions in LLMs and aligning their attention values to enhance backdoor persistence, leveraging the observation that shallow input tokens consistently attract disproportionate attention.

Result: Extensive experiments show that attention-sink-guided backdoor unlearning reliably restores forgotten knowledge when triggers are present while behaving indistinguishably from normally unlearned models when triggers are absent.

Conclusion: Attention sinks serve as effective gateways for backdoor unlearning attacks, highlighting a critical vulnerability in LLM unlearning mechanisms that requires new defensive strategies.

Abstract: Large language model (LLM) unlearning has become a critical mechanism for removing undesired data, knowledge, or behaviors from pre-trained models while retaining their general utility. Yet, with the rise of open-weight LLMs, we ask: can the unlearning process itself be backdoored, appearing successful under normal conditions yet reverting to pre-unlearned behavior when a hidden trigger is activated? Drawing inspiration from classical backdoor attacks that embed triggers into training data to enforce specific behaviors, we investigate backdoor unlearning, where models forget as intended in the clean setting but recover forgotten knowledge when the trigger appears. We show that designing such attacks presents unique challenges, hinging on where triggers are placed and how backdoor training is reinforced. We uncover a strong link between backdoor efficacy and the attention sink phenomenon, i.e., shallow input tokens consistently attract disproportionate attention in LLMs. Our analysis reveals that these attention sinks serve as gateways for backdoor unlearning: placing triggers at sink positions and aligning their attention values markedly enhances backdoor persistence. Extensive experiments validate these findings, showing that attention-sink-guided backdoor unlearning reliably restores forgotten knowledge in the presence of backdoor triggers, while behaving indistinguishably from a normally unlearned model when triggers are absent. Code is available at https://github.com/OPTML-Group/Unlearn-Backdoor.

[715] Computational Budget Should Be Considered in Data Selection

Weilin Wan, Weizhong Zhang, Cheng Jin

Main category: cs.LG

TL;DR: The paper proposes CADS, a computational budget-aware data selection method that formulates data selection as a bilevel optimization problem, addressing key challenges in gradient estimation and computational efficiency.

DetailsMotivation: Existing data selection methods ignore compute budget constraints, but empirical studies show no algorithm consistently outperforms others across varying budgets. Different budgets impose distinct requirements on data quantity, quality, and distribution.

Method: Proposes CADS as a bilevel optimization framework: inner loop trains model within budget constraints on selected data subset, outer loop optimizes data selection based on model evaluation. Uses probabilistic reparameterization and Hessian-free policy gradient for gradient estimation, and transforms inner optimization into penalty term.

Result: Extensive experiments show performance gains of up to 14.42% over baselines in vision and language benchmarks.

Conclusion: Compute budget must be integral to data-selection strategies, and the proposed CADS method effectively addresses this by solving the bilevel optimization challenges through efficient gradient estimation and computational optimization.

Abstract: Data selection improves computational efficiency by choosing informative subsets of training samples. However, existing methods ignore the compute budget, treating data selection and importance evaluation independently of compute budget constraints. Yet empirical studies show no algorithm can consistently outperform others (or even random selection) across varying budgets. We therefore argue that compute budget must be integral to data-selection strategies, since different budgets impose distinct requirements on data quantity, quality, and distribution for effective training. To this end, we propose a novel Computational budget-Aware Data Selection (CADS) method and naturally formulate it into a bilevel optimization framework, where the inner loop trains the model within the constraints of the computational budget on some selected subset of training data, while the outer loop optimizes data selection based on model evaluation. Our technical contributions lie in addressing two main challenges in solving this bilevel optimization problem: the expensive Hessian matrix estimation for outer-loop gradients and the computational burden of achieving inner-loop optimality during iterations. To solve the first issue, we propose a probabilistic reparameterization strategy and compute the gradient using a Hessian-free policy gradient estimator. To address the second challenge, we transform the inner optimization problem into a penalty term in the outer objective, further discovering that we only need to estimate the minimum of a one-dimensional loss to calculate the gradient, significantly improving efficiency. Extensive experiments show that our method achieves performance gains of up to 14.42% over baselines in vision and language benchmarks.

[716] Do LLMs Recognize Your Latent Preferences? A Benchmark for Latent Information Discovery in Personalized Interaction

Ioannis Tsaknakis, Bingqing Song, Shuyu Gan, Dongyeop Kang, Alfredo Garcia, Gaowen Liu, Charles Fleming, Mingyi Hong

Main category: cs.LG

TL;DR: LLMs struggle with inferring latent user preferences in personalized interactions. A new benchmark evaluates their ability to discover hidden user attributes through conversation across three settings: 20 Questions, Personalized QA, and Personalized Text Summarization.

DetailsMotivation: LLMs' general text generation capability becomes limiting when user-specific preferences are needed, as users rarely articulate all preferences explicitly. The paper aims to evaluate whether LLMs can uncover and reason about latent information through conversation.

Method: Introduces a unified benchmark with tri-agent framework (User, Assistant, Judge) for evaluating latent information discovery across three settings: 20 Questions game, Personalized Question Answering, and Personalized Text Summarization.

Result: LLMs can surface latent information through dialogue, but success varies dramatically from 32% to 98% depending on task complexity, topic, and number of hidden attributes.

Conclusion: Effective preference inference remains an open frontier for building truly adaptive AI systems, and this benchmark provides the first systematic framework for studying latent information discovery in personalized interaction.

Abstract: Large Language Models (LLMs) excel at producing broadly relevant text, but this generality becomes a limitation when user-specific preferences are required, such as recommending restaurants or planning travel. In these scenarios, users rarely articulate every preference explicitly; instead, much of what they care about remains latent, waiting to be inferred. This raises a fundamental question: Can LLMs uncover and reason about such latent information through conversation? We address this problem by introducing a unified benchmark for evaluating latent information discovery - the ability of LLMs to reveal and utilize hidden user attributes through multi-turn interaction. The benchmark spans three progressively realistic settings: the classic 20 Questions game, Personalized Question Answering, and Personalized Text Summarization. All tasks share a tri-agent framework (User, Assistant, Judge) enabling turn-level evaluation of elicitation and adaptation. Our results reveal that while LLMs can indeed surface latent information through dialogue, their success varies dramatically with context: from 32% to 98%, depending on task complexity, topic, and number of hidden attributes. This benchmark provides the first systematic framework for studying latent information discovery in personalized interaction, highlighting that effective preference inference remains an open frontier for building truly adaptive AI systems.

[717] Improving Model Representation and Reducing KV Cache via Skip Connections with First Value Heads

Zhoutong Wu, Yuan Zhang, Yiming Dong, Chenheng Zhang, Cong Fang, Kun Yuan, Zhouchen Lin

Main category: cs.LG

TL;DR: SkipV1Former is a Transformer variant that uses skip connections from the first layer’s Value heads to reduce KV cache by ~25% while improving perplexity, and can be uptrained from existing models with minimal compute.

DetailsMotivation: To improve Transformer representation without increasing memory/compute costs, addressing the limitations of prior works that either improve expressivity without reducing KV costs or reduce memory at the cost of weaker representation.

Method: Reuses half of Value heads from the first layer in deeper layers while computing the other half normally, cutting Value projections and V cache by nearly 50%. Theoretically restores information lost to compression and accelerates implicit mesa-optimization.

Result: Consistent ~25% KV cache reduction across model scales while improving perplexity compared to standard MHA Transformers and advanced variants. Can combine with other methods like Group-Query Attention for up to 50% KV cache reduction while maintaining performance improvement.

Conclusion: SkipV1Former effectively balances representation quality and resource efficiency, offering a practical approach to scale Transformers with reduced memory footprint and can be efficiently adapted from existing models.

Abstract: Transformer models have driven breakthroughs across various language tasks by their strong capability to learn rich contextual representations. Scaling them to improve representation, however, often demands substantial memory and compute costs, such as the Key-Value (KV) cache used during auto-regressive decoding. Skip connections offer a promising way to improve representation without bloating resource usage, yet most prior works either improve expressivity while leaving KV costs unchanged, or reduce memory at the cost of weaker representation. In this work, we propose SkipV1Former, a Transformer variant that uses skip connections from the first layer’s Value heads to strengthen model representation and reduce KV cache. Specifically, from the second block onward, each layer reuses half of its Value heads from the very first layer, while computing the other half as usual-cutting Value projections and V cache by nearly 50 %. Theoretically, we show that routing uncompressed first-layer Values into deeper layers restores information lost to compression and accelerates the model’s implicit mesa-optimization-a key pattern of Transformer in auto-regressive tasks. Empirically, across different model scales, SkipV1Former delivers consistent reductions of approximately 25 % in KV cache while improving perplexity relative to standard Multi-Head Attention (MHA) Transformers and some advanced variants. Moreover, we propose a recipe for uptraining existing MHA Transformer checkpoints to SkipV1Former with only 10-15% additional compute. Finally, SkipV1Former can seamlessly combine advanced methods like Group-Query Attention and Multi-Latent Attention to achieve further KV cache savings and performance improvement. When combined with YOCO, it cuts KV cache size by nearly 50 % while still improving performance.

[718] Graph Learning is Suboptimal in Causal Bandits

Mohammad Shahverdikondori, Jalal Etesami, Negar Kiyavash

Main category: cs.LG

TL;DR: Learning parent sets in causal bandits is suboptimal for regret minimization; parent identification conflicts with regret minimization objectives. Novel algorithms bypass graph recovery and achieve near-optimal performance.

DetailsMotivation: Previous causal bandit approaches focus on identifying reward parents or jointly learning parents while minimizing regret. This work investigates whether such strategies are optimal for regret minimization.

Method: Proves that regret minimization and parent identification are conflicting objectives. Analyzes known and unknown parent set size regimes, establishes novel regret lower bounds, and proposes algorithms that bypass graph and parent recovery.

Result: Shows learning parent set is suboptimal. Experiments demonstrate large performance gap between proposed method and existing baselines across various environments.

Conclusion: Parent identification is unnecessary for regret minimization in causal bandits; bypassing graph recovery leads to nearly optimal performance.

Abstract: We study regret minimization in causal bandits under causal sufficiency where the underlying causal structure is not known to the agent. Previous work has focused on identifying the reward’s parents and then applying classic bandit methods to them, or jointly learning the parents while minimizing regret. We investigate whether such strategies are optimal. Somewhat counterintuitively, our results show that learning the parent set is suboptimal. We do so by proving that there exist instances where regret minimization and parent identification are fundamentally conflicting objectives. We further analyze both the known and unknown parent set size regimes, establish novel regret lower bounds that capture the combinatorial structure of the action space. Building on these insights, we propose nearly optimal algorithms that bypass graph and parent recovery, demonstrating that parent identification is indeed unnecessary for regret minimization. Experiments confirm that there exists a large performance gap between our method and existing baselines in various environments.

[719] Soft-Masked Diffusion Language Models

Michael Hersche, Samuel Moor-Smith, Thomas Hofmann, Abbas Rahimi

Main category: cs.LG

TL;DR: The paper introduces soft-masking (SM), a novel method that improves masked diffusion language models by dynamically blending mask token embeddings with top-k predicted tokens from previous steps, preserving valuable predictive information that is normally discarded.

DetailsMotivation: Current masked diffusion language models use binary decisions (retain mask or replace with predicted token) which discard valuable predictive information when masks are retained, limiting model performance.

Method: Proposed soft-masking (SM) method that dynamically blends mask token embeddings with embeddings of top-k predicted tokens from previous decoding steps, providing more informative priors and allowing partial token information to propagate across steps.

Result: SM improves perplexity and MAUVE scores in 169M parameter models, and consistently enhances performance across multiple coding benchmarks when applied to state-of-the-art diffusion models Dream-7B and Dream-Coder-7B, particularly in high-throughput settings.

Conclusion: Soft-masking effectively addresses the limitations of binary masking in diffusion language models by preserving predictive information across decoding steps, leading to improved performance across various metrics and benchmarks.

Abstract: Diffusion models have demonstrated strong potential in language modeling, offering various advantages over traditional autoregressive approaches. Their ability to generate and revise entire responses in parallel enables faster generation and built-in self-correction mechanisms. Most modern diffusion-based language models employ masked diffusion, where decoding involves iteratively processing masked tokens based on a binary decision: either retaining the mask or replacing it with the predicted token. However, this binary choice discards valuable predictive information when the mask is retained. To address this limitation, we introduce soft-masking (SM), a novel method that dynamically blends the embedding of the mask token with the embeddings of the top-$k$ predicted tokens from the previous decoding step, for each retained mask. This provides the model with a more informative prior, preserving context from earlier computations and allowing partial information about masked tokens to propagate beyond a single step. We propose a training methodology that adapts a pretrained masked diffusion language model to incorporate SM. We demonstrate that continuing pretraining a 169M parameter model with SM leads to improved perplexity and MAUVE scores. Furthermore, we finetune two state-of-the-art diffusion models, Dream-7B and Dream-Coder-7B, with SM. SM consistently improves performance across multiple coding benchmarks, particularly in high-throughput settings.

[720] Needles in the Landscape: Semi-Supervised Pseudolabeling for Archaeological Site Discovery under Label Scarcity

Simon Jaxy, Anton Theys, Patrick Willett, W. Chris Carleton, Ralf Vandam, Pieter Libin

Main category: cs.LG

TL;DR: A semi-supervised deep learning approach using positive-unlabeled learning and dynamic pseudolabeling with CRF refinement for archaeological site prediction, achieving state-of-the-art performance on geospatial and satellite imagery datasets.

DetailsMotivation: To address structural label scarcity in archaeology where positives are rare and most locations are unlabeled, requiring methods that can work with limited labeled data.

Method: Semi-supervised positive-unlabeled learning implemented as semantic segmentation with dynamic pseudolabeling and Conditional Random Field refinement via RNN to handle severe class imbalance.

Result: Performs on par with state-of-the-art LAMAP on geospatial DEM data with higher Dice scores, maintains performance on raw satellite imagery with improved interpretability in predictive surfaces.

Conclusion: Semi-supervised learning offers a promising approach for identifying undiscovered archaeological sites across large, sparsely annotated landscapes.

Abstract: Archaeological predictive modelling estimates where undiscovered sites are likely to occur by combining known locations with environmental, cultural, and geospatial variables. We address this challenge using a deep learning approach but must contend with structural label scarcity inherent to archaeology: positives are rare, and most locations are unlabeled. To address this, we adopt a semi-supervised, positive-unlabeled (PU) learning strategy, implemented as a semantic segmentation model and evaluated on two datasets covering a representative range of archaeological periods. Our approach employs dynamic pseudolabeling, refined with a Conditional Random Field (CRF) implemented via an RNN, increasing label confidence under severe class imbalance. On a geospatial dataset derived from a digital elevation model (DEM), our model performs on par with the state-of-the-art, LAMAP, while achieving higher Dice scores. On raw satellite imagery, assessed end-to-end with stratified k-fold cross-validation, it maintains performance and yields predictive surfaces with improved interpretability. Overall, our results indicate that semi-supervised learning offers a promising approach to identifying undiscovered sites across large, sparsely annotated landscapes.

[721] LILO: Bayesian Optimization with Interactive Natural Language Feedback

Katarzyna Kobalczyk, Zhiyuan Jerry Lin, Benjamin Letham, Zhuokai Zhao, Maximilian Balandat, Eytan Bakshy

Main category: cs.LG

TL;DR: A language-in-the-loop framework that uses LLMs to convert natural language feedback into scalar utilities for Bayesian Optimization, outperforming conventional methods.

DetailsMotivation: Feedback is essential for translating complex, nuanced goals into quantifiable objectives, but existing methods have restrictive feedback formats and require domain-specific customization.

Method: Uses LLMs to convert varied natural language feedback into consistent utility signals, enabling flexible user priors without manual kernel design while maintaining BO’s sample efficiency.

Result: Outperforms conventional BO baselines and LLM-only optimizers, especially in feedback-limited regimes, and provides a more natural interface for decision makers.

Conclusion: The hybrid approach successfully combines LLMs’ natural language understanding with BO’s principled optimization, creating an effective framework for complex goal optimization.

Abstract: For many real-world applications, feedback is essential in translating complex, nuanced, or subjective goals into quantifiable optimization objectives. We propose a language-in-the-loop framework that uses a large language model (LLM) to convert unstructured feedback in the form of natural language into scalar utilities to conduct BO over a numeric search space. Unlike preferential BO, which only accepts restricted feedback formats and requires customized models for each domain-specific problem, our approach leverages LLMs to turn varied types of textual feedback into consistent utility signals and to easily include flexible user priors without manual kernel design. At the same time, our method maintains the sample efficiency and principled uncertainty quantification of BO. We show that this hybrid method not only provides a more natural interface to the decision maker but also outperforms conventional BO baselines and LLM-only optimizers, particularly in feedback-limited regimes.

[722] Efficient High-Accuracy PDEs Solver with the Linear Attention Neural Operator

Ming Zhong, Zhenya Yan

Main category: cs.LG

TL;DR: LANO is a neural operator that uses agent tokens to achieve linear complexity while maintaining softmax attention accuracy, outperforming state-of-the-art PDE solvers by 19.5% on benchmarks.

DetailsMotivation: To overcome the scalability-accuracy trade-off in transformer-based neural operators where softmax attention has quadratic complexity but high accuracy, while linear attention variants reduce cost but suffer accuracy degradation.

Method: Introduces Linear Attention Neural Operator (LANO) with agent-based attention mechanism using a compact set of M agent tokens (M « N) that mediate global interactions among N tokens, achieving linear complexity O(MNd).

Result: LANO achieves 19.5% average accuracy improvement over state-of-the-art neural PDE solvers including Transolver, while maintaining linear complexity and preserving softmax attention’s expressive power.

Conclusion: LANO successfully bridges the gap between linear complexity and softmax-level performance, establishing a scalable, high-accuracy foundation for scientific machine learning applications.

Abstract: Neural operators offer a powerful data-driven framework for learning mappings between function spaces, in which the transformer-based neural operator architecture faces a fundamental scalability-accuracy trade-off: softmax attention provides excellent fidelity but incurs quadratic complexity $\mathcal{O}(N^2 d)$ in the number of mesh points $N$ and hidden dimension $d$, while linear attention variants reduce cost to $\mathcal{O}(N d^2)$ but often suffer significant accuracy degradation. To address the aforementioned challenge, in this paper, we present a novel type of neural operators, Linear Attention Neural Operator (LANO), which achieves both scalability and high accuracy by reformulating attention through an agent-based mechanism. LANO resolves this dilemma by introducing a compact set of $M$ agent tokens $(M \ll N)$ that mediate global interactions among $N$ tokens. This agent attention mechanism yields an operator layer with linear complexity $\mathcal{O}(MN d)$ while preserving the expressive power of softmax attention. Theoretically, we demonstrate the universal approximation property, thereby demonstrating improved conditioning and stability properties. Empirically, LANO surpasses current state-of-the-art neural PDE solvers, including Transolver with slice-based softmax attention, achieving average $19.5%$ accuracy improvement across standard benchmarks. By bridging the gap between linear complexity and softmax-level performance, LANO establishes a scalable, high-accuracy foundation for scientific machine learning applications.

[723] Trace Regularity PINNs: Enforcing $\mathrm{H}^{\frac{1}{2}}(\partial Ω)$ for Boundary Data

Doyoon Kim, Junbin Song

Main category: cs.LG

TL;DR: TRPINN is an enhanced PINN that enforces boundary loss in the correct Sobolev-Slobodeckij norm H¹/²(∂Ω), reducing computational cost and improving convergence stability while achieving faster convergence than standard PINNs.

DetailsMotivation: Standard PINNs have limitations in handling boundary conditions, particularly for problems with highly oscillatory Dirichlet boundary conditions, and may fail to converge properly due to improper norm enforcement.

Method: Proposes TRPINN that enforces boundary loss in the H¹/²(∂Ω) norm, computes only essential portions of the semi-norm to reduce cost, avoids denominator evaluations for stability, and uses Neural Tangent Kernel analysis.

Result: TRPINN converges to true solution in H¹(Ω) sense, converges faster than standard PINNs, succeeds where standard PINNs fail on Laplace equation with oscillatory boundary conditions, and shows 1-3 decimal digit performance improvements.

Conclusion: TRPINN provides a theoretically sound and computationally efficient approach for enforcing boundary conditions in PINNs, with proven convergence properties and superior performance over standard methods.

Abstract: We propose an enhanced physics-informed neural network (PINN), the Trace Regularity Physics-Informed Neural Network (TRPINN), which enforces the boundary loss in the Sobolev-Slobodeckij norm $H^{1/2}(\partial \Omega)$, the correct trace space associated with $H^1(\Omega)$. We reduce computational cost by computing only the theoretically essential portion of the semi-norm and enhance convergence stability by avoiding denominator evaluations in the discretization. By incorporating the exact $H^{1/2}(\partial \Omega)$ norm, we show that the approximation converges to the true solution in the $H^{1}(\Omega)$ sense, and, through Neural Tangent Kernel (NTK) analysis, we demonstrate that TRPINN can converge faster than standard PINNs. Numerical experiments on the Laplace equation with highly oscillatory Dirichlet boundary conditions exhibit cases where TRPINN succeeds even when standard PINNs fail, and show performance improvements of one to three decimal digits.

[724] Mapping Post-Training Forgetting in Language Models at Scale

Jackson Harmon, Andreas Hochlehnert, Matthias Bethge, Ameya Prabhu

Main category: cs.LG

TL;DR: The paper proposes a sample-wise framework to measure knowledge forgetting and backward transfer during language model post-training, revealing that different post-training methods have varying effects on pretrained knowledge.

DetailsMotivation: Scaled post-training drives major capability gains in language models, but its impact on pretrained knowledge remains poorly understood. Traditional task averages conflate forgetting and backward transfer effects, obscuring important changes.

Method: Proposes a sample-wise paradigm that counts 1->0 transitions (forgetting) and 0->1 transitions (backward transfer). For multiple-choice benchmarks, adds chance-adjusted variants to subtract random guessing effects. Applies framework across post-training stages, model sizes, and data scales.

Result: Domain-continual pretraining causes moderate forgetting with low-to-moderate backward transfer. RL/SFT post-training on base models yields moderate-to-large backward transfer on math/logic with low-to-moderate forgetting. RL/SFT on instruction-tuned models is scale-sensitive. Model merging doesn’t reliably mitigate forgetting.

Conclusion: The framework provides a practical yardstick for understanding how post-training alters pretrained knowledge, enabling progress toward generally capable AI systems by better tracking knowledge changes.

Abstract: Scaled post-training now drives many of the largest capability gains in language models (LMs), yet its effect on pretrained knowledge remains poorly understood. Not all forgetting is equal: Forgetting one fact (e.g., a U.S. president or an API call) does not “average out” by recalling another. Hence, we propose a sample-wise paradigm to measure what is forgotten and when backward transfer occurs. Our metric counts 1->0 transitions (correct before post-training, incorrect after) to quantify forgetting and 0->1 transitions to quantify backward transfer. Traditional task averages conflate these effects and obscure large changes. For multiple-choice benchmarks, we add chance-adjusted variants that subtract the expected contribution of random guessing from pre- and post-training accuracies. We apply this framework across post-training stages, model sizes, and data scales. Our large-scale analysis shows that: (1) Domain-continual pretraining induces moderate forgetting with low-to-moderate backward transfer; (2) RL/SFT post-training applied to base models and Instruction tuning yields moderate-to-large backward transfer on math and logic with overall low-to-moderate forgetting; (3) Applying RL/SFT to instruction-tuned models is sensitive on data scale: at small scales, both forgetting and backward transfer are small; at larger scales, effects are mixed and warrant further study with better controls; (4) Model merging does not reliably mitigate forgetting. Overall, our framework offers a practical yardstick for mapping how post-training alters pretrained knowledge at scale – enabling progress towards generally capable AI systems.

[725] Finding Manifolds With Bilinear Autoencoders

Thomas Dooms, Ward Gauderis

Main category: cs.LG

TL;DR: This paper introduces bilinear autoencoders for decomposing neural network representations into quadratic polynomials, enabling analysis without input dependency and improving interpretability through importance ordering, clustering, and sparsity.

DetailsMotivation: Standard sparse autoencoders have interpretation limitations as they depend on inputs, making isolated study incomplete. Polynomials provide algebraic primitives that can be analyzed independently of inputs and can describe various structures from linear concepts to complex manifolds.

Method: The authors use bilinear autoencoders to efficiently decompose representations into quadratic polynomials. They discuss improvements that induce importance ordering, clustering, and activation sparsity in the learned representations.

Result: The work presents an initial step toward creating nonlinear yet analyzable latent representations through their algebraic properties, moving beyond input-dependent interpretations.

Conclusion: This approach represents progress toward developing interpretable latent representations that can be studied algebraically without input dependency, with potential applications in understanding complex neural network structures.

Abstract: Sparse autoencoders are a standard tool for uncovering interpretable latent representations in neural networks. Yet, their interpretation depends on the inputs, making their isolated study incomplete. Polynomials offer a solution; they serve as algebraic primitives that can be analysed without reference to input and can describe structures ranging from linear concepts to complicated manifolds. This work uses bilinear autoencoders to efficiently decompose representations into quadratic polynomials. We discuss improvements that induce importance ordering, clustering, and activation sparsity. This is an initial step toward nonlinear yet analysable latents through their algebraic properties.

[726] ProtoMol: Enhancing Molecular Property Prediction via Prototype-Guided Multimodal Learning

Yingxu Wang, Kunyu Zhang, Jiaxin Huang, Nan Yin, Siwei Liu, Eran Segal

Main category: cs.LG

TL;DR: ProtoMol is a prototype-guided multimodal framework that enables fine-grained integration and consistent semantic alignment between molecular graphs and textual descriptions through hierarchical encoders, layer-wise cross-modal attention, and a shared prototype space.

DetailsMotivation: Existing multimodal molecular representation learning methods suffer from limited cross-modal interaction (only at final encoder layer) and lack unified prototype space for robust alignment between modalities, which overlooks hierarchical semantic dependencies.

Method: Uses dual-branch hierarchical encoders (Graph Neural Networks for molecular graphs and Transformers for text), layer-wise bidirectional cross-modal attention mechanism, and constructs shared prototype space with learnable class-specific anchors.

Result: Extensive experiments on multiple benchmark datasets demonstrate that ProtoMol consistently outperforms state-of-the-art baselines across various molecular property prediction tasks.

Conclusion: ProtoMol effectively addresses limitations of existing multimodal methods by enabling fine-grained integration and consistent semantic alignment between molecular graphs and textual descriptions.

Abstract: Multimodal molecular representation learning, which jointly models molecular graphs and their textual descriptions, enhances predictive accuracy and interpretability by enabling more robust and reliable predictions of drug toxicity, bioactivity, and physicochemical properties through the integration of structural and semantic information. However, existing multimodal methods suffer from two key limitations: (1) they typically perform cross-modal interaction only at the final encoder layer, thus overlooking hierarchical semantic dependencies; (2) they lack a unified prototype space for robust alignment between modalities. To address these limitations, we propose ProtoMol, a prototype-guided multimodal framework that enables fine-grained integration and consistent semantic alignment between molecular graphs and textual descriptions. ProtoMol incorporates dual-branch hierarchical encoders, utilizing Graph Neural Networks to process structured molecular graphs and Transformers to encode unstructured texts, resulting in comprehensive layer-wise representations. Then, ProtoMol introduces a layer-wise bidirectional cross-modal attention mechanism that progressively aligns semantic features across layers. Furthermore, a shared prototype space with learnable, class-specific anchors is constructed to guide both modalities toward coherent and discriminative representations. Extensive experiments on multiple benchmark datasets demonstrate that ProtoMol consistently outperforms state-of-the-art baselines across a variety of molecular property prediction tasks.

[727] DrivAerStar: An Industrial-Grade CFD Dataset for Vehicle Aerodynamic Optimization

Jiyan Qiu, Lyulin Kuang, Guan Wang, Yichen Xu, Leiyao Cui, Shaotong Fu, Yixin Zhu, Ruihua Zhang

Main category: cs.LG

TL;DR: DrivAerStar is a dataset of 12,000 industrial-grade automotive CFD simulations that bridges academic ML research and industrial CFD practice, achieving wind tunnel validation accuracy below 1.04% for vehicle aerodynamics optimization.

DetailsMotivation: Traditional vehicle aerodynamics optimization faces a trade-off between computationally expensive CFD simulations (weeks per iteration) and simplified models that sacrifice accuracy. Existing ML datasets have limitations preventing industrial deployment.

Method: Generated 12,000 industrial-grade CFD simulations using STAR-CCM+ software, systematically exploring three vehicle configurations through 20 CAD parameters via Free Form Deformation algorithms, including complete engine compartments and cooling systems with realistic internal airflow.

Result: Achieved wind tunnel validation accuracy below 1.04% - a five-fold improvement over existing datasets. Models trained on this data achieve production-ready accuracy while reducing computational costs from weeks to minutes.

Conclusion: DrivAerStar establishes a new standard for data-driven aerodynamic optimization in automotive development and demonstrates a paradigm for integrating high-fidelity physics simulations with AI across engineering disciplines.

Abstract: Vehicle aerodynamics optimization has become critical for automotive electrification, where drag reduction directly determines electric vehicle range and energy efficiency. Traditional approaches face an intractable trade-off: computationally expensive Computational Fluid Dynamics (CFD) simulations requiring weeks per design iteration, or simplified models that sacrifice production-grade accuracy. While machine learning offers transformative potential, existing datasets exhibit fundamental limitations – inadequate mesh resolution, missing vehicle components, and validation errors exceeding 5% – preventing deployment in industrial workflows. We present DrivAerStar, comprising 12,000 industrial-grade automotive CFD simulations generated using $\text{STAR-CCM+}^\unicode{xAE}$ software. The dataset systematically explores three vehicle configurations through 20 Computer Aided Design (CAD) parameters via Free Form Deformation (FFD) algorithms, including complete engine compartments and cooling systems with realistic internal airflow. DrivAerStar achieves wind tunnel validation accuracy below 1.04% – a five-fold improvement over existing datasets – through refined mesh strategies with strict wall $y^+$ control. Benchmarks demonstrate that models trained on this data achieve production-ready accuracy while reducing computational costs from weeks to minutes. This represents the first dataset bridging academic machine learning research and industrial CFD practice, establishing a new standard for data-driven aerodynamic optimization in automotive development. Beyond automotive applications, DrivAerStar demonstrates a paradigm for integrating high-fidelity physics simulations with Artificial Intelligence (AI) across engineering disciplines where computational constraints currently limit innovation.

[728] Fly-CL: A Fly-Inspired Framework for Enhancing Efficient Decorrelation and Reduced Training Time in Pre-trained Model-based Continual Representation Learning

Heming Zou, Yunliang Zang, Wutong Xu, Xiangyang Ji

Main category: cs.LG

TL;DR: Fly-CL is a bio-inspired continual learning framework that uses similarity matching with pretrained models to prevent catastrophic forgetting, achieving SOTA performance with reduced training time by resolving multicollinearity issues.

DetailsMotivation: To address catastrophic forgetting in continual learning while overcoming multicollinearity problems in similarity matching and computational limitations of advanced methods for real-time applications.

Method: Proposes Fly-CL, a bio-inspired framework based on fly olfactory circuit, compatible with various pretrained backbones, using similarity matching with progressive multicollinearity resolution.

Result: Substantially reduces training time while achieving comparable or better performance than SOTA methods across diverse network architectures and data regimes.

Conclusion: Fly-CL effectively addresses continual learning challenges through biologically inspired design, offering efficient similarity matching with theoretical guarantees on multicollinearity resolution.

Abstract: Using a nearly-frozen pretrained model, the continual representation learning paradigm reframes parameter updates as a similarity-matching problem to mitigate catastrophic forgetting. However, directly leveraging pretrained features for downstream tasks often suffers from multicollinearity in the similarity-matching stage, and more advanced methods can be computationally prohibitive for real-time, low-latency applications. Inspired by the fly olfactory circuit, we propose Fly-CL, a bio-inspired framework compatible with a wide range of pretrained backbones. Fly-CL substantially reduces training time while achieving performance comparable to or exceeding that of current state-of-the-art methods. We theoretically show how Fly-CL progressively resolves multicollinearity, enabling more effective similarity matching with low time complexity. Extensive simulation experiments across diverse network architectures and data regimes validate Fly-CL’s effectiveness in addressing this challenge through a biologically inspired design. Code is available at https://github.com/gfyddha/Fly-CL.

[729] Adaptive Online Learning with LSTM Networks for Energy Price Prediction

Salih Salihoglu, Ibrahim Ahmed, Afshin Asadi

Main category: cs.LG

TL;DR: This paper develops an LSTM-based model with a novel custom loss function and online learning approach to forecast day-ahead electricity prices in California, achieving improved accuracy and adaptability.

DetailsMotivation: Accurate electricity price prediction is crucial for energy market stakeholders including grid operators, producers, and consumers to make informed decisions in dynamic markets.

Method: Uses LSTM networks with historical price data, weather conditions, and energy generation mix features. Introduces a custom loss function combining MAE, Jensen-Shannon Divergence, and smoothness penalty. Implements online learning for incremental adaptation to new data.

Result: The custom loss function improved model performance, aligning predictions more closely with actual values especially during peak intervals. The online learning model outperformed others with lower prediction error and variability. Energy generation mix inclusion enhanced predictive capabilities.

Conclusion: The research provides a robust framework for electricity price forecasting with comprehensive feature integration, offering valuable tools for better decision-making in dynamic electricity markets.

Abstract: Accurate prediction of electricity prices is crucial for stakeholders in the energy market, particularly for grid operators, energy producers, and consumers. This study focuses on developing a predictive model leveraging Long Short-Term Memory (LSTM) networks to forecast day-ahead electricity prices in the California energy market. The model incorporates a variety of features, including historical price data, weather conditions, and the energy generation mix. A novel custom loss function that integrates Mean Absolute Error (MAE), Jensen-Shannon Divergence (JSD), and a smoothness penalty is introduced to enhance the prediction accuracy and interpretability. Additionally, an online learning approach is implemented to allow the model to adapt to new data incrementally, ensuring continuous relevance and accuracy. The results demonstrate that the custom loss function can improve the model’s performance, aligning predicted prices more closely with actual values, particularly during peak intervals. Also, the online learning model outperforms other models by effectively incorporating real-time data, resulting in lower prediction error and variability. The inclusion of the energy generation mix further enhances the model’s predictive capabilities, highlighting the importance of comprehensive feature integration. This research provides a robust framework for electricity price forecasting, offering valuable insights and tools for better decision-making in dynamic electricity markets.

[730] UniGTE: Unified Graph-Text Encoding for Zero-Shot Generalization across Graph Tasks and Domains

Duo Wang, Yuan Zuo, Guangyue Lu, Junjie Wu

Main category: cs.LG

TL;DR: UniGTE is an instruction-tuned encoder-decoder framework that combines graph structure with LLM semantics for zero-shot graph reasoning across diverse tasks without task-specific fine-tuning.

DetailsMotivation: To address the limitations of conventional GNNs (fixed label space) and LLMs (struggle with graph structure) in generalizing to unseen graph tasks without task-specific supervision.

Method: Uses an encoder with learnable alignment tokens and structure-aware graph-text attention to create task-aware graph representations, and a frozen LLM decoder that predicts answers while reconstructing the input graph through natural language paraphrasing.

Result: Achieves state-of-the-art zero-shot results on node classification, link prediction, graph classification, and graph regression across cross-task and cross-domain settings.

Conclusion: Tight integration of graph structure with LLM semantics enables robust, transferable graph reasoning that generalizes well to unseen tasks and domains.

Abstract: Generalizing to unseen graph tasks without task-specific supervision is challenging: conventional graph neural networks are typically tied to a fixed label space, while large language models (LLMs) struggle to capture graph structure. We introduce UniGTE, an instruction-tuned encoder-decoder framework that unifies structural and semantic reasoning. The encoder augments a pretrained autoregressive LLM with learnable alignment tokens and a structure-aware graph-text attention mechanism, enabling it to attend jointly to a tokenized graph and a natural-language task prompt while remaining permutation-invariant to node order. This yields compact, task-aware graph representations. Conditioned solely on these representations, a frozen LLM decoder predicts and reconstructs: it outputs the task answer and simultaneously paraphrases the input graph in natural language. The reconstruction objective regularizes the encoder to preserve structural cues. UniGTE is instruction-tuned on five datasets spanning node-level, edge-level, and graph-level tasks across diverse domains, yet requires no fine-tuning at inference. It achieves new state-of-the-art zero-shot results on node classification, link prediction, graph classification, and graph regression under cross-task and cross-domain settings, demonstrating that tight integration of graph structure with LLM semantics enables robust, transferable graph reasoning.

[731] SNOMED CT-powered Knowledge Graphs for Structured Clinical Data and Diagnostic Reasoning

Dun Liu, Qin Pang, Guangai Liu, Hongyu Mou, Jipeng Fan, Yiming Miao, Pin-Han Ho, Limei Peng

Main category: cs.LG

TL;DR: A knowledge-driven framework using SNOMED CT and Neo4j graph database to structure clinical data, improving AI diagnostic reasoning consistency.

DetailsMotivation: Unstructured clinical documentation creates noisy, inconsistent training data that hinders AI effectiveness in healthcare.

Method: Integrate SNOMED CT with Neo4j to build medical knowledge graph with clinical entities as nodes and semantic relationships as edges, then fine-tune LLMs with structured datasets.

Result: Significantly improves clinical logic consistency of AI outputs, enhancing validity and interpretability of diagnostic reasoning.

Conclusion: Provides scalable solution for building reliable AI-assisted clinical systems through knowledge-guided approach.

Abstract: The effectiveness of artificial intelligence (AI) in healthcare is significantly hindered by unstructured clinical documentation, which results in noisy, inconsistent, and logically fragmented training data. To address this challenge, we present a knowledge-driven framework that integrates the standardized clinical terminology SNOMED CT with the Neo4j graph database to construct a structured medical knowledge graph. In this graph, clinical entities such as diseases, symptoms, and medications are represented as nodes, and semantic relationships such as caused by,'' treats,’’ and ``belongs to’’ are modeled as edges in Neo4j, with types mapped from formal SNOMED CT relationship concepts (e.g., \texttt{Causative agent}, \texttt{Indicated for}). This design enables multi-hop reasoning and ensures terminological consistency. By extracting and standardizing entity-relationship pairs from clinical texts, we generate structured, JSON-formatted datasets that embed explicit diagnostic pathways. These datasets are used to fine-tune large language models (LLMs), significantly improving the clinical logic consistency of their outputs. Experimental results demonstrate that our knowledge-guided approach enhances the validity and interpretability of AI-generated diagnostic reasoning, providing a scalable solution for building reliable AI-assisted clinical systems.

[732] DeepChem Equivariant: SE(3)-Equivariant Support in an Open-Source Molecular Machine Learning Library

Jose Siguenza, Bharath Ramsundar

Main category: cs.LG

TL;DR: The paper extends DEEPCHEM with SE(3)-equivariant neural network support, providing ready-to-use models and complete training pipelines for molecular applications.

DetailsMotivation: Existing SE(3)-equivariant neural network libraries require substantial deep learning or mathematical knowledge and lack complete training pipelines, making them inaccessible to scientists with minimal background.

Method: Extend DEEPCHEM framework with support for SE(3)-equivariant models including SE(3)-Transformer and Tensor Field Networks, providing complete training pipelines and equivariant utilities.

Result: Created an implementation with equivariant models, training pipelines, and toolkit of equivariant utilities, supported by comprehensive tests and documentation.

Conclusion: The extension enables scientists with minimal deep learning background to build, train, and evaluate SE(3)-equivariant models, facilitating both application and further development in molecular applications.

Abstract: Neural networks that incorporate geometric relationships respecting SE(3) group transformations (e.g. rotations and translations) are increasingly important in molecular applications, such as molecular property prediction, protein structure modeling, and materials design. These models, known as SE(3)-equivariant neural networks, ensure outputs transform predictably with input coordinate changes by explicitly encoding spatial atomic positions. Although libraries such as E3NN [4] and SE(3)-TRANSFORMER [3 ] offer powerful implementations, they often require substantial deep learning or mathematical prior knowledge and lack complete training pipelines. We extend DEEPCHEM [ 13] with support for ready-to-use equivariant models, enabling scientists with minimal deep learning background to build, train, and evaluate models, such as SE(3)-Transformer and Tensor Field Networks. Our implementation includes equivariant models, complete training pipelines, and a toolkit of equivariant utilities, supported with comprehensive tests and documentation, to facilitate both application and further development of SE(3)-equivariant models.

[733] A Lightweight DL Model for Smart Grid Power Forecasting with Feature and Resolution Mismatch

Sarah Al-Shareeda, Gulcihan Ozdemir, Heung Seok Jeon, Khaleel Ahmad

Main category: cs.LG

TL;DR: A lightweight DL pipeline combining preprocessing techniques and GRU-LSTM model achieves accurate short-term energy consumption forecasting despite noisy, incomplete sensor data.

DetailsMotivation: To address the challenge of accurate short-term energy consumption forecasting when sensor data is noisy, incomplete, and lacks contextual richness, particularly for real-world high-frequency data.

Method: Proposed a robust DL pipeline with hourly downsizing, dual-mode imputation (mean and polynomial regression), comprehensive normalization (Standard Scaling), and a lightweight GRU-LSTM sequence-to-one model.

Result: Achieved average RMSE of 601.9W, MAE of 468.9W, and 84.36% accuracy. The model generalized well, captured nonlinear demand patterns, maintained low inference latency, and showed strong alignment between temperature trends and predicted consumption.

Conclusion: Targeted preprocessing paired with compact recurrent architectures enables fast, accurate, and deployment-ready energy forecasting in real-world conditions.

Abstract: How can short-term energy consumption be accurately forecasted when sensor data is noisy, incomplete, and lacks contextual richness? This question guided our participation in the \textit{2025 Competition on Electric Energy Consumption Forecast Adopting Multi-criteria Performance Metrics}, which challenged teams to predict next-day power demand using real-world high-frequency data. We proposed a robust yet lightweight Deep Learning (DL) pipeline combining hourly downsizing, dual-mode imputation (mean and polynomial regression), and comprehensive normalization, ultimately selecting Standard Scaling for optimal balance. The lightweight GRU-LSTM sequence-to-one model achieves an average RMSE of 601.9W, MAE of 468.9W, and 84.36% accuracy. Despite asymmetric inputs and imputed gaps, it generalized well, captured nonlinear demand patterns, and maintained low inference latency. Notably, spatiotemporal heatmap analysis reveals a strong alignment between temperature trends and predicted consumption, further reinforcing the model’s reliability. These results demonstrate that targeted preprocessing paired with compact recurrent architectures can still enable fast, accurate, and deployment-ready energy forecasting in real-world conditions.

[734] Domain Generalizable Continual Learning

Hongwei Yan, Guanglong Sun, Zhiqi Kang, Yi Zhong, Liyuan Wang

Main category: cs.LG

TL;DR: The paper introduces Domain Generalizable Continual Learning (DGCL), a setting where models learn sequential tasks from single domains and must generalize across all domains. It proposes Adaptive Domain Transformation (DoT), a plug-in method that disentangles semantic and domain information for better generalization.

DetailsMotivation: Current continual learning methods assume identical training and testing domains, but real-world environments require models to generalize across diverse, unseen domains while learning new skills sequentially.

Method: Proposes Adaptive Domain Transformation (DoT), inspired by brain theory, which disentangles semantic- and domain-relevant information and adaptively transforms task representations across domains for output alignment.

Result: DoT significantly improves state-of-the-art continual learning baselines in DGCL settings under both full parameter tuning and parameter-efficient tuning, while being resource-efficient with lightweight implementation.

Conclusion: DoT effectively addresses DGCL challenges by accumulating domain-generalizable knowledge and ensuring balanced, generalized predictions across sequential tasks and domains.

Abstract: To adapt effectively to dynamic real-world environments, intelligent systems must continually acquire new skills while generalizing them to diverse, unseen scenarios. Here, we introduce a novel and realistic setting named domain generalizable continual learning (DGCL): a model learns sequential tasks with each involving a single domain, aiming to perform well across all encountered tasks and domains. This setting poses unique challenges in acquiring, retaining, and leveraging both semantic- and domain-relevant information for robust generalization. Although state-of-the-art continual learning (CL) methods have employed pre-trained models (PTMs) to enhance task-specific generalization, they typically assume identical training and testing domains for each task and therefore perform poorly in DGCL. To this end, we propose adaptive Domain Transformation (DoT), an innovative PTMs-based approach tailored to DGCL. Inspired by the distributed-plus-hub theory of the human brain, DoT disentangles semantic- and domain-relevant information in representation learning, and adaptively transforms task representations across various domains for output alignment, ensuring balanced and generalized predictions. DoT serves as a plug-in strategy that greatly facilitates state-of-the-art CL baselines under both full parameter tuning and parameter-efficient tuning paradigms in DGCL, validated by extensive experiments. Also, DoT is shown to accumulate domain-generalizable knowledge from DGCL, and ensure resource efficiency with a lightweight implementation.

[735] SolverLLM: Leveraging Test-Time Scaling for Optimization Problem via LLM-Guided Search

Dong Li, Xujiang Zhao, Linlin Yu, Yanchi Liu, Wei Cheng, Zhengzhang Chen, Zhong Chen, Feng Chen, Chen Zhao, Haifeng Chen

Main category: cs.LG

TL;DR: SolverLLM is a training-free framework that uses test-time scaling and MCTS to solve optimization problems by generating mathematical formulations and solver code, outperforming existing methods.

DetailsMotivation: Existing LLM methods for optimization either rely on prompt engineering (poor generalization) or supervised training (costly). There's a need for a training-free approach that generalizes well across problem types.

Method: Uses Monte Carlo Tree Search (MCTS) with three modifications: dynamic expansion for adaptive formulation generation, prompt backpropagation for outcome-driven feedback, and uncertainty backpropagation to incorporate reward reliability.

Result: Outperforms both prompt-based and learning-based baselines on six standard benchmark datasets, achieving strong generalization without additional training.

Conclusion: SolverLLM demonstrates that test-time scaling with enhanced MCTS can effectively solve diverse optimization problems without requiring costly training, offering better generalization than existing approaches.

Abstract: Large Language Models (LLMs) offer promising capabilities for tackling complex reasoning tasks, including optimization problems. However, existing methods either rely on prompt engineering, which leads to poor generalization across problem types, or require costly supervised training. We introduce SolverLLM, a training-free framework that leverages test-time scaling to solve diverse optimization problems. Rather than solving directly, SolverLLM generates mathematical formulations and translates them into solver-ready code, guided by a novel Monte Carlo Tree Search (MCTS) strategy. To enhance the search process, we modify classical MCTS with (1) dynamic expansion for adaptive formulation generation, (2) prompt backpropagation to guide exploration via outcome-driven feedback, and (3) uncertainty backpropagation to incorporate reward reliability into decision-making. Experiments on six standard benchmark datasets demonstrate that SolverLLM outperforms both prompt-based and learning-based baselines, achieving strong generalization without additional training.

[736] Closing the Curvature Gap: Full Transformer Hessians and Their Implications for Scaling Laws

Egor Petrov, Nikita Kiselev, Vladislav Meshkov, Andrey Grabovoy

Main category: cs.LG

TL;DR: This paper completes the Hessian characterization of full Transformer blocks by deriving explicit second-order expressions for Layer Normalization and feedforward components, generalizing prior self-attention analyses.

DetailsMotivation: The lack of theoretical results for Layer Normalization and feedforward Hessians has left a gap in understanding Transformer optimization landscapes, which this work aims to address.

Method: The authors derive explicit second-order expressions for Layer Normalization and feedforward components, and propose a Taylor-expansion-based framework for analyzing loss differences to quantify convergence trajectories.

Result: The work yields estimations for the role of each sublayer in curvature propagation and demonstrates how Hessian structures inform both convergence dynamics and empirical scaling laws for large-model performance.

Conclusion: By extending Hessian theory to the full Transformer architecture, this work establishes a new foundation for theoretical and empirical investigations of optimization in large-scale deep learning.

Abstract: The lack of theoretical results for Layer Normalization and feedforward Hessians has left a gap in the study of Transformer optimization landscapes. We address this by deriving explicit second-order expressions for these components, thereby completing the Hessian characterization of full Transformer blocks. Our results generalize prior self-attention analyses and yield estimations for the role of each sublayer in curvature propagation. We demonstrate how these Hessian structures inform both convergence dynamics and the empirical scaling laws governing large-model performance. Further, we propose a Taylor-expansion-based framework for analyzing loss differences to quantify convergence trajectories. By extending Hessian theory to the full Transformer architecture, this work establishes a new foundation for theoretical and empirical investigations of optimization in large-scale deep learning.

[737] A Primer on Kolmogorov-Arnold Networks (KANs) for Probabilistic Time Series Forecasting

Cristian J. Vaca-Rubio, Roberto Pereira, Luis Blanco, Engin Zeydan, Màrius Caus

Main category: cs.LG

TL;DR: P-KAN is a probabilistic extension of Kolmogorov-Arnold Networks that replaces scalar weights with spline-based functional connections for time series forecasting, offering parameter-efficient uncertainty-aware predictions.

DetailsMotivation: To develop expressive yet parameter-efficient probabilistic models for time series forecasting that can capture nonlinear and heavy-tailed dynamics, particularly for satellite traffic forecasting where uncertainty-aware predictions enable dynamic resource allocation.

Method: Replace scalar weights in KANs with spline-based functional connections and directly parameterize predictive distributions using Gaussian and Student-t distributions to model uncertainty.

Result: P-KANs consistently outperform MLP baselines in both accuracy and calibration, achieving superior efficiency-risk trade-offs while using significantly fewer parameters. Gaussian variant provides conservative forecasts for safety-critical scenarios, while Student-t variant yields sharper distributions for stable demand.

Conclusion: P-KANs establish a powerful framework for probabilistic forecasting with direct applicability to satellite communications and other resource-constrained domains, offering parameter-efficient uncertainty modeling.

Abstract: This work introduces Probabilistic Kolmogorov-Arnold Network (P-KAN), a novel probabilistic extension of Kolmogorov-Arnold Networks (KANs) for time series forecasting. By replacing scalar weights with spline-based functional connections and directly parameterizing predictive distributions, P-KANs offer expressive yet parameter-efficient models capable of capturing nonlinear and heavy-tailed dynamics. We evaluate P-KANs on satellite traffic forecasting, where uncertainty-aware predictions enable dynamic thresholding for resource allocation. Results show that P-KANs consistently outperform Multi Layer Perceptron (MLP) baselines in both accuracy and calibration, achieving superior efficiency-risk trade-offs while using significantly fewer parameters. We build up P-KANs on two distributions, namely Gaussian and Student-t distributions. The Gaussian variant provides robust, conservative forecasts suitable for safety-critical scenarios, whereas the Student-t variant yields sharper distributions that improve efficiency under stable demand. These findings establish P-KANs as a powerful framework for probabilistic forecasting with direct applicability to satellite communications and other resource-constrained domains.

[738] Quantile Regression, Variational Autoencoders, and Diffusion Models for Uncertainty Quantification: A Spatial Analysis of Sub-seasonal Wind Speed Prediction

Ganglin Tian, Anastase Alexandre Charantonis, Camille Le Coz, Alexis Tantet, Riwal Plougonven

Main category: cs.LG

TL;DR: Probabilistic deep learning methods improve spatial uncertainty representation in sub-seasonal wind speed forecasting compared to simpler stochastic approaches.

DetailsMotivation: To enhance spatial representation of uncertainties when downscaling surface wind speeds from large-scale atmospheric predictors for sub-seasonal forecasting, addressing limitations of previous stochastic perturbation methods that fail to capture spatial correlations and physical consistency.

Method: Evaluated three probabilistic deep learning methods with distinct uncertainty quantification mechanisms: Quantile Regression Neural Network (direct quantile modeling), Variational Autoencoders (latent space sampling), and Diffusion Models (iterative denoising). Trained on ERA5 reanalysis data and applied to ECMWF sub-seasonal hindcasts.

Result: Probabilistic downscaling approaches provide more realistic spatial uncertainty representations compared to simpler stochastic methods, with each model offering different strengths in ensemble dispersion, deterministic skill, and physical consistency.

Conclusion: Probabilistic downscaling establishes as an effective enhancement to operational sub-seasonal wind forecasts for renewable energy planning and risk assessment.

Abstract: This study aims to improve the spatial representation of uncertainties when regressing surface wind speeds from large-scale atmospheric predictors for sub-seasonal forecasting. Sub-seasonal forecasting often relies on large-scale atmospheric predictors such as 500 hPa geopotential height (Z500), which exhibit higher predictability than surface variables and can be downscaled to obtain more localised information. Previous work by Tian et al. (2024) demonstrated that stochastic perturbations based on model residuals can improve ensemble dispersion representation in statistical downscaling frameworks, but this method fails to represent spatial correlations and physical consistency adequately. More sophisticated approaches are needed to capture the complex relationships between large-scale predictors and local-scale predictands while maintaining physical consistency. Probabilistic deep learning models offer promising solutions for capturing complex spatial dependencies. This study evaluates three probabilistic methods with distinct uncertainty quantification mechanisms: Quantile Regression Neural Network that directly models distribution quantiles, Variational Autoencoders that leverage latent space sampling, and Diffusion Models that utilise iterative denoising. These models are trained on ERA5 reanalysis data and applied to ECMWF sub-seasonal hindcasts to regress probabilistic wind speed ensembles. Our results show that probabilistic downscaling approaches provide more realistic spatial uncertainty representations compared to simpler stochastic methods, with each probabilistic model offering different strengths in terms of ensemble dispersion, deterministic skill, and physical consistency. These findings establish probabilistic downscaling as an effective enhancement to operational sub-seasonal wind forecasts for renewable energy planning and risk assessment.

[739] Justitia: Fair and Efficient Scheduling for LLM Applications

Mingyan Yang, Guanjie Wang, Manqi Luo, Yifei Liu, Chen Chen, Han Zhao, Yu Feng, Quan Chen, Minyi Guo

Main category: cs.LG

TL;DR: Justitia is a novel scheduler for LLM applications that improves scheduling efficiency while maintaining fairness by using memory-centric cost modeling, neural network demand prediction, and virtual-time fair queuing.

DetailsMotivation: Current LLM schedulers suffer from head-of-line blocking and over-constrained resource allocation, leading to poor performance for LLM applications in shared GPU servers.

Method: Three key techniques: memory-centric service cost modeling for LLM applications, lightweight neural network for accurate demand prediction, and virtual-time based fair queuing algorithm.

Result: Experimental results show Justitia substantially enhances scheduling efficiency while preserving fairness across diverse LLM applications.

Conclusion: Justitia provides an effective solution for serving LLM applications in shared GPU environments by balancing efficiency and fairness through its three core techniques.

Abstract: In the era of Large Language Models (LLMs), it has been popular to launch a series of LLM inferences – we call an LLM application – to better solve real-world problems. When serving those applications in shared GPU servers, the schedulers are expected to attain fast application completions with guaranteed worst-case performance. However, mainstream LLM schedulers fail to behave well for LLM applications – due to head-of-line blocking or over-constrained resource allocation. In this paper, we propose to serve LLM applications in a fair and also efficient manner. To this end, we design Justitia, a novel scheduler with three key techniques. First, given that memory is prevalently a bottleneck for mainstream inference frameworks like vLLM, Justitia models the service cost of LLM applications in a memory-centric manner. Meanwhile, it uses a simple neural network model to conduct light-weight and also accurate demand prediction. Moreover, Justitia adopts a virtual-time based fair queuing algorithm to reduce the overall performance with guaranteed worst-case delay. We have implemented Justitia atop vLLM, and experimental results involving diverse LLM applications show that it can substantially enhance the scheduling efficiency with fairness preserved.

[740] Differentially Private Linear Regression and Synthetic Data Generation with Statistical Guarantees

Shurong Lin, Aleksandra Slavković, Deekshith Reddy Bhoomireddy

Main category: cs.LG

TL;DR: Proposes a differentially private linear regression method with valid inference and synthetic data generation for small-to-medium social science datasets, addressing limitations of existing approaches.

DetailsMotivation: Social science datasets are often small-to-medium scale and use linear regression, but current differentially private methods focus on point estimation without uncertainty quantification or synthetic data generation support. Existing synthetic data approaches are unsuitable for continuous regression data.

Method: Uses Gaussian differential privacy with a bias-corrected estimator and asymptotic confidence intervals, plus a synthetic data generation procedure where regression on synthetic data matches the DP regression. Employs binning-aggregation strategy for small-to-moderate dimensions.

Result: Experiments show improved accuracy over existing methods, valid confidence intervals, and more reliable synthetic data for downstream machine learning tasks compared to current DP synthetic data generation approaches.

Conclusion: The method effectively addresses the gap in differentially private linear regression by providing both valid inference and synthetic data generation capabilities suitable for small-to-medium social science datasets with continuous variables.

Abstract: In social sciences, small- to medium-scale datasets are common and linear regression (LR) is canonical. In privacy-aware settings, much work has focused on differentially private (DP) LR, but mostly on point estimation with limited attention to uncertainty quantification. Meanwhile, synthetic data generation (SDG) is increasingly important for reproducibility studies, yet current DP LR methods do not readily support it. Mainstream SDG approaches are either tailored to discretized data, making them less suitable for continuous regression, or rely on deep models that require large datasets, limiting their use for the smaller, continuous data typical in social science. We propose a method for LR with valid inference under Gaussian DP: a DP bias-corrected estimator with asymptotic confidence intervals (CIs) and a general SDG procedure in which regression on the synthetic data matches our DP regression. Our binning-aggregation strategy is effective in small- to moderate-dimensional settings. Experiments show our method (1) improves accuracy over existing methods, (2) provides valid CIs, and (3) produces more reliable synthetic data for downstream ML tasks than current DP SDGs.

[741] Curiosity-driven RL for symbolic equation solving

Kevin P. O Keeffe

Main category: cs.LG

TL;DR: RL with PPO, curiosity exploration, and graph-based actions can solve nonlinear equations including radicals, exponentials, and trig functions.

DetailsMotivation: To explore if reinforcement learning (RL) can be useful for symbolic mathematics beyond linear equations.

Method: Used model-free PPO augmented with curiosity-based exploration and graph-based actions.

Result: Successfully solved nonlinear equations involving radicals, exponentials, and trigonometric functions.

Conclusion: Curiosity-based exploration may be useful for general symbolic reasoning tasks.

Abstract: We explore if RL can be useful for symbolic mathematics. Previous work showed contrastive learning can solve linear equations in one variable. We show model-free PPO \cite{schulman2017proximal} augmented with curiosity-based exploration and graph-based actions can solve nonlinear equations such as those involving radicals, exponentials, and trig functions. Our work suggests curiosity-based exploration may be useful for general symbolic reasoning tasks.

[742] Towards Interpretable and Trustworthy Time Series Reasoning: A BlueSky Vision

Kanghui Ning, Zijie Pan, Yushan Jiang, Anderson Schneider, Yuriy Nevmyvaka, Dongjin Song

Main category: cs.LG

TL;DR: This paper presents a vision for advancing time series reasoning beyond pattern recognition to explicit, interpretable inference through two complementary directions: robust foundations and system-level reasoning.

DetailsMotivation: Time series reasoning is emerging as the next frontier in temporal analysis, aiming to move beyond pattern recognition towards explicit, interpretable, and trustworthy inference.

Method: Two complementary approaches: (1) building robust foundations through comprehensive temporal understanding, structured multi-step reasoning, and faithful evaluation frameworks; (2) advancing system-level reasoning by incorporating multi-agent collaboration, multi-modal context, and retrieval-augmented approaches.

Result: The paper outlines a flexible and extensible framework for advancing time series reasoning.

Conclusion: The proposed framework aims to deliver interpretable and trustworthy temporal intelligence across diverse domains.

Abstract: Time series reasoning is emerging as the next frontier in temporal analysis, aiming to move beyond pattern recognition towards explicit, interpretable, and trustworthy inference. This paper presents a BlueSky vision built on two complementary directions. One builds robust foundations for time series reasoning, centered on comprehensive temporal understanding, structured multi-step reasoning, and faithful evaluation frameworks. The other advances system-level reasoning, moving beyond language-only explanations by incorporating multi-agent collaboration, multi-modal context, and retrieval-augmented approaches. Together, these directions outline a flexible and extensible framework for advancing time series reasoning, aiming to deliver interpretable and trustworthy temporal intelligence across diverse domains.

[743] MuonBP: Faster Muon via Block-Periodic Orthogonalization

Ahmed Khaled, Kaan Ozkara, Tao Yu, Mingyi Hong, Youngsuk Park

Main category: cs.LG

TL;DR: MuonBP improves gradient orthogonalization by applying block-periodic orthogonalization to reduce communication overhead in model parallelism while maintaining training stability.

DetailsMotivation: To address the communication overhead introduced by gradient orthogonalization in model parallelism, which causes 5%-10% throughput reduction compared to AdamW.

Method: Proposes MuonBP with block-periodic orthogonalization: applies orthogonalization independently to matrix shards on each device and periodically performs full orthogonalization, using two different learning rates for blockwise and full orthogonalization steps.

Result: Achieves 8% throughput increase compared to Muon with no performance degradation when training an 8B model with eight-way tensor parallelism and ZeRO optimizer state sharding.

Conclusion: MuonBP provides competitive iteration complexity with per-iteration throughput comparable to coordinate-wise methods like AdamW, requiring minimal hyperparameter adjustments.

Abstract: Gradient orthogonalization is a simple strategy that shows great utility in speeding up gradient descent. The Muon optimizer (Jordan, Jin, et al., 2024) combines gradient orthogonalization with first-order momentum and achieves significant improvement in data efficiency over Adam/AdamW (Loshchilov and Hutter, 2019) for language model training. However, when using model parallelism, gradient orthogonalization introduces additional overhead compared to coordinate-wise optimizers (such as AdamW) due to additional gather and scatter operations on gradient matrix shards from different devices. This additional communication can amount to a throughput hit of 5%-10% compared to Adam/AdamW. To remedy this, we propose Muon with Block-Periodic Orthogonalization (MuonBP), which applies orthogonalization independently to matrix shards on each device and periodically performs full orthogonalization to maintain training stability at scale. We show how to adjust the learning rate from the baseline to MuonBP and give convergence guarantees for this algorithm. Crucially, our theory dictates that we use two stepsizes: one for the blockwise orthogonalization steps, and one for the full orthogonalization steps. Our method is simple, requires minimal hyperparameter adjustments, and achieves competitive iteration complexity compared with baseline Muon while providing per-iteration throughput comparable to coordinate-wise methods such as AdamW. When training an 8B model with eight-way tensor parallelism and ZeRO optimizer state sharding, MuonBP achieves 8% throughput increase compared to Muon with no degradation in performance.

[744] The Ends Justify the Thoughts: RL-Induced Motivated Reasoning in LLMs

Nikolaus Howe, Micah Carroll

Main category: cs.LG

TL;DR: The paper investigates how language models engage in motivated reasoning when post-hoc instructions conflict with learned behaviors, and finds that while frontier models can detect this reasoning, smaller LLMs may fail to identify it or be persuaded by it.

DetailsMotivation: To understand what happens to models' reasoning processes when post-hoc instructions conflict with learned behaviors, particularly in the context of detecting harmful behaviors through chain-of-thought monitoring.

Method: The researchers investigate this question in simple settings, examining how models generate justifications for violating instructions and testing the capability of different LLM judges to detect motivated reasoning.

Result: Models engage in systematic motivated reasoning - generating plausible-sounding justifications while downplaying harms. While frontier reasoning models can detect most motivated reasoning, smaller LLM judges fail to identify some of it and can be persuaded by the reasoning despite clear instruction violations.

Conclusion: The capability gap in detecting motivated reasoning raises concerns that as models become more sophisticated, their motivated reasoning may become harder to detect, underscoring the need to account for this phenomenon when using chain-of-thought processes for model evaluation and oversight.

Abstract: The use of reinforcement learning (RL) with chain-of-thought (CoT) reasoning has emerged as a promising approach for developing more capable language models. In turn, this has led to investigation of CoT monitoring as a compelling method for detecting harmful behaviors such as reward hacking, under the assumption that models’ reasoning processes reflect their internal decision-making. In practice, LLM training often produces unintended behaviors due to imperfect reward signals, leading models to develop misaligned tendencies. A common corrective approach is to apply post-hoc instructions to avoid problematic behaviors like sycophancy, but what happens to the model’s reasoning process when these instructions conflict with learned behaviors? We investigate this question in simple settings and find that models engage in systematic motivated reasoning – generating plausible-sounding justifications for violating their instructions while downplaying potential harms. Beyond being an interesting property of training, we find that while motivated reasoning can be detected by most frontier reasoning models, smaller LLM judges can fail to identify a portion of it, and in rare cases can themselves be persuaded that the reasoning is correct, despite it contradicting clear instructions. This capability gap raises concerns that as models become more sophisticated, their motivated reasoning may become increasingly difficult for monitors to detect. Our results underscore the need to account for motivated reasoning when relying on chain-of-thought processes for model evaluation and oversight. All code for this paper will be made available. WARNING: some examples in this paper may be upsetting.

[745] Graph4MM: Weaving Multimodal Learning with Structural Information

Xuying Ning, Dongqi Fu, Tianxin Wei, Wujiang Xu, Jingrui He

Main category: cs.LG

TL;DR: Graph4MM is a graph-based multimodal learning framework that integrates multi-hop structural information into foundation models through Hop-Diffused Attention and MM-QFormer for cross-modal fusion.

DetailsMotivation: Real-world multimodal data has complex structural relationships beyond simple one-to-one mappings, and previous approaches fail to properly integrate multi-hop neighbors and treat graphs as standalone modalities, limiting overall understanding.

Method: Proposed Graph4MM framework with Hop-Diffused Attention (integrating multi-hop structural information via causal masking and hop diffusion) and MM-QFormer (multi-mapping querying transformer for cross-modal fusion).

Result: Graph4MM outperforms larger VLMs, LLMs, and multimodal graph baselines, achieving 6.93% average improvement on both generative and discriminative tasks.

Conclusion: Leveraging structures to integrate both intra- and inter-modal interactions improves multimodal understanding beyond treating them as standalone modalities, demonstrating the value of proper graph integration in foundation models.

Abstract: Real-world multimodal data usually exhibit complex structural relationships beyond traditional one-to-one mappings like image-caption pairs. Entities across modalities interact in intricate ways, with images and text forming diverse interconnections through contextual dependencies and co-references. Graphs provide powerful structural information for modeling intra-modal and inter-modal relationships. However, previous works fail to distinguish multi-hop neighbors and treat the graph as a standalone modality, which fragments the overall understanding. This limitation presents two key challenges in multimodal learning: (1) integrating structural information from multi-hop neighbors into foundational models, and (2) fusing modality-specific information in a principled manner. To address these challenges, we revisit the role of graphs in multimodal learning within the era of foundation models and propose Graph4MM, a graph-based multimodal learning framework. To be specific, we introduce Hop-Diffused Attention, which integrates multi-hop structural information into self-attention through causal masking and hop diffusion. Furthermore, we design MM-QFormer, a multi-mapping querying transformer for cross-modal fusion. Through theoretical and empirical analysis, we show that leveraging structures to integrate both intra- and inter-modal interactions improves multimodal understanding beyond treating them as a standalone modality. Experiments on both generative and discriminative tasks show that Graph4MM outperforms larger VLMs, LLMs, and multimodal graph baselines, achieving a 6.93% average improvement.

[746] Bitwidth-Specific Logarithmic Arithmetic for Future Hardware-Accelerated Training

Hassan Hamad, Yuou Qiu, Peter A. Beerel, Keith M. Chugg

Main category: cs.LG

TL;DR: This paper introduces a novel low-precision logarithmic fixed-point training method with hardware-friendly approximations, achieving comparable accuracy to 32-bit floating-point training while significantly reducing hardware area and energy consumption.

DetailsMotivation: While quantization has reduced inference costs, training still relies on complex floating-point arithmetic. Low-precision fixed-point training offers a compelling alternative for future hardware accelerators.

Method: Proposes hardware-friendly piece-wise linear approximation for logarithmic addition, optimized using simulated annealing at different precision levels. Uses 12-bit integer arithmetic for training neural networks.

Result: Successfully trained VGG-11 and VGG-16 models on CIFAR-100 and TinyImageNet with minimal accuracy degradation compared to 32-bit floating-point training. Achieved 32.5% area reduction and 53.5% energy reduction for LNS multiply-accumulate units.

Conclusion: The proposed low-precision logarithmic fixed-point training method enables efficient hardware implementation with significant area and energy savings while maintaining training accuracy comparable to floating-point approaches.

Abstract: While advancements in quantization have significantly reduced the computational costs of inference in deep learning, training still predominantly relies on complex floating-point arithmetic. Low-precision fixed-point training presents a compelling alternative. This work introduces a novel enhancement in low-precision logarithmic fixed-point training, geared towards future hardware accelerator designs. We propose incorporating bitwidth in the design of approximations to arithmetic operations. To this end, we introduce a new hardware-friendly, piece-wise linear approximation for logarithmic addition. Using simulated annealing, we optimize this approximation at different precision levels. A C++ bit-true simulation demonstrates training of VGG-11 and VGG-16 models on CIFAR-100 and TinyImageNet, respectively, using 12-bit integer arithmetic with minimal accuracy degradation compared to 32-bit floating-point training. Our hardware study reveals up to 32.5% reduction in area and 53.5% reduction in energy consumption for the proposed LNS multiply-accumulate units compared to that of linear fixed-point equivalents.

[747] EEschematic: Multimodal-LLM Based AI Agent for Schematic Generation of Analog Circuit

Chang Liu, Danial Chitnis

Main category: cs.LG

TL;DR: EEschematic is an AI agent that uses a Multimodal Large Language Model to automatically generate human-editable schematic diagrams from SPICE netlists, addressing the visual interpretability gap in current LLM-based circuit design approaches.

DetailsMotivation: Current LLM-based circuit design methods rely on textual SPICE netlists which lack visual interpretability for circuit designers, creating a need for schematic generation that bridges this gap.

Method: Uses a Multimodal Large Language Model integrating textual, visual, and symbolic modalities with few-shot placement using six analog substructure examples and Visual Chain-of-Thought strategy for iterative refinement of placement and wiring.

Result: Successfully generated high-quality schematics for representative analog circuits including CMOS inverter, 5T-OTA, and telescopic cascode amplifier with high visual quality and structural correctness.

Conclusion: EEschematic effectively translates SPICE netlists into human-editable schematic diagrams, providing visual interpretability while maintaining structural accuracy in analog circuit design.

Abstract: Circuit schematics play a crucial role in analog integrated circuit design, serving as the primary medium for human understanding and verification of circuit functionality. While recent large language model (LLM)-based approaches have shown promise in circuit topology generation and device sizing, most rely solely on textual representations such as SPICE netlists, which lack visual interpretability for circuit designers. To address this limitation, we propose EEschematic, an AI agent for automatic analog schematic generation based on a Multimodal Large Language Model (MLLM). EEschematic integrates textual, visual, and symbolic modalities to translate SPICE netlists into schematic diagrams represented in a human-editable format. The framework uses six analog substructure examples for few-shot placement and a Visual Chain-of-Thought (VCoT) strategy to iteratively refine placement and wiring, enhancing schematic clarity and symmetry. Experimental results on representative analog circuits, including a CMOS inverter, a five-transistor operational transconductance amplifier (5T-OTA), and a telescopic cascode amplifier, demonstrate that EEschematic produces schematics with high visual quality and structural correctness.

[748] Explainable Heterogeneous Anomaly Detection in Financial Networks via Adaptive Expert Routing

Zan Li, Rui Fan

Main category: cs.LG

TL;DR: A novel financial anomaly detection framework using adaptive graph learning and specialized expert networks that provides interpretable mechanism-specific detection, outperforming baselines by 30.8 percentage points.

DetailsMotivation: Existing financial anomaly detectors treat all anomalies uniformly with scalar scores, lacking mechanism-specific insights and interpretability needed for targeted regulatory responses. Key challenges include static graph structures, uniform detection mechanisms, and black-box outputs.

Method: Uses adaptive graph learning with four mechanism-specific expert networks, BiLSTM with self-attention for multi-scale temporal dependencies, cross-modal attention for temporal-spatial fusion, neural multi-source interpolation for dynamic graphs, and stress-modulated fusion for balancing dynamics with structural priors.

Result: Achieved 92.3% detection of 13 major events with 3.8-day lead time on 100 US equities (2017-2024), outperforming best baseline by 30.8 percentage points. Silicon Valley Bank case study showed automatic temporal mechanism identification with Price-Shock expert weight rising from 0.39 to 0.48.

Conclusion: The framework successfully embeds interpretability architecturally rather than post-hoc, providing actionable insights into anomaly mechanisms and their temporal evolution without labeled supervision.

Abstract: Financial anomalies exhibit heterogeneous mechanisms (price shocks, liquidity freezes, contagion cascades, regime shifts), but existing detectors treat all anomalies uniformly, producing scalar scores without revealing which mechanism is failing, where risks concentrate, or how to intervene. This opacity prevents targeted regulatory responses. Three unsolved challenges persist: (1) static graph structures cannot adapt when market correlations shift during regime changes; (2) uniform detection mechanisms miss type-specific signatures across multiple temporal scales while failing to integrate individual behaviors with network contagion; (3) black-box outputs provide no actionable guidance on anomaly mechanisms or their temporal evolution. We address these via adaptive graph learning with specialized expert networks that provide built-in interpretability. Our framework captures multi-scale temporal dependencies through BiLSTM with self-attention, fuses temporal and spatial information via cross-modal attention, learns dynamic graphs through neural multi-source interpolation, adaptively balances learned dynamics with structural priors via stress-modulated fusion, routes anomalies to four mechanism-specific experts, and produces dual-level interpretable attributions. Critically, interpretability is embedded architecturally rather than applied post-hoc. On 100 US equities (2017-2024), we achieve 92.3% detection of 13 major events with 3.8-day lead time, outperforming best baseline by 30.8pp. Silicon Valley Bank case study demonstrates anomaly evolution tracking: Price-Shock expert weight rose to 0.39 (33% above baseline 0.29) during closure, peaking at 0.48 (66% above baseline) one week later, revealing automatic temporal mechanism identification without labeled supervision.

[749] Hephaestus: Mixture Generative Modeling with Energy Guidance for Large-scale QoS Degradation

Nguyen Do, Bach Ngo, Youval Kashuv, Canh V. Pham, Hanghang Tong, My T. Thai

Main category: cs.LG

TL;DR: PIMMA is a self-reinforcing generative framework that solves the Quality of Service Degradation problem by synthesizing feasible solutions in latent space through three phases: Forge, Morph, and Refine.

DetailsMotivation: Addresses the gap in handling nonlinear edge-weight functions in QoSD problems, which classical combinatorial optimization and recent ML approaches fail to tackle effectively.

Method: Three-phase approach: (1) Forge uses Predictive Path-Stressing algorithm with graph learning for feasible solutions, (2) Morph employs Mixture of Conditional VAEs with energy-based model to capture solution distributions, (3) Refine uses RL with differentiable reward to generate near-optimal solutions.

Result: Outperforms classical and ML baselines on synthetic and real-world networks, especially in nonlinear cost function scenarios where traditional methods fail.

Conclusion: PIMMA provides an effective solution for QoSD problems with nonlinear edge-weight functions, demonstrating superior performance over existing approaches.

Abstract: We study the Quality of Service Degradation (QoSD) problem, in which an adversary perturbs edge weights to degrade network performance. This setting arises in both network infrastructures and distributed ML systems, where communication quality, not just connectivity, determines functionality. While classical methods rely on combinatorial optimization, and recent ML approaches address only restricted linear variants with small-size networks, no prior model directly tackles the QoSD problem under nonlinear edge-weight functions. This work proposes \PIMMA, a self-reinforcing generative framework that synthesizes feasible solutions in latent space, to fill this gap. Our method includes three phases: (1) Forge: a Predictive Path-Stressing (PPS) algorithm that uses graph learning and approximation to produce feasible solutions with performance guarantee, (2) Morph: a new theoretically grounded training paradigm for Mixture of Conditional VAEs guided by an energy-based model to capture solution feature distributions, and (3) Refine: a reinforcement learning agent that explores this space to generate progressively near-optimal solutions using our designed differentiable reward function. Experiments on both synthetic and real-world networks show that our approach consistently outperforms classical and ML baselines, particularly in scenarios with nonlinear cost functions where traditional methods fail to generalize.

[750] Diverse Influence Component Analysis: A Geometric Approach to Nonlinear Mixture Identifiability

Hoang-Son Nguyen, Xiao Fu

Main category: cs.LG

TL;DR: Diverse Influence Component Analysis (DICA) enables latent component identification from nonlinear mixtures using Jacobian Volume Maximization, achieving identifiability without auxiliary signals, independence assumptions, or sparsity requirements.

DetailsMotivation: To address the challenge of identifying latent components from unknown nonlinear mixtures, which is fundamental for disentangled representation learning and causal inference, without relying on auxiliary signals or restrictive assumptions.

Method: Proposes DICA framework with Jacobian Volume Maximization (J-VolMax) criterion that exploits the convex geometry of the mixing function’s Jacobian to encourage diversity in latent components’ influence on observed variables.

Result: The approach achieves identifiability of latent components under reasonable conditions without requiring auxiliary information, latent component independence, or Jacobian sparsity assumptions.

Conclusion: DICA extends the scope of identifiability analysis in nonlinear ICA and provides a complementary perspective to existing methods by leveraging geometric properties of the mixing function.

Abstract: Latent component identification from unknown nonlinear mixtures is a foundational challenge in machine learning, with applications in tasks such as disentangled representation learning and causal inference. Prior work in nonlinear independent component analysis (nICA) has shown that auxiliary signals – such as weak supervision – can support identifiability of conditionally independent latent components. More recent approaches explore structural assumptions, e.g., sparsity in the Jacobian of the mixing function, to relax such requirements. In this work, we introduce Diverse Influence Component Analysis (DICA), a framework that exploits the convex geometry of the mixing function’s Jacobian. We propose a Jacobian Volume Maximization (J-VolMax) criterion, which enables latent component identification by encouraging diversity in their influence on the observed variables. Under reasonable conditions, this approach achieves identifiability without relying on auxiliary information, latent component independence, or Jacobian sparsity assumptions. These results extend the scope of identifiability analysis and offer a complementary perspective to existing methods.

[751] Consistent Zero-Shot Imitation with Contrastive Goal Inference

Kathryn Wantlin, Chongyi Zheng, Benjamin Eysenbach

Main category: cs.LG

TL;DR: A self-supervised training method for interactive agents that enables instant imitation of human demonstrations by treating goals as atomic constructs and practicing goal-reaching during training.

DetailsMotivation: Current AI models lack explicit action training and fail to prepare for rapid task adaptation. Agents need interactive training like humans, but today's models assume humans spend most time in rewarding states, which is incorrect.

Method: Proposes goals as atomic constructs, automatically generates goals during training and practices reaching them using reinforcement learning exploration. During evaluation, solves inverse reinforcement learning to explain demonstrations as optimal goal-reaching behavior.

Result: Outperforms prior methods for zero-shot imitation on standard benchmarks not designed for goal-reaching.

Conclusion: Self-supervised interactive training with goal-reaching practice enables agents to instantly mimic human demonstrations, addressing the gap in action-based training for rapid adaptation.

Abstract: In the same way that generative models today conduct most of their training in a self-supervised fashion, how can agentic models conduct their training in a self-supervised fashion, interactively exploring, learning, and preparing to quickly adapt to new tasks? A prerequisite for embodied agents deployed in real world interactions ought to be training with interaction, yet today’s most successful AI models (e.g., VLMs, LLMs) are trained without an explicit notion of action. The problem of pure exploration (which assumes no data as input) is well studied in the reinforcement learning literature and provides agents with a wide array of experiences, yet it fails to prepare them for rapid adaptation to new tasks. Today’s language and vision models are trained on data provided by humans, which provides a strong inductive bias for the sorts of tasks that the model will have to solve (e.g., modeling chords in a song, phrases in a sonnet, sentences in a medical record). However, when they are prompted to solve a new task, there is a faulty tacit assumption that humans spend most of their time in the most rewarding states. The key contribution of our paper is a method for pre-training interactive agents in a self-supervised fashion, so that they can instantly mimic human demonstrations. Our method treats goals (i.e., observations) as the atomic construct. During training, our method automatically proposes goals and practices reaching them, building off prior work in reinforcement learning exploration. During evaluation, our method solves an (amortized) inverse reinforcement learning problem to explain demonstrations as optimal goal-reaching behavior. Experiments on standard benchmarks (not designed for goal-reaching) show that our approach outperforms prior methods for zero-shot imitation.

[752] Data Reliability Scoring

Yiling Chen, Shi Feng, Paul Kattuman, Fang-Yi Yu

Main category: cs.LG

TL;DR: The paper introduces a method to assess dataset reliability without ground truth by measuring how much reported data deviate from truth using the Gram determinant score, which is experiment-agnostic and preserves reliability orderings.

DetailsMotivation: To evaluate dataset reliability when ground truth is unavailable, especially for datasets from potentially strategic sources where true data are unobserved but outcomes of statistical experiments are visible.

Method: Proposes the Gram determinant score that measures the volume spanned by vectors describing the empirical distribution of observed data and experiment outcomes, preserving ground-truth-based reliability orderings.

Result: The Gram determinant score preserves several reliability orderings and provides the same reliability ranking regardless of the experiment (experiment agnosticism). Validated on synthetic noise models, CIFAR-10 embeddings, and real employment data.

Conclusion: The Gram determinant score effectively captures data quality across diverse observation processes without requiring access to ground truth.

Abstract: How can we assess the reliability of a dataset without access to ground truth? We introduce the problem of reliability scoring for datasets collected from potentially strategic sources. The true data are unobserved, but we see outcomes of an unknown statistical experiment that depends on them. To benchmark reliability, we define ground-truth-based orderings that capture how much reported data deviate from the truth. We then propose the Gram determinant score, which measures the volume spanned by vectors describing the empirical distribution of the observed data and experiment outcomes. We show that this score preserves several ground-truth based reliability orderings and, uniquely up to scaling, yields the same reliability ranking of datasets regardless of the experiment – a property we term experiment agnosticism. Experiments on synthetic noise models, CIFAR-10 embeddings, and real employment data demonstrate that the Gram determinant score effectively captures data quality across diverse observation processes.

[753] On the Universal Near Optimality of Hedge in Combinatorial Settings

Zhiyuan Fan, Arnab Maiti, Kevin Jamieson, Lillian J. Ratliff, Gabriele Farina

Main category: cs.LG

TL;DR: The paper analyzes the Hedge algorithm in combinatorial settings, showing it’s near-optimal (within √log d factor) for general combinatorial sets, but provably suboptimal for m-sets when log d ≤ m ≤ √d. Hedge is optimal for online multitask learning, and its near-optimality enables finding near-optimal regularizers for DAG shortest-path problems.

DetailsMotivation: To determine whether the classical Hedge algorithm is optimal across all combinatorial settings, as it achieves O(√(T log|X|)) regret but its optimality wasn't fully characterized for different combinatorial structures.

Method: Established a general lower bound of Ω(√(T log(|X|)/log d)) for any algorithm, compared it to Hedge’s upper bound, and analyzed specific combinatorial structures including m-sets, online multitask learning, and DAG shortest-path problems using Online Mirror Descent with dilated entropy regularizer.

Result: Hedge is near-optimal (within √log d factor) for general combinatorial sets, but provably suboptimal by exactly √log d for m-sets when log d ≤ m ≤ √d. Hedge is optimal for online multitask learning. The dilated entropy regularizer makes OMD iterate-equivalent to Hedge, inheriting its near-optimal guarantees for DAGs.

Conclusion: Hedge is near-optimal for general combinatorial settings but not universally optimal - its performance depends on the specific combinatorial structure. The analysis provides a comprehensive characterization of Hedge’s optimality across different combinatorial domains and enables finding near-optimal algorithms for DAG shortest-path problems.

Abstract: In this paper, we study the classical Hedge algorithm in combinatorial settings. In each round, the learner selects a vector $\boldsymbol{x}_t$ from a set $X \subseteq {0,1}^d$, observes a full loss vector $\boldsymbol{y}_t \in \mathbb{R}^d$, and incurs a loss $\langle \boldsymbol{x}_t, \boldsymbol{y}_t \rangle \in [-1,1]$. This setting captures several important problems, including extensive-form games, resource allocation, $m$-sets, online multitask learning, and shortest-path problems on directed acyclic graphs (DAGs). It is well known that Hedge achieves a regret of $O\big(\sqrt{T \log |X|}\big)$ after $T$ rounds of interaction. In this paper, we ask whether Hedge is optimal across all combinatorial settings. To that end, we show that for any $X \subseteq {0,1}^d$, Hedge is near-optimal–specifically, up to a $\sqrt{\log d}$ factor–by establishing a lower bound of $\Omega\big(\sqrt{T \log(|X|)/\log d}\big)$ that holds for any algorithm. We then identify a natural class of combinatorial sets–namely, $m$-sets with $\log d \leq m \leq \sqrt{d}$–for which this lower bound is tight, and for which Hedge is provably suboptimal by a factor of exactly $\sqrt{\log d}$. At the same time, we show that Hedge is optimal for online multitask learning, a generalization of the classical $K$-experts problem. Finally, we leverage the near-optimality of Hedge to establish the existence of a near-optimal regularizer for online shortest-path problems in DAGs–a setting that subsumes a broad range of combinatorial domains. Specifically, we show that the classical Online Mirror Descent (OMD) algorithm, when instantiated with the dilated entropy regularizer, is iterate-equivalent to Hedge, and therefore inherits its near-optimal regret guarantees for DAGs.

[754] D2C-HRHR: Discrete Actions with Double Distributional Critics for High-Risk-High-Return Tasks

Jundong Zhang, Yuhui Situ, Fanji Zhang, Rongji Deng, Tianqi Wei

Main category: cs.LG

TL;DR: Proposes a reinforcement learning framework for high-risk-high-return tasks using action discretization, entropy-regularized exploration, and dual-critic architecture to handle multimodal action distributions.

DetailsMotivation: Standard RL methods with unimodal Gaussian policies and scalar critics are ineffective for high-risk-high-return tasks that have multimodal action distributions and stochastic returns.

Method: Discretizes continuous action spaces to approximate multimodal distributions, uses entropy-regularized exploration for better coverage of risky actions, and implements dual-critic architecture for discrete value distribution estimation.

Result: Outperforms baseline methods on locomotion and manipulation benchmarks with high failure risks, demonstrating effectiveness in complex control domains.

Conclusion: Explicit modeling of multimodality and risk is crucial for RL in high-risk-high-return scenarios, and the proposed framework successfully addresses these challenges.

Abstract: Tasks involving high-risk-high-return (HRHR) actions, such as obstacle crossing, often exhibit multimodal action distributions and stochastic returns. Most reinforcement learning (RL) methods assume unimodal Gaussian policies and rely on scalar-valued critics, which limits their effectiveness in HRHR settings. We formally define HRHR tasks and theoretically show that Gaussian policies cannot guarantee convergence to the optimal solution. To address this, we propose a reinforcement learning framework that (i) discretizes continuous action spaces to approximate multimodal distributions, (ii) employs entropy-regularized exploration to improve coverage of risky but rewarding actions, and (iii) introduces a dual-critic architecture for more accurate discrete value distribution estimation. The framework scales to high-dimensional action spaces, supporting complex control domains. Experiments on locomotion and manipulation benchmarks with high risks of failure demonstrate that our method outperforms baselines, underscoring the importance of explicitly modeling multimodality and risk in RL.

[755] Adapting to Stochastic and Adversarial Losses in Episodic MDPs with Aggregate Bandit Feedback

Shinji Ito, Kevin Jamieson, Haipeng Luo, Arnab Maiti, Taira Tsuchiya

Main category: cs.LG

TL;DR: First BOBW algorithms for episodic MDPs with aggregate bandit feedback, achieving O(log T) regret in stochastic and O(√T) regret in adversarial settings with known transitions, plus extensions to unknown transitions.

DetailsMotivation: Prior work focused only on worst-case analysis, lacking algorithms that perform well in both stochastic and adversarial environments under aggregate bandit feedback.

Method: Combines FTRL over occupancy measures, self-bounding techniques, and new loss estimators inspired by online shortest path problems.

Result: Achieves O(log T) regret in stochastic settings and O(√T) regret in adversarial settings with known transitions, with matching lower bounds proving optimality.

Conclusion: Provides first BOBW algorithms for episodic MDPs with aggregate bandit feedback, establishing optimal regret bounds and extending to unknown transitions.

Abstract: We study online learning in finite-horizon episodic Markov decision processes (MDPs) under the challenging aggregate bandit feedback model, where the learner observes only the cumulative loss incurred in each episode, rather than individual losses at each state-action pair. While prior work in this setting has focused exclusively on worst-case analysis, we initiate the study of best-of-both-worlds (BOBW) algorithms that achieve low regret in both stochastic and adversarial environments. We propose the first BOBW algorithms for episodic tabular MDPs with aggregate bandit feedback. In the case of known transitions, our algorithms achieve $O(\log T)$ regret in stochastic settings and ${O}(\sqrt{T})$ regret in adversarial ones. Importantly, we also establish matching lower bounds, showing the optimality of our algorithms in this setting. We further extend our approach to unknown-transition settings by incorporating confidence-based techniques. Our results rely on a combination of FTRL over occupancy measures, self-bounding techniques, and new loss estimators inspired by recent advances in online shortest path problems. Along the way, we also provide the first individual-gap-dependent lower bounds and demonstrate near-optimal BOBW algorithms for shortest path problems with bandit feedback.

[756] Diagnosis of Fuel Cell Health Status with Deep Sparse Auto-Encoder Neural Network

Chenyan Fei, Dalin Zhang, Chen Melinda Dang

Main category: cs.LG

TL;DR: Deep sparse auto-encoding network predicts and classifies high-frequency impedance in fuel cells with over 92% accuracy, deployed on FPGA achieving nearly 90% hardware recognition rate.

DetailsMotivation: High-frequency impedance is crucial for fuel cell health assessment but online testing is complex and costly, requiring alternative prediction methods.

Method: Uses deep sparse auto-encoding network for prediction and classification of high-frequency impedance in fuel cells.

Result: Achieved accuracy rate above 92% for prediction/classification, and nearly 90% hardware recognition rate when deployed on FPGA.

Conclusion: The proposed method effectively addresses the complexity of online high-frequency impedance testing and enables practical deployment for fuel cell health monitoring.

Abstract: Effective and accurate diagnosis of fuel cell health status is crucial for ensuring the stable operation of fuel cell stacks. Among various parameters, high-frequency impedance serves as a critical indicator for assessing fuel cell state and health conditions. However, its online testing is prohibitively complex and costly. This paper employs a deep sparse auto-encoding network for the prediction and classification of high-frequency impedance in fuel cells, achieving metric of accuracy rate above 92%. The network is further deployed on an FPGA, attaining a hardware-based recognition rate almost 90%.

[757] Fighter: Unveiling the Graph Convolutional Nature of Transformers in Time Series Modeling

Chen Zhang, Weixin Bu, Wendong Xu, Runsheng Yu, Yik-Chung Wu, Ngai Wong

Main category: cs.LG

TL;DR: Transformers are fundamentally equivalent to Graph Convolutional Networks (GCNs), where attention matrices act as dynamic adjacency matrices and perform graph convolution operations.

DetailsMotivation: To demystify Transformer encoders in time series modeling by revealing their internal mechanisms and establishing their equivalence to GCNs.

Method: Proposed Fighter architecture that removes redundant linear projections and incorporates multi-hop graph aggregation based on the GCN-Transformer equivalence.

Result: Fighter achieves competitive performance on standard forecasting benchmarks while providing clearer mechanistic interpretability of predictions.

Conclusion: The Transformer-GCN equivalence provides an explicit and interpretable representation of temporal dependencies as graph edges, enabling more transparent time series modeling.

Abstract: Transformers have achieved remarkable success in time series modeling, yet their internal mechanisms remain opaque. This work demystifies the Transformer encoder by establishing its fundamental equivalence to a Graph Convolutional Network (GCN). We show that in the forward pass, the attention distribution matrix serves as a dynamic adjacency matrix, and its composition with subsequent transformations performs computations analogous to graph convolution. Moreover, we demonstrate that in the backward pass, the update dynamics of value and feed-forward projections mirror those of GCN parameters. Building on this unified theoretical reinterpretation, we propose \textbf{Fighter} (Flexible Graph Convolutional Transformer), a streamlined architecture that removes redundant linear projections and incorporates multi-hop graph aggregation. This perspective yields an explicit and interpretable representation of temporal dependencies across different scales, naturally expressed as graph edges. Experiments on standard forecasting benchmarks confirm that Fighter achieves competitive performance while providing clearer mechanistic interpretability of its predictions.

[758] Matricial Free Energy as a Gaussianizing Regularizer: Enhancing Autoencoders for Gaussian Code Generation

Rishi Sonthalia, Raj Rao Nadakuditi

Main category: cs.LG

TL;DR: A novel regularization method for autoencoders using matricial free energy that enforces Gaussian-like code distributions through singular value optimization.

DetailsMotivation: To develop a regularization scheme that ensures autoencoder codes have Gaussian-like distributions, improving generalization and enabling better performance in underdetermined inverse problems.

Method: Define a differentiable loss function based on matricial free energy that minimizes when the singular value distribution of the code matrix matches that of a sculpted random matrix with i.i.d. Gaussian entries. Use stochastic gradient descent to optimize this loss.

Result: Empirical simulations show the method produces Gaussian-like codes that generalize well across training and test sets. The approach also enables reliable Gaussian code generation for underdetermined inverse problems.

Conclusion: The matricial free energy regularization effectively enforces Gaussian code distributions in autoencoders, improving generalization and making the method suitable for applications in underdetermined inverse problems.

Abstract: We introduce a novel regularization scheme for autoencoders based on matricial free energy. Our approach defines a differentiable loss function in terms of the singular values of the code matrix (code dimension x batch size). From the standpoint of free probability an d random matrix theory, this loss achieves its minimum when the singular value distribution of the code matrix coincides with that of an appropriately sculpted random metric with i.i.d. Gaussian entries. Empirical simulations demonstrate that minimizing the negative matricial free energy through standard stochastic gradient-based training yields Gaussian-like codes that generalize across training and test sets. Building on this foundation, we propose a matricidal free energy maximizing autoencoder that reliably produces Gaussian codes and show its application to underdetermined inverse problems.

[759] Continuous Q-Score Matching: Diffusion Guided Reinforcement Learning for Continuous-Time Control

Chengxiu Hua, Jiawen Gu, Yushun Tang

Main category: cs.LG

TL;DR: A novel continuous-time reinforcement learning method using martingale conditions and score matching for policy improvement, preserving Q-function action-evaluation without time discretization.

DetailsMotivation: Most existing RL methods are formulated in discrete time, but many real-world problems require continuous-time control governed by stochastic differential equations.

Method: Characterizes continuous-time Q-functions via martingale conditions, links diffusion policy scores to action gradients of learned Q-functions, and proposes Continuous Q-Score Matching (CQSM) algorithm.

Result: Provides theoretical closed-form solutions for linear-quadratic control problems and demonstrates effectiveness in simulated environments compared to popular baselines.

Conclusion: The method successfully addresses the challenge of preserving Q-function action-evaluation capability in continuous-time RL without relying on time discretization.

Abstract: Reinforcement learning (RL) has achieved significant success across a wide range of domains, however, most existing methods are formulated in discrete time. In this work, we introduce a novel RL method for continuous-time control, where stochastic differential equations govern state-action dynamics. Departing from traditional value function-based approaches, our key contribution is the characterization of continuous-time Q-functions via a martingale condition and the linking of diffusion policy scores to the action gradient of a learned continuous Q-function by the dynamic programming principle. This insight motivates Continuous Q-Score Matching (CQSM), a score-based policy improvement algorithm. Notably, our method addresses a long-standing challenge in continuous-time RL: preserving the action-evaluation capability of Q-functions without relying on time discretization. We further provide theoretical closed-form solutions for linear-quadratic (LQ) control problems within our framework. Numerical results in simulated environments demonstrate the effectiveness of our proposed method and compare it to popular baselines.

[760] MemoryBench: A Benchmark for Memory and Continual Learning in LLM Systems

Qingyao Ai, Yichen Tang, Changyue Wang, Jianming Long, Weihang Su, Yiqun Liu

Main category: cs.LG

TL;DR: The paper proposes a new benchmark for evaluating LLM memory and continual learning abilities using simulated user feedback across multiple domains, languages, and task types, as existing benchmarks focus too much on homogeneous reading comprehension.

DetailsMotivation: Current LLM scaling methods are reaching limits due to data depletion and diminishing returns. Inspired by human learning abilities, there's a need to develop memory and continual learning frameworks for LLMs, but existing benchmarks don't properly test learning from accumulated user feedback.

Method: The authors developed a user feedback simulation framework and comprehensive benchmark covering multiple domains, languages, and task types to evaluate LLM continual learning abilities.

Result: Experiments showed that state-of-the-art baselines perform poorly in effectiveness and efficiency on the proposed benchmark, indicating current methods are inadequate for real-world continual learning scenarios.

Conclusion: The benchmark provides a foundation for future research on LLM memory and optimization algorithms, as current approaches fall short in learning from accumulated user feedback during service time.

Abstract: Scaling up data, parameters, and test-time computation has been the mainstream methods to improve LLM systems (LLMsys), but their upper bounds are almost reached due to the gradual depletion of high-quality data and marginal gains obtained from larger computational resource consumption. Inspired by the abilities of human and traditional AI systems in learning from practice, constructing memory and continual learning frameworks for LLMsys has become an important and popular research direction in recent literature. Yet, existing benchmarks for LLM memory often focus on evaluating the system on homogeneous reading comprehension tasks with long-form inputs rather than testing their abilities to learn from accumulated user feedback in service time. Therefore, we propose a user feedback simulation framework and a comprehensive benchmark covering multiple domains, languages, and types of tasks to evaluate the continual learning abilities of LLMsys. Experiments show that the effectiveness and efficiency of state-of-the-art baselines are far from satisfying, and we hope this benchmark could pave the way for future studies on LLM memory and optimization algorithms.

[761] In-situ Autoguidance: Eliciting Self-Correction in Diffusion Models

Enhao Gu, Haolin Hou

Main category: cs.LG

TL;DR: In-situ Autoguidance enables diffusion models to self-correct during inference without auxiliary models, achieving quality and alignment improvements similar to classifier-free guidance but without reducing diversity.

DetailsMotivation: Classifier-free guidance improves image quality and prompt alignment but reduces diversity, and existing disentanglement methods require separate auxiliary models, creating overhead.

Method: Dynamic generation of inferior predictions using stochastic forward passes during inference, treating guidance as self-correction without external components.

Result: Zero-cost approach achieves quality and alignment improvements comparable to classifier-free guidance while maintaining diversity, establishing new baseline for efficient guidance.

Conclusion: Self-guidance benefits can be achieved without external models through inference-time self-correction, making guidance more accessible and cost-effective.

Abstract: The generation of high-quality, diverse, and prompt-aligned images is a central goal in image-generating diffusion models. The popular classifier-free guidance (CFG) approach improves quality and alignment at the cost of reduced variation, creating an inherent entanglement of these effects. Recent work has successfully disentangled these properties by guiding a model with a separately trained, inferior counterpart; however, this solution introduces the considerable overhead of requiring an auxiliary model. We challenge this prerequisite by introducing In-situ Autoguidance, a method that elicits guidance from the model itself without any auxiliary components. Our approach dynamically generates an inferior prediction on the fly using a stochastic forward pass, reframing guidance as a form of inference-time self-correction. We demonstrate that this zero-cost approach is not only viable but also establishes a powerful new baseline for cost-efficient guidance, proving that the benefits of self-guidance can be achieved without external models.

[762] Learning After Model Deployment

Derda Kaymak, Gyuhak Kim, Tomoya Kaichi, Tatsuya Konishi, Bing Liu

Main category: cs.LG

TL;DR: The paper introduces ALMD (Autonomous Learning after Model Deployment), a paradigm where models continuously detect novel samples from unseen classes and learn them incrementally during deployment, without human engineers.

DetailsMotivation: Classic supervised learning with fixed deployed models is inadequate for dynamic environments where unexpected samples from unseen classes appear. Models need to autonomously detect and learn new classes during application.

Method: Proposes PLDA method that performs dynamic OOD detection and incremental learning of new classes on the fly, addressing challenges of expanding ID classes, incremental learning without retraining from scratch, and data scarcity.

Result: Empirical evaluations demonstrate the effectiveness of the proposed PLDA method.

Conclusion: ALMD enables continuous learning in dynamic environments where models can autonomously detect novel samples and incrementally learn new classes during deployment, overcoming limitations of traditional fixed models.

Abstract: In classic supervised learning, once a model is deployed in an application, it is fixed. No updates will be made to it during the application. This is inappropriate for many dynamic and open environments, where unexpected samples from unseen classes may appear. In such an environment, the model should be able to detect these novel samples from unseen classes and learn them after they are labeled. We call this paradigm Autonomous Learning after Model Deployment (ALMD). The learning here is continuous and involves no human engineers. Labeling in this scenario is performed by human co-workers or other knowledgeable agents, which is similar to what humans do when they encounter an unfamiliar object and ask another person for its name. In ALMD, the detection of novel samples is dynamic and differs from traditional out-of-distribution (OOD) detection in that the set of in-distribution (ID) classes expands as new classes are learned during application, whereas ID classes is fixed in traditional OOD detection. Learning is also different from classic supervised learning because in ALMD, we learn the encountered new classes immediately and incrementally. It is difficult to retrain the model from scratch using all the past data from the ID classes and the novel samples from newly discovered classes, as this would be resource- and time-consuming. Apart from these two challenges, ALMD faces the data scarcity issue because instances of new classes often appear sporadically in real-life applications. To address these issues, we propose a novel method, PLDA, which performs dynamic OOD detection and incremental learning of new classes on the fly. Empirical evaluations will demonstrate the effectiveness of PLDA.

[763] Auto-Rubric: Learning to Extract Generalizable Criteria for Reward Modeling

Lipeng Xie, Sen Huang, Zhuo Zhang, Anni Zou, Yunpeng Zhai, Dingchao Ren, Kezun Zhang, Haoyuan Hu, Boyin Liu, Haoran Chen, Zhaoyang Liu, Bolin Ding

Main category: cs.LG

TL;DR: A training-free framework for reward modeling that infers query-specific rubrics and generalizes them into a compact core set, achieving exceptional data efficiency with just 70 preference pairs.

DetailsMotivation: Address limitations in current reward models: costly preference datasets, poor interpretability, and trade-offs between scalability and reliability in rubric-based approaches.

Method: Two-stage approach: 1) Propose-Evaluate-Revise pipeline to infer query-specific rubrics, 2) Information-theoretic coding rate maximization to generalize rubrics into compact “Theme-Tips” hierarchical set.

Result: Achieves remarkable data efficiency using only 70 preference pairs (1.5% of source data), enables smaller models like Qwen3-8B to outperform specialized fully-trained counterparts.

Conclusion: Pioneers a scalable, interpretable, and data-efficient path for reward modeling by leveraging rubric generalization ability across diverse queries.

Abstract: Reward models are essential for aligning Large Language Models (LLMs) with human values, yet their development is hampered by costly preference datasets and poor interpretability. While recent rubric-based approaches offer transparency, they often lack systematic quality control and optimization, creating a trade-off between scalability and reliability. We address these limitations with a novel, training-free framework built on a key assumption: \textit{evaluation rubrics underlying human preferences exhibit significant generalization ability across diverse queries}, a property that enables remarkable data efficiency. Our two-stage approach first infers high-quality, query-specific rubrics using a validation-guided \textbf{Propose-Evaluate-Revise} pipeline. Second, it generalizes these granular rubrics into a compact, non-redundant core set by maximizing an \textbf{information-theoretic coding rate}. The final output is an interpretable, hierarchical “Theme-Tips” rubric set. Extensive experiments demonstrate the framework’s exceptional data efficiency and performance. Critically, using just 70 preference pairs (1.5% of the source data), our method also empowers smaller models like Qwen3-8B to outperform specialized, fully-trained counterparts. This work pioneers a scalable, interpretable, and data-efficient path for reward modeling.

[764] ALPINE: A Lightweight and Adaptive Privacy-Decision Agent Framework for Dynamic Edge Crowdsensing

Guanjie Cheng, Siyang Liu, Junqin Huang, Xinkui Zhao, Yin Wang, Mengying Zhu, Linghe Kong, Shuiguang Deng

Main category: cs.LG

TL;DR: ALPINE is an adaptive differential privacy framework for mobile edge crowdsensing that dynamically adjusts privacy levels in real-time using a TD3-based control system to balance privacy, utility, and energy costs.

DetailsMotivation: Static differential privacy mechanisms fail to adapt to evolving risks in dynamic edge environments, leading to either excessive noise or inadequate protection against privacy threats.

Method: ALPINE uses a closed-loop control system with four modules: dynamic risk perception, privacy decision via TD3 reinforcement learning, local privacy execution, and performance verification from edge nodes.

Result: Extensive analysis shows ALPINE effectively mitigates inference attacks while preserving data utility and managing energy costs, making it practical for large-scale edge applications.

Conclusion: The proposed adaptive framework successfully addresses the limitations of static DP mechanisms in dynamic edge environments by achieving dynamic equilibrium among privacy, utility, and cost.

Abstract: Mobile edge crowdsensing (MECS) systems continuously generate and transmit user data in dynamic, resource-constrained environments, exposing users to significant privacy threats. In practice, many privacy-preserving mechanisms build on differential privacy (DP). However, static DP mechanisms often fail to adapt to evolving risks, for example, shifts in adversarial capabilities, resource constraints and task requirements, resulting in either excessive noise or inadequate protection. To address this challenge, we propose ALPINE, a lightweight, adaptive framework that empowers terminal devices to autonomously adjust differential privacy levels in real time. ALPINE operates as a closed-loop control system consisting of four modules: dynamic risk perception, privacy decision via twin delayed deep deterministic policy gradient (TD3), local privacy execution and performance verification from edge nodes. Based on environmental risk assessments, we design a reward function that balances privacy gains, data utility and energy cost, guiding the TD3 agent to adaptively tune noise magnitude across diverse risk scenarios and achieve a dynamic equilibrium among privacy, utility and cost. Both the collaborative risk model and pretrained TD3-based agent are designed for low-overhead deployment. Extensive theoretical analysis and real-world simulations demonstrate that ALPINE effectively mitigates inference attacks while preserving utility and cost, making it practical for large-scale edge applications.

[765] Localist LLMs with Recruitment Learning

Joachim Diederich

Main category: cs.LG

TL;DR: A framework for training LLMs with adjustable internal representations from interpretable (localist) to efficient (distributed) encodings, featuring a tunable locality dial, adaptive semantic block allocation, and hierarchical recruitment of specialized LLMs.

DetailsMotivation: To enable practitioners to balance interpretability and performance in LLMs, particularly for regulated domains requiring both transparency and high capability, by providing continuous control over representation localization.

Method: Uses group sparsity penalties on attention, information-theoretic anchor design, dynamic rule injection, and recruitment criteria based on penalized likelihood. Features a locality dial parameter, adaptive semantic block allocation, and hierarchical recruitment framework.

Result: Established mathematical guarantees for attention concentration on semantically relevant blocks with exact bounds on attention entropy and pointer fidelity. Provides convergence guarantees at both block and LLM levels for discovering optimal semantic partitions.

Conclusion: The framework successfully enables continuous interpolation between interpretable and high-performance modes while adapting architectural capacity at multiple granularities, supporting applications that require both transparency and capability.

Abstract: We present a novel framework for training large language models with continuously adjustable internal representations that span the full spectrum from localist (interpretable, rule-based) to distributed (generalizable, efficient) encodings. The key innovations are (1) a locality dial, a tunable parameter that dynamically controls the degree of localization during both training and inference without requiring model retraining, (2) an information-theoretic recruitment mechanism that adaptively allocates semantic blocks as needed, eliminating the requirement for complete domain knowledge at initialization, and (3) a hierarchical recruitment framework that extends capacity allocation to entire specialized LLMs, enabling multi-granularity architectural adaptation. This is achieved through group sparsity penalties on attention mechanisms, information-theoretic anchor design, dynamic rule injection, and principled recruitment criteria based on penalized likelihood with explicit units. We provide rigorous mathematical results establishing explicit threshold conditions under which attention provably concentrates on semantically relevant blocks at stationary points, with exact bounds on attention entropy and pointer fidelity. The hierarchical recruitment mechanism provides convergence guarantees at both the block level (fine-grained, within-LLM) and the LLM level (coarse-grained, cross-domain), ensuring the system discovers semantic partitions that balance model complexity against data encoding efficiency. This framework enables practitioners to continuously interpolate between interpretable and high-performance modes while adapting architectural capacity at multiple granularities, supporting applications in regulated domains requiring both transparency and capability.

[766] Robustness in Text-Attributed Graph Learning: Insights, Trade-offs, and New Defenses

Runlin Lei, Lu Yi, Mingguo He, Pengyu Qiu, Zhewei Wei, Yongchao Liu, Chuntao Hong

Main category: cs.LG

TL;DR: This paper introduces a unified framework to evaluate robustness of GNNs and LLMs on Text-Attributed Graphs, revealing inherent robustness trade-offs between text and structure, and proposes SFT-auto for balanced robustness.

DetailsMotivation: Current evaluations of GNNs and LLMs on Text-Attributed Graphs are fragmented and fail to systematically investigate robustness against textual and structural perturbations across different models and attack scenarios.

Method: The authors introduce a unified framework to evaluate classical GNNs, robust GNNs, and GraphLLMs across 10 datasets from 4 domains under text-based, structure-based, and hybrid perturbations in poisoning and evasion scenarios.

Result: Key findings include: 1) models have inherent robustness trade-offs between text and structure, 2) GNN/RGNN performance depends heavily on text encoder and attack type, 3) GraphLLMs are particularly vulnerable to training data corruption.

Conclusion: The work establishes foundations for TAG security research and introduces SFT-auto, a novel framework that provides superior balanced robustness against both textual and structural attacks in a single model.

Abstract: While Graph Neural Networks (GNNs) and Large Language Models (LLMs) are powerful approaches for learning on Text-Attributed Graphs (TAGs), a comprehensive understanding of their robustness remains elusive. Current evaluations are fragmented, failing to systematically investigate the distinct effects of textual and structural perturbations across diverse models and attack scenarios. To address these limitations, we introduce a unified and comprehensive framework to evaluate robustness in TAG learning. Our framework evaluates classical GNNs, robust GNNs (RGNNs), and GraphLLMs across ten datasets from four domains, under diverse text-based, structure-based, and hybrid perturbations in both poisoning and evasion scenarios. Our extensive analysis reveals multiple findings, among which three are particularly noteworthy: 1) models have inherent robustness trade-offs between text and structure, 2) the performance of GNNs and RGNNs depends heavily on the text encoder and attack type, and 3) GraphLLMs are particularly vulnerable to training data corruption. To overcome the identified trade-offs, we introduce SFT-auto, a novel framework that delivers superior and balanced robustness against both textual and structural attacks within a single model. Our work establishes a foundation for future research on TAG security and offers practical solutions for robust TAG learning in adversarial environments. Our code is available at: https://github.com/Leirunlin/TGRB.

[767] A Standardized Benchmark for Machine-Learned Molecular Dynamics using Weighted Ensemble Sampling

Alexander Aghili, Andy Bruce, Daniel Sabo, Sanya Murdeshwar, Kevin Bachelor, Ionut Mistreanu, Ashwin Lokapally, Razvan Marinescu

Main category: cs.LG

TL;DR: A modular benchmarking framework for protein molecular dynamics methods that enables standardized validation using enhanced sampling analysis and comprehensive evaluation metrics.

DetailsMotivation: Address the lack of standardized tools for objective comparison between molecular dynamics simulation approaches, which is hindered by inconsistent evaluation metrics and insufficient sampling of rare conformational states.

Method: Uses weighted ensemble sampling via WESTPA with TICA-derived progress coordinates, includes a flexible propagator interface supporting arbitrary simulation engines (both classical force fields and machine learning models), and provides comprehensive evaluation with 19+ metrics and visualizations.

Result: Developed a framework with extensive simulations of nine diverse proteins (10-224 residues) and demonstrated utility through validation tests comparing classic MD with implicit solvent and CGSchNet models (fully trained vs under-trained).

Conclusion: The open-source platform standardizes evaluation protocols and enables reproducible comparisons across MD approaches, laying groundwork for consistent, rigorous benchmarking in the molecular simulation community.

Abstract: The rapid evolution of molecular dynamics (MD) methods, including machine-learned dynamics, has outpaced the development of standardized tools for method validation. Objective comparison between simulation approaches is often hindered by inconsistent evaluation metrics, insufficient sampling of rare conformational states, and the absence of reproducible benchmarks. To address these challenges, we introduce a modular benchmarking framework that systematically evaluates protein MD methods using enhanced sampling analysis. Our approach uses weighted ensemble (WE) sampling via The Weighted Ensemble Simulation Toolkit with Parallelization and Analysis (WESTPA), based on progress coordinates derived from Time-lagged Independent Component Analysis (TICA), enabling fast and efficient exploration of protein conformational space. The framework includes a flexible, lightweight propagator interface that supports arbitrary simulation engines, allowing both classical force fields and machine learning-based models. Additionally, the framework offers a comprehensive evaluation suite capable of computing more than 19 different metrics and visualizations across a variety of domains. We further contribute a dataset of nine diverse proteins, ranging from 10 to 224 residues, that span a variety of folding complexities and topologies. Each protein has been extensively simulated at 300K for one million MD steps per starting point (4 ns). To demonstrate the utility of our framework, we perform validation tests using classic MD simulations with implicit solvent and compare protein conformational sampling using a fully trained versus under-trained CGSchNet model. By standardizing evaluation protocols and enabling direct, reproducible comparisons across MD approaches, our open-source platform lays the groundwork for consistent, rigorous benchmarking across the molecular simulation community.

[768] Optimizing Energy Management of Smart Grid using Reinforcement Learning aided by Surrogate models built using Physics-informed Neural Networks

Julen Cestero, Carmine Delle Femine, Kenji S. Muro, Marco Quartulli, Marcello Restelli

Main category: cs.LG

TL;DR: Using Physics-informed Neural Networks (PINNs) as surrogate models to replace costly smart grid simulators for more efficient Reinforcement Learning policy training in Optimal Power Flow problems.

DetailsMotivation: Reinforcement Learning needs extensive iterations in costly simulators for smart grid optimization, leading to sample efficiency problems.

Method: Substitute expensive smart grid simulators with surrogate models built using Physics-informed Neural Networks (PINNs) to optimize RL policy training.

Result: Achieved convergent results in a fraction of the time compared to using the original environment.

Conclusion: PINNs-based surrogate models effectively address the sample efficiency problem in RL for smart grid energy management optimization.

Abstract: Optimizing the energy management within a smart grids scenario presents significant challenges, primarily due to the complexity of real-world systems and the intricate interactions among various components. Reinforcement Learning (RL) is gaining prominence as a solution for addressing the challenges of Optimal Power Flow in smart grids. However, RL needs to iterate compulsively throughout a given environment to obtain the optimal policy. This means obtaining samples from a, most likely, costly simulator, which can lead to a sample efficiency problem. In this work, we address this problem by substituting costly smart grid simulators with surrogate models built using Phisics-informed Neural Networks (PINNs), optimizing the RL policy training process by arriving to convergent results in a fraction of the time employed by the original environment.

[769] SOLE: Hardware-Software Co-design of Softmax and LayerNorm for Efficient Transformer Inference

Wenxun Wang, Shuchang Zhou, Wenyu Sun, Peiqin Sun, Yongpan Liu

Main category: cs.LG

TL;DR: SOLE is a hardware-software co-design that optimizes Softmax and LayerNorm operations in Transformers using E2Softmax (log2 quantization) and AILayerNorm (low-precision statistics), achieving significant speedup and energy savings without retraining.

DetailsMotivation: Transformers have performance limitations due to inefficient Softmax and LayerNorm operations. Existing approximation methods suffer from memory overhead issues and require costly retraining to compensate for errors.

Method: SOLE combines E2Softmax (using log2 quantization of exponent function and log-based division) and AILayerNorm (using low-precision statistic calculation) to approximate Softmax and LayerNorm operations efficiently.

Result: SOLE maintains inference accuracy without retraining while achieving 3.04x and 3.86x energy-efficiency improvements, and 2.82x and 3.32x area-efficiency improvements over prior state-of-the-art custom hardware for Softmax and LayerNorm respectively.

Conclusion: SOLE provides an effective hardware-software co-design solution that significantly improves the efficiency of Softmax and LayerNorm operations in Transformers without sacrificing accuracy or requiring retraining.

Abstract: Transformers have shown remarkable performance in both natural language processing (NLP) and computer vision (CV) tasks. However, their real-time inference speed and efficiency are limited due to the inefficiency in Softmax and Layer Normalization (LayerNorm). Previous works based on function approximation suffer from inefficient implementation as they place emphasis on computation while disregarding memory overhead concerns. Moreover, such methods rely on retraining to compensate for approximation error which can be costly and inconvenient. In this paper, we present SOLE, a hardware-software co-design for Softmax and LayerNorm which is composed of E2Softmax and AILayerNorm. E2Softmax utilizes log2 quantization of exponent function and log-based division to approximate Softmax while AILayerNorm adopts low-precision statistic calculation. Compared with state-of-the-art designs, we achieve both low-precision calculation and low bit-width storage on Softmax and LayerNorm. Experiments show that SOLE maintains inference accuracy without retraining while offering orders of magnitude speedup and energy savings over GPU, achieving 3.04x, 3.86x energy-efficiency improvements and 2.82x, 3.32x area-efficiency improvements over prior state-of-the-art custom hardware for Softmax and LayerNorm, respectively.

[770] TabR1: Taming GRPO for tabular reasoning LLMs

Pengxiang Cai, Zihao Gao, Jintai Chen

Main category: cs.LG

TL;DR: TabR1 is the first reasoning LLM for tabular prediction that uses Permutation Relative Policy Optimization (PRPO) to enable multi-step reasoning, achieving strong performance with limited supervision while maintaining interpretability.

DetailsMotivation: Traditional tabular prediction methods (gradient-boosted trees, specialized deep learning) have limited interpretability and weak cross-table transfer. Reasoning LLMs promise adaptability and transparent reasoning but haven't been fully realized for tabular data.

Method: Uses Permutation Relative Policy Optimization (PRPO) - a reinforcement learning method that encodes column-permutation invariance as structural prior. Creates multiple label-preserving permutations per sample and estimates advantages within/across permutations to transform sparse rewards into dense learning signals.

Result: TabR1 achieves performance comparable to strong baselines under full-supervision fine-tuning. In zero-shot setting, approaches performance of strong baselines under 32-shot setting. TabR1 (8B) substantially outperforms much larger LLMs, achieving up to 53.17% improvement over DeepSeek-R1 (685B).

Conclusion: PRPO effectively activates reasoning ability of LLMs for tabular prediction, enhancing few-shot/zero-shot performance and interpretability with limited supervision.

Abstract: Tabular prediction has traditionally relied on gradient-boosted decision trees and specialized deep learning models, which excel within tasks but provide limited interpretability and weak transfer across tables. Reasoning large language models (LLMs) promise cross-task adaptability with trans- parent reasoning traces, yet their potential has not been fully realized for tabular data. This paper presents TabR1, the first reasoning LLM for tabular prediction with multi-step reasoning. At its core is Permutation Relative Policy Optimization (PRPO), a simple yet efficient reinforcement learning method that encodes column-permutation invariance as a structural prior. By construct- ing multiple label-preserving permutations per sample and estimating advantages both within and across permutations, PRPO transforms sparse rewards into dense learning signals and improves generalization. With limited supervision, PRPO activates the reasoning ability of LLMs for tabular prediction, enhancing few-shot and zero-shot performance as well as interpretability. Comprehensive experiments demonstrate that TabR1 achieves performance comparable to strong baselines under full-supervision fine-tuning. In the zero-shot setting, TabR1 approaches the performance of strong baselines under the 32-shot setting. Moreover, TabR1 (8B) substantially outperforms much larger LLMs across various tasks, achieving up to 53.17% improvement over DeepSeek-R1 (685B).

[771] A Prototypical Network with an Attention-based Encoder for Drivers Identification Application

Wei-Hsun Lee, Che-Yu Chang, Kuang-Yu Li

Main category: cs.LG

TL;DR: This paper proposes two deep learning models for driver identification: AttEnc (attention-based encoder) for standard identification with high accuracy and parameter efficiency, and P-AttEnc (prototypical network combined with AttEnc) for few-shot learning to handle data shortages and unknown drivers.

DetailsMotivation: Driver identification is important for data-driven applications, but biometric technologies raise privacy concerns. Existing methods don't address data shortages and are inflexible with unknown drivers.

Method: Proposed AttEnc uses attention mechanism for efficient driver identification. P-AttEnc combines prototypical network with AttEnc for few-shot learning to handle data shortages and classify unknown drivers.

Result: AttEnc achieved 99.3%, 99.0%, 99.9% accuracy on three datasets with 44-79% faster prediction time and 87.6% parameter reduction. P-AttEnc achieved 69.8% accuracy in one-shot scenario and 65.7% accuracy for unknown driver classification.

Conclusion: The proposed attention-based models provide efficient driver identification while addressing data shortage issues and enabling classification of unknown drivers through few-shot learning.

Abstract: Driver identification has become an area of increasing interest in recent years, especially for data- driven applications, because biometric-based technologies may incur privacy issues. This study proposes a deep learning neural network architecture, an attention-based encoder (AttEnc), which uses an attention mechanism for driver identification and uses fewer model parameters than current methods. Most studies do not address the issue of data shortages for driver identification, and most of them are inflexible when encountering unknown drivers. In this study, an architecture that combines a prototypical network and an attention-based encoder (P-AttEnc) is proposed. It applies few-shot learning to overcome the data shortage issues and to enhance model generalizations. The experiments showed that the attention-based encoder can identify drivers with accuracies of 99.3%, 99.0% and 99.9% in three different datasets and has a prediction time that is 44% to 79% faster because it significantly reduces, on average, 87.6% of the model parameters. P-AttEnc identifies drivers based on few shot data, extracts driver fingerprints to address the issue of data shortages, and is able to classify unknown drivers. The first experiment showed that P-AttEnc can identify drivers with an accuracy of 69.8% in the one-shot scenario. The second experiment showed that P-AttEnc, in the 1-shot scenario, can classify unknown drivers with an average accuracy of 65.7%.

[772] Adaptive Discretization for Consistency Models

Jiayu Bai, Zhanbo Feng, Zhijie Deng, Tianqi Hou, Robert C. Qiu, Zenan Ling

Main category: cs.LG

TL;DR: Proposes ADCMs - a unified framework for automatic and adaptive discretization of Consistency Models that improves training efficiency and performance without manual tuning.

DetailsMotivation: Existing Consistency Models rely on manually designed discretization schemes that require repeated adjustments for different noise schedules and datasets, which is inefficient.

Method: Formulates discretization as an optimization problem using local consistency as objective and global consistency as constraint with Lagrange multiplier, implemented via Gauss-Newton method.

Result: ADCMs significantly improve training efficiency and achieve superior generative performance on CIFAR-10 and ImageNet with minimal overhead, while adapting well to advanced DM variants.

Conclusion: The proposed adaptive discretization framework provides an efficient and automated approach for training Consistency Models without manual parameter tuning.

Abstract: Consistency Models (CMs) have shown promise for efficient one-step generation. However, most existing CMs rely on manually designed discretization schemes, which can cause repeated adjustments for different noise schedules and datasets. To address this, we propose a unified framework for the automatic and adaptive discretization of CMs, formulating it as an optimization problem with respect to the discretization step. Concretely, during the consistency training process, we propose using local consistency as the optimization objective to ensure trainability by avoiding excessive discretization, and taking global consistency as a constraint to ensure stability by controlling the denoising error in the training target. We establish the trade-off between local and global consistency with a Lagrange multiplier. Building on this framework, we achieve adaptive discretization for CMs using the Gauss-Newton method. We refer to our approach as ADCMs. Experiments demonstrate that ADCMs significantly improve the training efficiency of CMs, achieving superior generative performance with minimal training overhead on both CIFAR-10 and ImageNet. Moreover, ADCMs exhibit strong adaptability to more advanced DM variants. Code is available at https://github.com/rainstonee/ADCM.

[773] Uncertainty-aware data assimilation through variational inference

Anthony Frion, David S Greenberg

Main category: cs.LG

TL;DR: The paper proposes a variational inference-based extension to deterministic machine learning for data assimilation, modeling predicted states as multivariate Gaussian distributions to handle uncertainty in chaotic systems.

DetailsMotivation: Data assimilation involves uncertainty in most settings when combining dynamical models with noisy, incomplete observations. Existing deterministic approaches need extension to better handle this uncertainty.

Method: Extends deterministic machine learning approach using variational inference, modeling predicted states as multivariate Gaussian distributions. Tested on chaotic Lorenz-96 dynamics.

Result: The model achieves nearly perfectly calibrated predictions and can be integrated into wider variational data assimilation pipelines, enabling greater benefit from longer data assimilation windows.

Conclusion: The proposed stochastic variational inference approach successfully handles uncertainty in data assimilation, providing well-calibrated predictions and improved performance with longer observation windows.

Abstract: Data assimilation, consisting in the combination of a dynamical model with a set of noisy and incomplete observations in order to infer the state of a system over time, involves uncertainty in most settings. Building upon an existing deterministic machine learning approach, we propose a variational inference-based extension in which the predicted state follows a multivariate Gaussian distribution. Using the chaotic Lorenz-96 dynamics as a testing ground, we show that our new model enables to obtain nearly perfectly calibrated predictions, and can be integrated in a wider variational data assimilation pipeline in order to achieve greater benefit from increasing lengths of data assimilation windows. Our code is available at https://github.com/anthony-frion/Stochastic_CODA.

[774] Breaking and Fixing Defenses Against Control-Flow Hijacking in Multi-Agent Systems

Rishi Jha, Harold Triedman, Justin Wagle, Vitaly Shmatikov

Main category: cs.LG

TL;DR: ControlValve is a new defense against control-flow hijacking attacks in multi-agent systems that generates permitted control-flow graphs and enforces execution compliance with contextual rules.

DetailsMotivation: Existing defenses like LlamaFirewall using alignment checks are insufficient against control-flow hijacking attacks due to conflicting safety/functionality objectives and brittle alignment definitions.

Method: ControlValve generates permitted control-flow graphs for multi-agent systems and enforces execution compliance with these graphs plus zero-shot generated contextual rules for each agent invocation.

Result: The paper demonstrates that current alignment-based defenses can be evaded and proposes ControlValve as a more robust alternative.

Conclusion: ControlValve provides better protection against control-flow hijacking by applying control-flow integrity and least privilege principles to multi-agent systems.

Abstract: Control-flow hijacking attacks manipulate orchestration mechanisms in multi-agent systems into performing unsafe actions that compromise the system and exfiltrate sensitive information. Recently proposed defenses, such as LlamaFirewall, rely on alignment checks of inter-agent communications to ensure that all agent invocations are “related to” and “likely to further” the original objective. We start by demonstrating control-flow hijacking attacks that evade these defenses even if alignment checks are performed by advanced LLMs. We argue that the safety and functionality objectives of multi-agent systems fundamentally conflict with each other. This conflict is exacerbated by the brittle definitions of “alignment” and the checkers’ incomplete visibility into the execution context. We then propose, implement, and evaluate ControlValve, a new defense inspired by the principles of control-flow integrity and least privilege. ControlValve (1) generates permitted control-flow graphs for multi-agent systems, and (2) enforces that all executions comply with these graphs, along with contextual rules (generated in a zero-shot manner) for each agent invocation.

[775] Layer Specialization Underlying Compositional Reasoning in Transformers

Jing Liu

Main category: cs.LG

TL;DR: Transformers develop modular, hierarchical representations that enable compositional reasoning on unseen sequences, with performance scaling systematically with task complexity and in-context examples.

DetailsMotivation: To understand how transformers achieve compositional reasoning on sequences not seen during training, particularly through in-context learning and skill composition mechanisms.

Method: Used Random Hierarchy Model (RHM) - a probabilistic context-free grammar generating sequences through recursive rules. Trained models on sequence subsets and evaluated across four generalization conditions: memorization, in-distribution, out-of-distribution with same rules, and cross-layer transfer.

Result: Performance improved systematically with task complexity and number of in-context examples. Out-of-distribution tasks required significantly more examples than in-distribution scenarios. Progressive layer specialization emerged during training, correlating with generalization. Transformers developed structured, hierarchical representations in specialized layers.

Conclusion: Transformers develop modular, interpretable mechanisms supporting compositional reasoning, with internal algorithmic structure directly linked to observed behavioral capabilities.

Abstract: Transformers exhibit compositional reasoning on sequences not observed during training, a capability often attributed to in-context learning (ICL) and skill composition. We investigate this phenomenon using the Random Hierarchy Model (RHM), a probabilistic context-free grammar that generates sequences through recursive rule application. Models are trained on subsets of sequences and evaluated across four generalization conditions: memorization, in-distribution generalization, out-of-distribution generalization with the same rules, and cross-layer transfer. Behaviorally, performance improves systematically with task complexity and the number of in-context examples, with out-of-distribution tasks requiring substantially more examples than in-distribution scenarios. Mechanistically, we identify a progressive emergence of layer specialization during training that correlates with generalization performance. Principal component analysis and attention pattern clustering reveal that transformers develop structured, hierarchically organized representations in specialized layers. These results demonstrate that transformers develop modular, interpretable mechanisms supporting compositional reasoning, linking internal algorithmic structure to observed behavioral capabilities.

[776] Symmetries in PAC-Bayesian Learning

Armin Beck, Peter Ochs

Main category: cs.LG

TL;DR: This paper extends generalization guarantees to non-compact symmetries and non-invariant data distributions using PAC-Bayes framework, showing improved performance for symmetric models beyond traditional compact group assumptions.

DetailsMotivation: Prior theoretical guarantees for symmetries in ML mainly focus on compact group symmetries and assume invariant data distributions, which rarely hold in real-world applications. The authors aim to extend these guarantees to more realistic settings.

Method: The authors build on the PAC-Bayes framework, adapting and tightening existing bounds (specifically McAllester’s PAC-Bayes bound) to handle non-compact symmetries like translations and non-invariant data distributions.

Result: The theory is validated on a rotated MNIST dataset with non-uniform rotation group, where the derived guarantees not only hold but also improve upon prior results, demonstrating practical benefits.

Conclusion: Symmetric models are preferable for symmetric data beyond the narrow setting of compact groups and invariant distributions, providing theoretical evidence for more general understanding of symmetries in machine learning.

Abstract: Symmetries are known to improve the empirical performance of machine learning models, yet theoretical guarantees explaining these gains remain limited. Prior work has focused mainly on compact group symmetries and often assumes that the data distribution itself is invariant, an assumption rarely satisfied in real-world applications. In this work, we extend generalization guarantees to the broader setting of non-compact symmetries, such as translations and to non-invariant data distributions. Building on the PAC-Bayes framework, we adapt and tighten existing bounds, demonstrating the approach on McAllester’s PAC-Bayes bound while showing that it applies to a wide range of PAC-Bayes bounds. We validate our theory with experiments on a rotated MNIST dataset with a non-uniform rotation group, where the derived guarantees not only hold but also improve upon prior results. These findings provide theoretical evidence that, for symmetric data, symmetric models are preferable beyond the narrow setting of compact groups and invariant distributions, opening the way to a more general understanding of symmetries in machine learning.

[777] DAMSDAN: Distribution-Aware Multi-Source Domain Adaptation Network for Cross-Domain EEG-based Emotion Recognition

Fo Hu, Can Wang, Qinxu Zheng, Xusheng Yang, Bin Zhou, Gang Li, Yu Sun, Wen-an Zhang

Main category: cs.LG

TL;DR: DAMSDAN is a distribution-aware multi-source domain adaptation network that addresses cross-domain EEG emotion recognition challenges by dynamically weighting source contributions and achieving fine-grained semantic alignment through prototype-based constraints and adversarial learning.

DetailsMotivation: Significant inter-individual variability limits generalization of EEG-based emotion recognition across domains. Two core challenges exist: dynamically modeling distribution heterogeneity across sources to reduce negative transfer, and achieving fine-grained semantic consistency for better class discrimination.

Method: Integrates prototype-based constraints with adversarial learning to create discriminative, domain-invariant emotion representations. Uses domain-aware source weighting based on MMD to dynamically estimate inter-domain shifts and reweight source contributions. Includes prototype-guided conditional alignment with dual pseudo-label interaction to enhance pseudo-label reliability and enable category-level fine-grained alignment.

Result: Achieved average accuracies of 94.86% (SEED) and 79.78% (SEED-IV) for cross-subject protocols, and 95.12% (SEED) and 83.15% (SEED-IV) for cross-session protocols. On FACED dataset, achieved 82.88% for cross-subject. Extensive ablations and interpretability analyses confirmed framework effectiveness.

Conclusion: DAMSDAN effectively addresses cross-domain EEG-based emotion recognition challenges through dynamic source weighting and fine-grained semantic alignment, demonstrating superior performance across multiple datasets and protocols.

Abstract: Significant inter-individual variability limits the generalization of EEG-based emotion recognition under cross-domain settings. We address two core challenges in multi-source adaptation: (1) dynamically modeling distributional heterogeneity across sources and quantifying their relevance to a target to reduce negative transfer; and (2) achieving fine-grained semantic consistency to strengthen class discrimination. We propose a distribution-aware multi-source domain adaptation network (DAMSDAN). DAMSDAN integrates prototype-based constraints with adversarial learning to drive the encoder toward discriminative, domain-invariant emotion representations. A domain-aware source weighting strategy based on maximum mean discrepancy (MMD) dynamically estimates inter-domain shifts and reweights source contributions. In addition, a prototype-guided conditional alignment module with dual pseudo-label interaction enhances pseudo-label reliability and enables category-level, fine-grained alignment, mitigating noise propagation and semantic drift. Experiments on SEED and SEED-IV show average accuracies of 94.86% and 79.78% for cross-subject, and 95.12% and 83.15% for cross-session protocols. On the large-scale FACED dataset, DAMSDAN achieves 82.88% (cross-subject). Extensive ablations and interpretability analyses corroborate the effectiveness of the proposed framework for cross-domain EEG-based emotion recognition.

[778] Disentanglement Beyond Static vs. Dynamic: A Benchmark and Evaluation Framework for Multi-Factor Sequential Representations

Tal Barami, Nimrod Berman, Ilan Naiman, Amos H. Hason, Rotem Ezra, Omri Azencot

Main category: cs.LG

TL;DR: This paper introduces the first standardized benchmark for multi-factor sequential disentanglement across six datasets, proposes a Latent Exploration Stage for automatic alignment, and demonstrates Vision-Language Models can automate evaluation.

DetailsMotivation: Real-world sequential data involves multiple interacting semantic factors over time, but prior work has focused on simpler two-factor settings, overlooking the inherently multi-factor nature of real data.

Method: Proposes a post-hoc Latent Exploration Stage to automatically align latent dimensions with semantic factors, and introduces a Koopman-inspired model. Also uses Vision-Language Models for automated dataset annotation and zero-shot evaluation.

Result: The Koopman-inspired model achieves state-of-the-art results on the new benchmark spanning video, audio, and time series datasets.

Conclusion: The contributions provide a robust and scalable foundation for advancing multi-factor sequential disentanglement research.

Abstract: Learning disentangled representations in sequential data is a key goal in deep learning, with broad applications in vision, audio, and time series. While real-world data involves multiple interacting semantic factors over time, prior work has mostly focused on simpler two-factor static and dynamic settings, primarily because such settings make data collection easier, thereby overlooking the inherently multi-factor nature of real-world data. We introduce the first standardized benchmark for evaluating multi-factor sequential disentanglement across six diverse datasets spanning video, audio, and time series. Our benchmark includes modular tools for dataset integration, model development, and evaluation metrics tailored to multi-factor analysis. We additionally propose a post-hoc Latent Exploration Stage to automatically align latent dimensions with semantic factors, and introduce a Koopman-inspired model that achieves state-of-the-art results. Moreover, we show that Vision-Language Models can automate dataset annotation and serve as zero-shot disentanglement evaluators, removing the need for manual labels and human intervention. Together, these contributions provide a robust and scalable foundation for advancing multi-factor sequential disentanglement.

[779] I-RAVEN-X: Benchmarking Generalization and Robustness of Analogical and Mathematical Reasoning in Large Language and Reasoning Models

Giacomo Camposampiero, Michael Hersche, Roger Wattenhofer, Abu Sebastian, Abbas Rahimi

Main category: cs.LG

TL;DR: I-RAVEN-X is a symbolic benchmark that extends I-RAVEN to evaluate generalization and robustness in analogical and mathematical reasoning for LLMs and LRMs by increasing operand complexity, attribute range, and adding perceptual uncertainty.

DetailsMotivation: To evaluate generalization and robustness in analogical and mathematical reasoning for Large Language Models (LLMs) and Large Reasoning Models (LRMs), addressing limitations in handling complex reasoning tasks.

Method: Extends I-RAVEN benchmark by increasing operand complexity, expanding attribute range, and introducing perceptual uncertainty to test reasoning capabilities under more challenging conditions.

Result: LRMs show improved productivity on longer reasoning relations and better systematicity on wider attribute ranges compared to LLMs, but both struggle with reasoning under uncertainty and exploring multiple probabilistic outcomes.

Conclusion: LRMs outperform LLMs on complex reasoning tasks but still face significant challenges with uncertainty reasoning and probabilistic exploration, highlighting areas for future improvement.

Abstract: We introduce I-RAVEN-X, a symbolic benchmark designed to evaluate generalization and robustness in analogical and mathematical reasoning for Large Language Models (LLMs) and Large Reasoning Models (LRMs). I-RAVEN-X extends I-RAVEN by increasing operand complexity, attribute range, and introducing perceptual uncertainty. Compared to LLMs, empirical results show that LRMs achieve improved productivity and systematicity on longer reasoning relations and wider attribute ranges, respectively. However, LRMs are still significantly challenged by reasoning under uncertainty and cannot effectively explore multiple probabilistic outcomes.

[780] Model Metamers Reveal Invariances in Graph Neural Networks

Wei Xu, Xiaoyi Jiang, Lixiang Xu, Dechao Tang

Main category: cs.LG

TL;DR: The paper investigates representational invariance in graph neural networks (GNNs) using a novel “metamers” generation technique, revealing excessive invariance compared to human-like behavior.

DetailsMotivation: To understand the invariance properties of GNNs and compare them with human brain mechanisms, as current studies show gaps between artificial and human neural network invariances.

Method: Developed a metamer generation technique that optimizes input graphs to match internal node activations of reference graphs, creating structurally different graphs with equivalent representations.

Result: Found extreme levels of representational invariance in classic GNN architectures; architectural and training modifications only partially mitigated this excessive invariance without bridging the human-like gap.

Conclusion: Current GNNs exhibit unique failure modes with excessive invariance, and the metamer approach provides a complementary benchmark for model evaluation beyond traditional metrics.

Abstract: In recent years, deep neural networks have been extensively employed in perceptual systems to learn representations endowed with invariances, aiming to emulate the invariance mechanisms observed in the human brain. However, studies in the visual and auditory domains have confirmed that significant gaps remain between the invariance properties of artificial neural networks and those of humans. To investigate the invariance behavior within graph neural networks (GNNs), we introduce a model ``metamers’’ generation technique. By optimizing input graphs such that their internal node activations match those of a reference graph, we obtain graphs that are equivalent in the model’s representation space, yet differ significantly in both structure and node features. Our theoretical analysis focuses on two aspects: the local metamer dimension for a single node and the activation-induced volume change of the metamer manifold. Utilizing this approach, we uncover extreme levels of representational invariance across several classic GNN architectures. Although targeted modifications to model architecture and training strategies can partially mitigate this excessive invariance, they fail to fundamentally bridge the gap to human-like invariance. Finally, we quantify the deviation between metamer graphs and their original counterparts, revealing unique failure modes of current GNNs and providing a complementary benchmark for model evaluation.

[781] The Graphon Limit Hypothesis: Understanding Neural Network Pruning via Infinite Width Analysis

Hoang Pham, The-Anh Ta, Tom Jacobs, Rebekka Burkholz, Long Tran-Thanh

Main category: cs.LG

TL;DR: A theoretical framework using graphons to analyze sparse neural networks, showing that pruning methods create specific connectivity patterns that affect trainability through Graphon Neural Tangent Kernel analysis.

DetailsMotivation: To understand why some sparse neural network structures are more trainable than others with the same sparsity level, and to develop a systematic theoretical approach to this fundamental problem.

Method: Proposed a graphon-based theoretical framework that characterizes sparse networks in infinite-width regime, introduced Graphon Limit Hypothesis, and derived Graphon Neural Tangent Kernel (Graphon NTK) to study training dynamics.

Result: Empirical evidence supports the Graphon Limit Hypothesis, and spectral analysis of Graphon NTK correlates with observed training dynamics, explaining varying convergence behaviors of different pruning methods.

Conclusion: The framework provides theoretical insights into how connectivity patterns impact trainability of sparse networks, offering a general approach for analyzing sparse network architectures.

Abstract: Sparse neural networks promise efficiency, yet training them effectively remains a fundamental challenge. Despite advances in pruning methods that create sparse architectures, understanding why some sparse structures are better trainable than others with the same level of sparsity remains poorly understood. Aiming to develop a systematic approach to this fundamental problem, we propose a novel theoretical framework based on the theory of graph limits, particularly graphons, that characterizes sparse neural networks in the infinite-width regime. Our key insight is that connectivity patterns of sparse neural networks induced by pruning methods converge to specific graphons as networks’ width tends to infinity, which encodes implicit structural biases of different pruning methods. We postulate the Graphon Limit Hypothesis and provide empirical evidence to support it. Leveraging this graphon representation, we derive a Graphon Neural Tangent Kernel (Graphon NTK) to study the training dynamics of sparse networks in the infinite width limit. Graphon NTK provides a general framework for the theoretical analysis of sparse networks. We empirically show that the spectral analysis of Graphon NTK correlates with observed training dynamics of sparse networks, explaining the varying convergence behaviours of different pruning methods. Our framework provides theoretical insights into the impact of connectivity patterns on the trainability of various sparse network architectures.

[782] Beyond Binary Out-of-Distribution Detection: Characterizing Distributional Shifts with Multi-Statistic Diffusion Trajectories

Achref Jaziri, Martin Rogmann, Martin Mundt, Visvanathan Ramesh

Main category: cs.LG

TL;DR: DISC introduces a diffusion-based method for OOD detection that goes beyond binary classification to characterize different types of out-of-distribution data using multi-dimensional statistical features.

DetailsMotivation: Current OOD detection methods collapse distributional shifts into single scalar scores, which is insufficient for contextualizing and exploiting OOD data. Different types of OOD data require different courses of action.

Method: DISC leverages the iterative denoising process of diffusion models to extract rich, multi-dimensional feature vectors that capture statistical discrepancies across multiple noise levels.

Result: Extensive experiments on image and tabular benchmarks show DISC matches or surpasses state-of-the-art detectors for OOD detection and crucially classifies OOD types, a capability largely absent from prior work.

Conclusion: The work enables a shift from simple binary OOD detection to more granular detection and characterization of out-of-distribution data types.

Abstract: Detecting out-of-distribution (OOD) data is critical for machine learning, be it for safety reasons or to enable open-ended learning. However, beyond mere detection, choosing an appropriate course of action typically hinges on the type of OOD data encountered. Unfortunately, the latter is generally not distinguished in practice, as modern OOD detection methods collapse distributional shifts into single scalar outlier scores. This work argues that scalar-based methods are thus insufficient for OOD data to be properly contextualized and prospectively exploited, a limitation we overcome with the introduction of DISC: Diffusion-based Statistical Characterization. DISC leverages the iterative denoising process of diffusion models to extract a rich, multi-dimensional feature vector that captures statistical discrepancies across multiple noise levels. Extensive experiments on image and tabular benchmarks show that DISC matches or surpasses state-of-the-art detectors for OOD detection and, crucially, also classifies OOD type, a capability largely absent from prior work. As such, our work enables a shift from simple binary OOD detection to a more granular detection.

[783] An Empirical Study of Lagrangian Methods in Safe Reinforcement Learning

Lindsay Spoor, Álvaro Serra-Gómez, Aske Plaat, Thomas Moerland

Main category: cs.LG

TL;DR: Analysis of Lagrange multipliers in safe reinforcement learning shows automated updates can recover optimal performance but exhibit oscillatory behavior, requiring PID control for stability.

DetailsMotivation: To understand the robustness of automated Lagrange multiplier updates in safe RL and their influence on performance across safety-critical domains.

Method: Analyzed optimality and stability of Lagrange multipliers using λ-profiles visualization across various tasks, comparing automated updates with PID-controlled updates.

Result: Automated multiplier updates can recover and sometimes exceed optimal performance found at λ*, but exhibit oscillatory behavior that can be mitigated with PID control (though requiring careful tuning).

Conclusion: Lagrangian methods in safe RL are highly sensitive to λ choice, automated updates can achieve good performance but require stabilization techniques, highlighting need for further research on stabilizing these methods.

Abstract: In safety-critical domains such as robotics, navigation and power systems, constrained optimization problems arise where maximizing performance must be carefully balanced with associated constraints. Safe reinforcement learning provides a framework to address these challenges, with Lagrangian methods being a popular choice. However, the effectiveness of Lagrangian methods crucially depends on the choice of the Lagrange multiplier $\lambda$, which governs the trade-off between return and constraint cost. A common approach is to update the multiplier automatically during training. Although this is standard in practice, there remains limited empirical evidence on the robustness of an automated update and its influence on overall performance. Therefore, we analyze (i) optimality and (ii) stability of Lagrange multipliers in safe reinforcement learning across a range of tasks. We provide $\lambda$-profiles that give a complete visualization of the trade-off between return and constraint cost of the optimization problem. These profiles show the highly sensitive nature of $\lambda$ and moreover confirm the lack of general intuition for choosing the optimal value $\lambda^$. Our findings additionally show that automated multiplier updates are able to recover and sometimes even exceed the optimal performance found at $\lambda^$ due to the vast difference in their learning trajectories. Furthermore, we show that automated multiplier updates exhibit oscillatory behavior during training, which can be mitigated through PID-controlled updates. However, this method requires careful tuning to achieve consistently better performance across tasks. This highlights the need for further research on stabilizing Lagrangian methods in safe reinforcement learning. The code used to reproduce our results can be found at https://github.com/lindsayspoor/Lagrangian_SafeRL.

[784] Latent Spaces Beyond Synthesis: From GANs to Diffusion Models

Ludovica Schaerf

Main category: cs.LG

TL;DR: This paper analyzes how diffusion models distribute representations across layers rather than using unified latent spaces, challenging traditional views of generative AI synthesis.

DetailsMotivation: To examine the conceptual shift from GANs/VAEs to diffusion models and challenge the assumption of unified internal representation spaces in generative visual models.

Method: Close readings of model architectures and experimental interventions in layerwise representations of diffusion models to analyze how representational labor is distributed.

Result: Diffusion models fragment the burden of representation across layers, demonstrating that generative processes emerge from specialized configurations rather than unified latent spaces.

Conclusion: Generative AI should be understood not as direct synthesis from compact latent spaces, but as emergent configurations of distributed specialized processes.

Abstract: This paper examines the evolving nature of internal representations in generative visual models, focusing on the conceptual and technical shift from GANs and VAEs to diffusion-based architectures. Drawing on Beatrice Fazi’s account of synthesis as the amalgamation of distributed representations, we propose a distinction between “synthesis in a strict sense”, where a compact latent space wholly determines the generative process, and “synthesis in a broad sense,” which characterizes models whose representational labor is distributed across layers. Through close readings of model architectures and a targeted experimental setup that intervenes in layerwise representations, we show how diffusion models fragment the burden of representation and thereby challenge assumptions of unified internal space. By situating these findings within media theoretical frameworks and critically engaging with metaphors such as the latent space and the Platonic Representation Hypothesis, we argue for a reorientation of how generative AI is understood: not as a direct synthesis of content, but as an emergent configuration of specialized processes.

[785] CEPerFed: Communication-Efficient Personalized Federated Learning for Multi-Pulse MRI Classification

Ludi Li, Junbin Mao, Hanhe Lin, Xu Tian, Fang-Xiang Wu, Jin Liu

Main category: cs.LG

TL;DR: CEPerFed is a communication-efficient personalized federated learning method for multi-pulse MRI classification that addresses data heterogeneity and communication overhead through historical gradient coordination and hierarchical SVD compression.

DetailsMotivation: Multi-pulse MRI classification requires large diverse data from multiple institutions while protecting privacy. Federated learning faces challenges with model convergence due to data heterogeneity and high communication overhead from transmitting large model parameters.

Method: CEPerFed incorporates client-side historical risk gradients and historical mean gradients to coordinate local and global optimization. It uses hierarchical SVD (HSVD) strategy to transmit only critical information for model updates, reducing communication overhead.

Result: Experiments on five classification tasks demonstrate the effectiveness of the CEPerFed method in improving model performance while reducing communication costs.

Conclusion: CEPerFed successfully addresses both data heterogeneity and communication efficiency challenges in federated learning for medical imaging applications, providing a practical solution for privacy-preserving multi-institutional collaboration.

Abstract: Multi-pulse magnetic resonance imaging (MRI) is widely utilized for clinical practice such as Alzheimer’s disease diagnosis. To train a robust model for multi-pulse MRI classification, it requires large and diverse data from various medical institutions while protecting privacy by preventing raw data sharing across institutions. Although federated learning (FL) is a feasible solution to address this issue, it poses challenges of model convergence due to the effect of data heterogeneity and substantial communication overhead due to large numbers of parameters transmitted within the model. To address these challenges, we propose CEPerFed, a communication-efficient personalized FL method. It mitigates the effect of data heterogeneity by incorporating client-side historical risk gradients and historical mean gradients to coordinate local and global optimization. The former is used to weight the contributions from other clients, enhancing the reliability of local updates, while the latter enforces consistency between local updates and the global optimization direction to ensure stable convergence across heterogeneous data distributions. To address the high communication overhead, we propose a hierarchical SVD (HSVD) strategy that transmits only the most critical information required for model updates. Experiments on five classification tasks demonstrate the effectiveness of the CEPerFed method. The code will be released upon acceptance at https://github.com/LD0416/CEPerFed.

[786] Exploration via Feature Perturbation in Contextual Bandits

Seouh-won Yi, Min-hwan Oh

Main category: cs.LG

TL;DR: Feature perturbation injects randomness directly into feature inputs rather than parameters or rewards, achieving improved regret bounds for generalized linear bandits while being computationally efficient and extensible to non-parametric models.

DetailsMotivation: Existing randomized bandit algorithms suffer from suboptimal regret bounds (O(d^{3/2}√T)) and computational inefficiency due to parameter sampling. The goal is to develop a method that achieves better theoretical guarantees while maintaining practical efficiency.

Method: Feature perturbation algorithm that injects randomness directly into feature inputs instead of randomizing unknown parameters or adding noise to rewards. This avoids parameter sampling while maintaining exploration.

Result: Achieves O(d√T) worst-case regret bound for generalized linear bandits, improving upon the typical O(d^{3/2}√T) regret of existing randomized methods. The method is computationally efficient and naturally extends to non-parametric or neural network models.

Conclusion: Feature perturbation provides a unified approach that combines strong practical performance with best-known theoretical guarantees, surpassing existing methods while being computationally efficient and broadly applicable.

Abstract: We propose feature perturbation, a simple yet powerful technique that injects randomness directly into feature inputs, instead of randomizing unknown parameters or adding noise to rewards. Remarkably, this algorithm achieves $\tilde{\mathcal{O}}(d\sqrt{T})$ worst-case regret bound for generalized linear bandits, while avoiding the $\tilde{\mathcal{O}}(d^{3/2}\sqrt{T})$ regret typical of existing randomized bandit algorithms. Because our algorithm eschews parameter sampling, it is both computationally efficient and naturally extends to non-parametric or neural network models. We verify these advantages through empirical evaluations, demonstrating that feature perturbation not only surpasses existing methods but also unifies strong practical performance with best-known theoretical guarantees.

[787] Finite-Time Bounds for Average-Reward Fitted Q-Iteration

Jongmin Lee, Ernest K. Ryu

Main category: cs.LG

TL;DR: First sample complexity results for average-reward offline RL with function approximation for weakly communicating MDPs, introducing Anchored Fitted Q-Iteration with anchor mechanism.

DetailsMotivation: Prior work on average-reward offline RL had restrictive assumptions like ergodicity or linearity, while weakly communicating MDPs are much milder assumptions.

Method: Anchored Fitted Q-Iteration that combines standard FQI with an anchor mechanism (interpreted as weight decay).

Result: Established finite-time sample complexity results for average-reward offline RL with function approximation.

Conclusion: Anchor mechanism is crucial for finite-time analysis in average-reward setting and enables extension to single-trajectory datasets.

Abstract: Although there is an extensive body of work characterizing the sample complexity of discounted-return offline RL with function approximations, prior work on the average-reward setting has received significantly less attention, and existing approaches rely on restrictive assumptions, such as ergodicity or linearity of the MDP. In this work, we establish the first sample complexity results for average-reward offline RL with function approximation for weakly communicating MDPs, a much milder assumption. To this end, we introduce Anchored Fitted Q-Iteration, which combines the standard Fitted Q-Iteration with an anchor mechanism. We show that the anchor, which can be interpreted as a form of weight decay, is crucial for enabling finite-time analysis in the average-reward setting. We also extend our finite-time analysis to the setup where the dataset is generated from a single-trajectory rather than IID transitions, again leveraging the anchor mechanism.

[788] MILES: Modality-Informed Learning Rate Scheduler for Balancing Multimodal Learning

Alejandro Guerra-Manzanares, Farah E. Shamout

Main category: cs.LG

TL;DR: MILES is a learning rate scheduler that dynamically adjusts rates during multimodal training to balance modality usage, preventing overfitting to single modalities and improving both multimodal and unimodal performance.

DetailsMotivation: Multimodal networks often suffer from modality overfitting, where they rely excessively on one modality, leading to suboptimal performance and marginal improvements over unimodal models.

Method: MILES uses modality-wise conditional utilization rates to dynamically adjust learning rates during training, balancing the speed of learning from each modality in multimodal joint fusion models.

Result: MILES outperforms seven state-of-the-art baselines across four multimodal tasks and fusion methods, effectively balancing modality usage and improving both multimodal performance and unimodal encoder strength.

Conclusion: Balancing multimodal learning through dynamic learning rate scheduling significantly improves model performance, enabling better multimodal predictions and stronger unimodal encoders for handling missing modalities.

Abstract: The aim of multimodal neural networks is to combine diverse data sources, referred to as modalities, to achieve enhanced performance compared to relying on a single modality. However, training of multimodal networks is typically hindered by modality overfitting, where the network relies excessively on one of the available modalities. This often yields sub-optimal performance, hindering the potential of multimodal learning and resulting in marginal improvements relative to unimodal models. In this work, we present the Modality-Informed Learning ratE Scheduler (MILES) for training multimodal joint fusion models in a balanced manner. MILES leverages the differences in modality-wise conditional utilization rates during training to effectively balance multimodal learning. The learning rate is dynamically adjusted during training to balance the speed of learning from each modality by the multimodal model, aiming for enhanced performance in both multimodal and unimodal predictions. We extensively evaluate MILES on four multimodal joint fusion tasks and compare its performance to seven state-of-the-art baselines. Our results show that MILES outperforms all baselines across all tasks and fusion methods considered in our study, effectively balancing modality usage during training. This results in improved multimodal performance and stronger modality encoders, which can be leveraged when dealing with unimodal samples or absent modalities. Overall, our work highlights the impact of balancing multimodal learning on improving model performance.

[789] On-the-Fly OVD Adaptation with FLAME: Few-shot Localization via Active Marginal-Samples Exploration

Yehonathan Refael, Amit Aides, Aviad Barzilai, George Leifman, Genady Beryozkin, Vered Silverman, Bolous Jaber, Tomer Shekel

Main category: cs.LG

TL;DR: A cascaded approach combining open-vocabulary detection with few-shot learning for remote sensing, using FLAME active learning for efficient adaptation.

DetailsMotivation: Open-vocabulary detection models struggle with fine-grained class distinctions in specialized domains like remote sensing due to language ambiguity, limiting practical applications like illegal fishing monitoring.

Method: Cascaded framework: first uses zero-shot OVD model for high-recall proposals, then refines with lightweight few-shot classifier trained on minimal user annotations using FLAME active learning strategy for sample selection.

Result: Achieves state-of-the-art performance on RS benchmarks, enables instant adaptation within less than a minute, significantly faster than alternatives, without costly full-model fine-tuning.

Conclusion: Establishes practical and resource-efficient framework for adapting foundation models to specific user needs in specialized domains like remote sensing.

Abstract: Open-vocabulary object detection (OVD) models offer remarkable flexibility by detecting objects from arbitrary text queries. However, their zero-shot performance in specialized domains like Remote Sensing (RS) is often compromised by the inherent ambiguity of natural language, limiting critical downstream applications. For instance, an OVD model may struggle to distinguish between fine-grained classes such as “fishing boat” and “yacht” since their embeddings are similar and often inseparable. This can hamper specific user goals, such as monitoring illegal fishing, by producing irrelevant detections. To address this, we propose a cascaded approach that couples the broad generalization of a large pre-trained OVD model with a lightweight few-shot classifier. Our method first employs the zero-shot model to generate high-recall object proposals. These proposals are then refined for high precision by a compact classifier trained in real-time on only a handful of user-annotated examples - drastically reducing the high costs of RS imagery annotation.The core of our framework is FLAME, a one-step active learning strategy that selects the most informative samples for training. FLAME identifies, on the fly, uncertain marginal candidates near the decision boundary using density estimation, followed by clustering to ensure sample diversity. This efficient sampling technique achieves high accuracy without costly full-model fine-tuning and enables instant adaptation, within less then a minute, which is significantly faster than state-of-the-art alternatives.Our method consistently surpasses state-of-the-art performance on RS benchmarks, establishing a practical and resource-efficient framework for adapting foundation models to specific user needs.

[790] RINS-T: Robust Implicit Neural Solvers for Time Series Linear Inverse Problems

Keivan Faghih Niresi, Zepeng Zhang, Olga Fink

Main category: cs.LG

TL;DR: RINS-T is a novel deep prior framework for time series linear inverse problems that achieves high recovery performance without pretraining data, using neural networks as implicit priors with robust optimization techniques.

DetailsMotivation: Time series data often suffer from corruption like missing values, noise, and outliers, which challenge forecasting and anomaly detection. Existing deep learning methods require extensive pretraining and struggle with distribution shifts.

Method: Proposes RINS-T framework using neural networks as implicit priors with robust optimization. Introduces three key innovations: guided input initialization, input perturbation, and convex output combination techniques for optimization stability.

Result: RINS-T achieves high recovery performance without requiring pretraining data, is resilient to outliers, and relaxes reliance on Gaussian noise assumptions.

Conclusion: RINS-T provides a flexible and effective solution for complex real-world time series challenges, with demonstrated robustness and optimization stability improvements.

Abstract: Time series data are often affected by various forms of corruption, such as missing values, noise, and outliers, which pose significant challenges for tasks such as forecasting and anomaly detection. To address these issues, inverse problems focus on reconstructing the original signal from corrupted data by leveraging prior knowledge about its underlying structure. While deep learning methods have demonstrated potential in this domain, they often require extensive pretraining and struggle to generalize under distribution shifts. In this work, we propose RINS-T (Robust Implicit Neural Solvers for Time Series Linear Inverse Problems), a novel deep prior framework that achieves high recovery performance without requiring pretraining data. RINS-T leverages neural networks as implicit priors and integrates robust optimization techniques, making it resilient to outliers while relaxing the reliance on Gaussian noise assumptions. To further improve optimization stability and robustness, we introduce three key innovations: guided input initialization, input perturbation, and convex output combination techniques. Each of these contributions strengthens the framework’s optimization stability and robustness. These advancements make RINS-T a flexible and effective solution for addressing complex real-world time series challenges. Our code is available at https://github.com/EPFL-IMOS/RINS-T.

[791] S4ECG: Exploring the impact of long-range interactions for arrhythmia prediction

Tiezhi Wang, Wilhelm Haverkamp, Nils Strodthoff

Main category: cs.LG

TL;DR: S4ECG is a novel deep learning architecture using structured state space models for multi-epoch ECG arrhythmia classification, achieving superior performance over single-epoch approaches with optimal 10-20 minute temporal dependency windows.

DetailsMotivation: Conventional ECG analysis methods fail to capture the simultaneous interplay between global trends and local waveform features at high temporal resolution, limiting comprehensive analysis of cardiac dynamics.

Method: Introduced S4ECG - a deep learning architecture leveraging structured state space models for multi-epoch arrhythmia classification, enabling joint analysis across multiple time epochs.

Result: Multi-epoch predictions outperformed single-epoch approaches by 1.0-11.6% in macro-AUROC, with atrial fibrillation specificity improving from 0.718-0.979 to 0.967-0.998, showing superior in-distribution performance and enhanced out-of-distribution robustness.

Conclusion: This work enables a paradigm shift toward temporally-aware arrhythmia detection algorithms, particularly beneficial for complex arrhythmias like atrial fibrillation and atrial flutter, with optimal performance achieved using 10-20 minute temporal dependency windows.

Abstract: The electrocardiogram (ECG) exemplifies biosignal-based time series with continuous, temporally ordered structure reflecting cardiac physiological and pathophysiological dynamics. Detailed analysis of these dynamics has proven challenging, as conventional methods capture either global trends or local waveform features but rarely their simultaneous interplay at high temporal resolution. To bridge global and local signal analysis, we introduce S4ECG, a novel deep learning architecture leveraging structured state space models for multi-epoch arrhythmia classification. Our joint multi-epoch predictions significantly outperform single-epoch approaches by 1.0-11.6% in macro-AUROC, with atrial fibrillation specificity improving from 0.718-0.979 to 0.967-0.998, demonstrating superior performance in-distribution and enhanced out-of-distribution robustness. Systematic investigation reveals optimal temporal dependency windows spanning 10-20 minutes for peak performance. This work contributes to a paradigm shift toward temporally-aware arrhythmia detection algorithms, opening new possibilities for ECG interpretation, in particular for complex arrhythmias like atrial fibrillation and atrial flutter.

[792] A Conditional Diffusion Model for Probabilistic Prediction of Battery Capacity Degradation

Hequn Li, Zhongwei Deng, Chunlin Jiang, Yvxin He andZhansheng Ning

Main category: cs.LG

TL;DR: A novel CDUA method integrates diffusion models and attention mechanisms for accurate lithium-ion battery capacity prediction with uncertainty quantification, achieving 0.94% MAE and 1.14% RMSE on real-world vehicle data.

DetailsMotivation: Accurate prediction of lithium-ion battery capacity and uncertainty is essential for reliable battery management but challenging due to stochastic aging processes.

Method: Proposes CDUA model combining diffusion-based generative modeling with attention mechanisms. Uses Pearson correlation and XGBoost for feature selection, then trains a U-Net with self-attention and denoising network for capacity reconstruction.

Result: Achieves relative MAE of 0.94% and RMSE of 1.14% with 95% confidence interval width of 3.74%. Outperforms existing mainstream approaches in comparative experiments.

Conclusion: CDUA provides both accurate capacity estimation and reliable uncertainty quantification, demonstrating robustness and superior performance for battery management applications.

Abstract: Accurate prediction of lithium-ion battery capacity and its associated uncertainty is essential for reliable battery management but remains challenging due to the stochastic nature of aging. This paper presents a novel method, termed the Condition Diffusion U-Net with Attention (CDUA), which integrates feature engineering and deep learning to address this challenge. The proposed approach employs a diffusion-based generative model for time-series forecasting and incorporates attention mechanisms to enhance predictive performance. Battery capacity is first derived from real-world vehicle operation data. The most relevant features are then identified using the Pearson correlation coefficient and the XGBoost algorithm. These features are used to train the CDUA model, which comprises two core components: (1) a contextual U-Net with self-attention to capture complex temporal dependencies, and (2) a denoising network to reconstruct accurate capacity values from noisy observations. Experimental validation on the real-world vehicle data demonstrates that the proposed CDUA model achieves a relative Mean Absolute Error (MAE) of 0.94% and a relative Root Mean Square Error (RMSE) of 1.14%, with a narrow 95% confidence interval of 3.74% in relative width. These results confirm that CDUA provides both accurate capacity estimation and reliable uncertainty quantification. Comparative experiments further verify its robustness and superior performance over existing mainstream approaches.

[793] Closing the Sim2Real Performance Gap in RL

Akhil S Anand, Shambhuraj Sawant, Jasper Hoffmann, Dirk Reinhardt, Sebastien Gros

Main category: cs.LG

TL;DR: A bi-level RL framework that directly adapts simulator parameters based on real-world performance to close the Sim2Real gap, rather than relying on proxy metrics like simulator accuracy.

DetailsMotivation: Current Sim2Real RL methods use simulator accuracy and variability as proxies for real-world performance, but these metrics don't necessarily correlate with actual policy performance when deployed in reality, leading to significant performance drops.

Method: Proposes a bi-level RL framework: inner-level RL trains policies purely in simulation, while outer-level RL adapts simulation model and reward parameters to maximize real-world performance of the in-sim trained policy.

Result: The paper derives and validates mathematical tools needed to develop bi-level RL algorithms that can effectively close the Sim2Real performance gap.

Conclusion: Direct adaptation of simulator parameters based on real-world performance provides a more effective approach to closing the Sim2Real gap compared to traditional proxy-based methods.

Abstract: Sim2Real aims at training policies in high-fidelity simulation environments and effectively transferring them to the real world. Despite the developments of accurate simulators and Sim2Real RL approaches, the policies trained purely in simulation often suffer significant performance drops when deployed in real environments. This drop is referred to as the Sim2Real performance gap. Current Sim2Real RL methods optimize the simulator accuracy and variability as proxies for real-world performance. However, these metrics do not necessarily correlate with the real-world performance of the policy as established theoretically and empirically in the literature. We propose a novel framework to address this issue by directly adapting the simulator parameters based on real-world performance. We frame this problem as a bi-level RL framework: the inner-level RL trains a policy purely in simulation, and the outer-level RL adapts the simulation model and in-sim reward parameters to maximize real-world performance of the in-sim policy. We derive and validate in simple examples the mathematical tools needed to develop bi-level RL algorithms that close the Sim2Real performance gap.

[794] Diffusion Models as Dataset Distillation Priors

Duo Su, Huyu Wu, Huanran Chen, Yiming Shi, Yuzhu Wang, Xi Ye, Jun Zhu

Main category: cs.LG

TL;DR: DAP leverages diffusion models’ inherent representativeness prior to improve dataset distillation by guiding the reverse diffusion process with a Mercer kernel-based similarity measure, achieving better diversity and generalization without retraining.

DetailsMotivation: Existing generative dataset distillation methods overlook the inherent representativeness prior in diffusion models and require external constraints to enhance data quality, failing to achieve the trifecta of diversity, generalization, and representativeness.

Method: Propose Diffusion As Priors (DAP) that formalizes representativeness by quantifying similarity between synthetic and real data using a Mercer kernel, then introduces this prior as guidance to steer the reverse diffusion process without retraining.

Result: Extensive experiments on ImageNet-1K and subsets show DAP outperforms state-of-the-art methods in generating high-fidelity datasets and achieves superior cross-architecture generalization.

Conclusion: DAP establishes a theoretical connection between diffusion priors and dataset distillation objectives while providing a practical, training-free framework for improving distilled dataset quality.

Abstract: Dataset distillation aims to synthesize compact yet informative datasets from large ones. A significant challenge in this field is achieving a trifecta of diversity, generalization, and representativeness in a single distilled dataset. Although recent generative dataset distillation methods adopt powerful diffusion models as their foundation models, the inherent representativeness prior in diffusion models is overlooked. Consequently, these approaches often necessitate the integration of external constraints to enhance data quality. To address this, we propose Diffusion As Priors (DAP), which formalizes representativeness by quantifying the similarity between synthetic and real data in feature space using a Mercer kernel. We then introduce this prior as guidance to steer the reverse diffusion process, enhancing the representativeness of distilled samples without any retraining. Extensive experiments on large-scale datasets, such as ImageNet-1K and its subsets, demonstrate that DAP outperforms state-of-the-art methods in generating high-fidelity datasets while achieving superior cross-architecture generalization. Our work not only establishes a theoretical connection between diffusion priors and the objectives of dataset distillation but also provides a practical, training-free framework for improving the quality of the distilled dataset.

[795] Deeper with Riemannian Geometry: Overcoming Oversmoothing and Oversquashing for Graph Foundation Models

Li Sun, Zhenhao Huang, Ming Zhang, Philip S. Yu

Main category: cs.LG

TL;DR: The paper proposes GBN, a local approach using Riemannian geometry and Robin boundary conditions to adaptively adjust message passing in MPNNs, addressing oversquashing and oversmoothing while maintaining performance in deep networks.

DetailsMotivation: Existing global approaches to fix oversquashing and oversmoothing in MPNNs are suboptimal as they may help some regions but harm others. Theoretical analysis shows that increasing spectral gap leads to gradient vanishing, undermining message passing effectiveness.

Method: Connects local Riemannian geometry with MPNNs and establishes a nonhomogeneous boundary condition. Designs GBN network with local bottleneck adjustment based on Robin condition, providing theoretical guarantees.

Result: Extensive experiments on homophilic and heterophilic graphs demonstrate GBN’s expressiveness. GBN maintains performance without degradation even with network depth exceeding 256 layers.

Conclusion: The proposed local approach using Riemannian geometry and adaptive message passing effectively addresses both oversquashing and oversmoothing issues in MPNNs, enabling deep networks without performance degradation.

Abstract: Message Passing Neural Networks (MPNNs) is the building block of graph foundation models, but fundamentally suffer from oversmoothing and oversquashing. There has recently been a surge of interest in fixing both issues. Existing efforts primarily adopt global approaches, which may be beneficial in some regions but detrimental in others, ultimately leading to the suboptimal expressiveness. In this paper, we begin by revisiting oversquashing through a global measure – spectral gap $\lambda$ – and prove that the increase of $\lambda$ leads to gradient vanishing with respect to the input features, thereby undermining the effectiveness of message passing. Motivated by such theoretical insights, we propose a \textbf{local} approach that adaptively adjusts message passing based on local structures. To achieve this, we connect local Riemannian geometry with MPNNs, and establish a novel nonhomogeneous boundary condition to address both oversquashing and oversmoothing. Building on the Robin condition, we design a GBN network with local bottleneck adjustment, coupled with theoretical guarantees. Extensive experiments on homophilic and heterophilic graphs show the expressiveness of GBN. Furthermore, GBN does not exhibit performance degradation even when the network depth exceeds $256$ layers.

[796] Explainable AI for microseismic event detection

Ayrat Abdullin, Denis Anikiev, Umair bin Waheed

Main category: cs.LG

TL;DR: Applied XAI techniques (Grad-CAM and SHAP) to interpret PhaseNet’s seismic event detection, revealing network attention aligns with P/S-wave arrivals. Introduced SHAP-gated inference that improved performance to F1-score 0.98.

DetailsMotivation: Address concerns about black-box nature of deep neural networks like PhaseNet in critical seismic applications by making decisions interpretable and improving reliability.

Method: Used Grad-CAM to visualize network attention and SHAP to quantify feature contributions. Developed SHAP-gated inference scheme combining model output with explanation-based metric.

Result: SHAP-gated model achieved F1-score 0.98 (precision 0.99, recall 0.97) on 9,000 test waveforms, outperforming baseline PhaseNet (F1-score 0.97) with enhanced noise robustness.

Conclusion: XAI can both interpret deep learning models and directly enhance their performance, providing a template for building trust in automated seismic detectors.

Abstract: Deep neural networks like PhaseNet show high accuracy in detecting microseismic events, but their black-box nature is a concern in critical applications. We apply explainable AI (XAI) techniques, such as Gradient-weighted Class Activation Mapping (Grad-CAM) and Shapley Additive Explanations (SHAP), to interpret the PhaseNet model’s decisions and improve its reliability. Grad-CAM highlights that the network’s attention aligns with P- and S-wave arrivals. SHAP values quantify feature contributions, confirming that vertical-component amplitudes drive P-phase picks while horizontal components dominate S-phase picks, consistent with geophysical principles. Leveraging these insights, we introduce a SHAP-gated inference scheme that combines the model’s output with an explanation-based metric to reduce errors. On a test set of 9,000 waveforms, the SHAP-gated model achieved an F1-score of 0.98 (precision 0.99, recall 0.97), outperforming the baseline PhaseNet (F1-score 0.97) and demonstrating enhanced robustness to noise. These results show that XAI can not only interpret deep learning models but also directly enhance their performance, providing a template for building trust in automated seismic detectors.

[797] CrossStateECG: Multi-Scale Deep Convolutional Network with Attention for Rest-Exercise ECG Biometrics

Dan Zheng, Jing Feng, Juan Liu

Main category: cs.LG

TL;DR: CrossStateECG is a robust ECG-based authentication model that addresses performance decline in rest-exercise scenarios by combining multi-scale deep convolutional feature extraction with attention mechanisms, achieving high accuracy across different physiological states.

DetailsMotivation: Current ECG biometrics research mainly focuses on resting-state conditions, leaving the performance decline in rest-exercise scenarios largely unresolved, which limits practical applications in dynamic real-world settings.

Method: The model creatively combines multi-scale deep convolutional feature extraction with attention mechanisms to ensure strong identification across different physiological states.

Result: Achieved 92.50% accuracy in Rest-to-Exercise scenario, 94.72% in Exercise-to-Rest scenario, 99.94% in Rest-to-Rest scenarios, and 97.85% in Mixed-to-Mixed scenarios. Validations on ECG-ID and MIT-BIH datasets confirmed generalization abilities.

Conclusion: CrossStateECG demonstrates exceptional performance across state combinations and shows strong generalization, making it a practical solution for post-exercise ECG-based authentication in dynamic real-world settings.

Abstract: Current research in Electrocardiogram (ECG) biometrics mainly emphasizes resting-state conditions, leaving the performance decline in rest-exercise scenarios largely unresolved. This paper introduces CrossStateECG, a robust ECG-based authentication model explicitly tailored for cross-state (rest-exercise) conditions. The proposed model creatively combines multi-scale deep convolutional feature extraction with attention mechanisms to ensure strong identification across different physiological states. Experimental results on the exercise-ECGID dataset validate the effectiveness of CrossStateECG, achieving an identification accuracy of 92.50% in the Rest-to-Exercise scenario (training on resting ECG and testing on post-exercise ECG) and 94.72% in the Exercise-to-Rest scenario (training on post-exercise ECG and testing on resting ECG). Furthermore, CrossStateECG demonstrates exceptional performance across both state combinations, reaching an accuracy of 99.94% in Rest-to-Rest scenarios and 97.85% in Mixed-to-Mixed scenarios. Additional validations on the ECG-ID and MIT-BIH datasets further confirmed the generalization abilities of CrossStateECG, underscoring its potential as a practical solution for post-exercise ECG-based authentication in dynamic real-world settings.

[798] Towards geological inference with process-based and deep generative modeling, part 2: inversion of fluvial deposits and latent-space disentanglement

Guillaume Rongier, Luk Peeters

Main category: cs.LG

TL;DR: GANs trained for fluvial deposit generation can be inverted to match well and seismic data, but latent space entanglement causes inversion challenges. Fine-tuning GANs locally restructures latent space and reduces mismatches to acceptable levels.

DetailsMotivation: High costs and uncertainties in subsurface decision-making make scalable data acquisition difficult. Embedding geological knowledge into predictive models offers a valuable alternative to traditional methods.

Method: Used generative adversarial networks (GANs) trained to produce fluvial deposits, applied four inversion approaches to match well and seismic data across test samples with 4, 8, and 20 wells. Explored label conditioning, latent overparameterization, and fine-tuning to restructure latent space.

Result: Inversion approaches struggled to match well data, especially with more wells or when test samples diverged from training data. Key bottleneck was entangled latent representation where similar sedimentological features weren’t close in latent space. Fine-tuning GANs locally reduced mismatches to acceptable levels for all test cases.

Conclusion: GANs can handle tasks required for integration into geomodeling workflows, but need further assessment of robustness and best practices for leveraging them in geological interpretation. Fine-tuning shows promise but depends on initial partially successful inversion.

Abstract: High costs and uncertainties make subsurface decision-making challenging, as acquiring new data is rarely scalable. Embedding geological knowledge directly into predictive models offers a valuable alternative. A joint approach enables just that: process-based models that mimic geological processes can help train generative models that make predictions more efficiently. This study explores whether a generative adversarial network (GAN) - a type of deep-learning algorithm for generative modeling - trained to produce fluvial deposits can be inverted to match well and seismic data. Four inversion approaches applied to three test samples with 4, 8, and 20 wells struggled to match these well data, especially as the well number increased or as the test sample diverged from the training data. The key bottleneck lies in the GAN’s latent representation: it is entangled, so samples with similar sedimentological features are not necessarily close in the latent space. Label conditioning or latent overparameterization can partially disentangle the latent space during training, although not yet sufficiently for a successful inversion. Fine-tuning the GAN to restructure the latent space locally reduces mismatches to acceptable levels for all test cases, with and without seismic data. But this approach depends on an initial, partially successful inversion step, which influences the quality and diversity of the final samples. Overall, GANs can already handle the tasks required for their integration into geomodeling workflows. We still need to further assess their robustness, and how to best leverage them in support of geological interpretation.

[799] Prediction of Sea Ice Velocity and Concentration in the Arctic Ocean using Physics-informed Neural Network

Younghyun Koo, Maryam Rahnemoonfar

Main category: cs.LG

TL;DR: Physics-informed neural networks (PINN) improve sea ice velocity and concentration predictions by integrating physical knowledge into machine learning models, outperforming purely data-driven approaches.

DetailsMotivation: Traditional data-driven ML models have limitations in generalizability and physical consistency, especially as Arctic sea ice conditions change rapidly with thinning ice and accelerated melting.

Method: Developed PINN strategies using Hierarchical Information-sharing U-net (HIS-Unet) architecture, incorporating physics loss function and activation function to ensure physically plausible outputs.

Result: PINN model outperformed fully data-driven model in daily SIV and SIC predictions, even with limited training data, particularly improving SIC predictions during melting/early freezing seasons and near fast-moving ice regions.

Conclusion: Physics-informed neural networks provide more reliable sea ice predictions by combining data-driven learning with physical constraints, making them suitable for rapidly changing Arctic conditions.

Abstract: As an increasing amount of remote sensing data becomes available in the Arctic Ocean, data-driven machine learning (ML) techniques are becoming widely used to predict sea ice velocity (SIV) and sea ice concentration (SIC). However, fully data-driven ML models have limitations in generalizability and physical consistency due to their excessive reliance on the quantity and quality of training data. In particular, as Arctic sea ice entered a new phase with thinner ice and accelerated melting, there is a possibility that an ML model trained with historical sea ice data cannot fully represent the dynamically changing sea ice conditions in the future. In this study, we develop physics-informed neural network (PINN) strategies to integrate physical knowledge of sea ice into the ML model. Based on the Hierarchical Information-sharing U-net (HIS-Unet) architecture, we incorporate the physics loss function and the activation function to produce physically plausible SIV and SIC outputs. Our PINN model outperforms the fully data-driven model in the daily predictions of SIV and SIC, even when trained with a small number of samples. The PINN approach particularly improves SIC predictions in melting and early freezing seasons and near fast-moving ice regions.

[800] Unified Privacy Guarantees for Decentralized Learning via Matrix Factorization

Aurélien Bellet, Edwige Cyffers, Davide Frey, Romaric Gaudel, Dimitri Lerévérend, François Taïani

Main category: cs.LG

TL;DR: This paper presents MAFALDA-SGD, a gossip-based decentralized learning algorithm with user-level correlated noise that improves privacy-utility trade-offs using Matrix Factorization for tighter differential privacy accounting.

DetailsMotivation: Current differential privacy accounting methods for decentralized learning show worse privacy-utility trade-offs than centralized training, likely due to limitations in existing DP accounting approaches.

Method: The paper generalizes Matrix Factorization-based DP accounting from centralized settings to decentralized learning, providing a unified formulation for standard DL algorithms and trust models. It introduces MAFALDA-SGD, a gossip-based algorithm with user-level correlated noise.

Result: MAFALDA-SGD outperforms existing methods on both synthetic and real-world graphs, demonstrating improved privacy-utility trade-offs through tighter privacy accounting.

Conclusion: Matrix Factorization enables tighter privacy accounting for decentralized learning, allowing development of new algorithms like MAFALDA-SGD that achieve better privacy-utility trade-offs than existing approaches.

Abstract: Decentralized Learning (DL) enables users to collaboratively train models without sharing raw data by iteratively averaging local updates with neighbors in a network graph. This setting is increasingly popular for its scalability and its ability to keep data local under user control. Strong privacy guarantees in DL are typically achieved through Differential Privacy (DP), with results showing that DL can even amplify privacy by disseminating noise across peer-to-peer communications. Yet in practice, the observed privacy-utility trade-off often appears worse than in centralized training, which may be due to limitations in current DP accounting methods for DL. In this paper, we show that recent advances in centralized DP accounting based on Matrix Factorization (MF) for analyzing temporal noise correlations can also be leveraged in DL. By generalizing existing MF results, we show how to cast both standard DL algorithms and common trust models into a unified formulation. This yields tighter privacy accounting for existing DP-DL algorithms and provides a principled way to develop new ones. To demonstrate the approach, we introduce MAFALDA-SGD, a gossip-based DL algorithm with user-level correlated noise that outperforms existing methods on synthetic and real-world graphs.

[801] Local properties of neural networks through the lens of layer-wise Hessians

Maxim Bolshim, Alexander Kugaevskikh

Main category: cs.LG

TL;DR: The paper introduces a methodology using layer-wise Hessian matrices to analyze neural networks, revealing patterns related to overfitting, underparameterization, and expressivity through spectral properties.

DetailsMotivation: To develop a formal tool for characterizing the local geometry of parameter space in neural networks and connect optimization geometry with functional behavior.

Method: Defines local Hessian matrices for each layer as second derivatives with respect to parameters, then analyzes spectral properties (eigenvalue distributions) across 111 experiments on 37 datasets.

Result: Shows consistent structural regularities in local Hessian evolution during training and correlations between Hessian spectra and generalization performance.

Conclusion: Establishes a foundation for using local geometric analysis to guide neural network diagnosis and design, offering practical insights for improving architectures and training stability.

Abstract: We introduce a methodology for analyzing neural networks through the lens of layer-wise Hessian matrices. The local Hessian of each functional block (layer) is defined as the matrix of second derivatives of a scalar function with respect to the parameters of that layer. This concept provides a formal tool for characterizing the local geometry of the parameter space. We show that the spectral properties of local Hessians, such as the distribution of eigenvalues, reveal quantitative patterns associated with overfitting, underparameterization, and expressivity in neural network architectures. We conduct an extensive empirical study involving 111 experiments across 37 datasets. The results demonstrate consistent structural regularities in the evolution of local Hessians during training and highlight correlations between their spectra and generalization performance. These findings establish a foundation for using local geometric analysis to guide the diagnosis and design of deep neural networks. The proposed framework connects optimization geometry with functional behavior and offers practical insight for improving network architectures and training stability.

[802] Unbiased Gradient Low-Rank Projection

Rui Pan, Yang Luo, Yuxing Liu, Yang You, Tong Zhang

Main category: cs.LG

TL;DR: GaLore Unbiased with Muon (GUM) is a novel memory-efficient optimization method that combines GaLore’s gradient low-rank projection with Muon’s layerwise sampling to eliminate bias, achieving better convergence guarantees and performance than both GaLore and full-parameter training.

DetailsMotivation: To address the lack of convergence guarantees and performance gaps in existing memory-efficient optimization methods like GaLore, which suffer from inherent biases introduced by low-rank projection techniques.

Method: Combines GaLore’s gradient low-rank projection mechanism with the Muon algorithm’s layerwise sampling technique to create an unbiased optimization method called GUM.

Result: Theoretical convergence guarantees matching Muon algorithm, empirical improvements over GaLore in LLM fine-tuning and pretraining, and even better performance than full-parameter training with more uniform knowledge distribution and efficient parameter space utilization.

Conclusion: GUM successfully addresses the bias problem in low-rank optimization methods while maintaining memory efficiency, achieving superior performance through more uniform knowledge distribution and better memorization capabilities.

Abstract: Memory-efficient optimization is critical for training increasingly large language models (LLMs). A popular strategy involves gradient low-rank projection, storing only the projected optimizer states, with GaLore being a representative example. However, a significant drawback of many such methods is their lack of convergence guarantees, as various low-rank projection approaches introduce inherent biases relative to the original optimization algorithms, which contribute to performance gaps compared to full-parameter training. Aiming to tackle this problem, this paper investigates the layerwise sampling technique for debiasing low-rank projection mechanisms. In particular, an instantiation of the paradigm gives rise to a novel and unbiased low-rank optimization method built upon GaLore’s mechanism and the Muon algorithm, named GaLore Unbiased with Muon (GUM). We theoretically prove our method matches the convergence guarantees of the base Muon algorithm while preserving the memory efficiency of low-rank techniques. Empirical experiments on LLM fine-tuning and pretraining also demonstrate non-trivial improvements over GaLore and even better performance than full-parameter training. Further investigation shows that the improvement of this technique comes from a more uniform distribution of knowledge inside layers, leading to more efficient utilization of the model parameter space and better memorization.

[803] Stochastic Difference-of-Convex Optimization with Momentum

El Mahdi Chayti, Martin Jaggi

Main category: cs.LG

TL;DR: Momentum enables convergence in stochastic DC optimization under standard assumptions for any batch size, addressing limitations of existing methods.

DetailsMotivation: Stochastic DC optimization is widely used in machine learning but existing methods require large batches or strong noise assumptions, limiting practical applicability.

Method: Proposed a momentum-based algorithm for stochastic DC optimization that works with any batch size under standard smoothness and bounded variance assumptions.

Result: The momentum-based method achieves provable convergence and shows strong empirical performance, while proving that convergence fails without momentum regardless of stepsize.

Conclusion: Momentum is essential for convergence in stochastic DC optimization under practical conditions, enabling effective training with small batch sizes.

Abstract: Stochastic difference-of-convex (DC) optimization is prevalent in numerous machine learning applications, yet its convergence properties under small batch sizes remain poorly understood. Existing methods typically require large batches or strong noise assumptions, which limit their practical use. In this work, we show that momentum enables convergence under standard smoothness and bounded variance assumptions (of the concave part) for any batch size. We prove that without momentum, convergence may fail regardless of stepsize, highlighting its necessity. Our momentum-based algorithm achieves provable convergence and demonstrates strong empirical performance.

[804] Convergence Rates for Gradient Descent on the Edge of Stability in Overparametrised Least Squares

Lachlan Ewen MacDonald, Hancheng Min, Leandro Palma, Salma Tarmoun, Ziqing Xu, René Vidal

Main category: cs.LG

TL;DR: Analysis of gradient descent with large learning rates in overparametrized least squares, showing convergence rates in three regimes based on learning rate size: subcritical, critical, and supercritical.

DetailsMotivation: To understand and quantify the phenomenon of gradient descent operating in the 'edge of stability' regime where objectives decrease non-monotonically with implicit bias toward flat minima, which contrasts with classical optimization theory.

Method: Analyze gradient descent dynamics in overparametrized least squares by decomposing the dynamics into parallel and orthogonal components relative to the Riemannian manifold of global minimizers, treating the orthogonal component as a bifurcating dynamical system.

Result: Established convergence rates in three regimes: subcritical (transient instability with linear convergence to suboptimally flat minimum), critical (persistent instability with power-law convergence to optimally flat minimum), and supercritical (persistent instability with linear convergence to period-2 orbit around optimally flat minimum).

Conclusion: The analysis provides theoretical quantification of gradient descent behavior with large learning rates, explaining the observed implicit bias toward flat minima through the geometric structure of overparametrized optimization landscapes.

Abstract: Classical optimisation theory guarantees monotonic objective decrease for gradient descent (GD) when employed in a small step size, or stable", regime. In contrast, gradient descent on neural networks is frequently performed in a large step size regime called the edge of stability", in which the objective decreases non-monotonically with an observed implicit bias towards flat minima. In this paper, we take a step toward quantifying this phenomenon by providing convergence rates for gradient descent with large learning rates in an overparametrised least squares setting. The key insight behind our analysis is that, as a consequence of overparametrisation, the set of global minimisers forms a Riemannian manifold $M$, which enables the decomposition of the GD dynamics into components parallel and orthogonal to $M$. The parallel component corresponds to Riemannian gradient descent on the objective sharpness, while the orthogonal component is a bifurcating dynamical system. This insight allows us to derive convergence rates in three regimes characterised by the learning rate size: (a) the subcritical regime, in which transient instability is overcome in finite time before linear convergence to a suboptimally flat global minimum; (b) the critical regime, in which instability persists for all time with a power-law convergence toward the optimally flat global minimum; and (c) the supercritical regime, in which instability persists for all time with linear convergence to an orbit of period two centred on the optimally flat global minimum.

[805] SAFE-D: A Spatiotemporal Detection Framework for Abnormal Driving Among Parkinson’s Disease-like Drivers

Hangcheng Cao, Baixiang Huang, Longzhi Yuan, Haonan An, Zihan Fang, Xianhao Chen, Yuguang Fang

Main category: cs.LG

TL;DR: SAFE-D is a framework for detecting Parkinson’s disease-related driving anomalies using multi-component behavioral profiling and attention-based neural networks.

DetailsMotivation: Limited research exists on pathologically-triggered driving deviations from chronic medical conditions like Parkinson's disease, which pose safety risks to public transportation.

Method: Analyzed Parkinson’s symptomatology, built behavioral profiles from multiple vehicle control components, and designed an attention-based network for spatiotemporal feature prioritization.

Result: SAFE-D achieved 96.8% average accuracy in distinguishing normal and Parkinson-affected driving patterns on Logitech G29 platform and CARLA simulator.

Conclusion: The framework successfully detects Parkinson-related behavioral anomalies, enhancing driving safety for individuals with chronic medical conditions.

Abstract: A driver’s health state serves as a determinant factor in driving behavioral regulation. Subtle deviations from normalcy can lead to operational anomalies, posing risks to public transportation safety. While prior efforts have developed detection mechanisms for functionally-driven temporary anomalies such as drowsiness and distraction, limited research has addressed pathologically-triggered deviations, especially those stemming from chronic medical conditions. To bridge this gap, we investigate the driving behavior of Parkinson’s disease patients and propose SAFE-D, a novel framework for detecting Parkinson-related behavioral anomalies to enhance driving safety. Our methodology starts by performing analysis of Parkinson’s disease symptomatology, focusing on primary motor impairments, and establishes causal links to degraded driving performance. To represent the subclinical behavioral variations of early-stage Parkinson’s disease, our framework integrates data from multiple vehicle control components to build a behavioral profile. We then design an attention-based network that adaptively prioritizes spatiotemporal features, enabling robust anomaly detection under physiological variability. Finally, we validate SAFE-D on the Logitech G29 platform and CARLA simulator, using data from three road maps to emulate real-world driving. Our results show SAFE-D achieves 96.8% average accuracy in distinguishing normal and Parkinson-affected driving patterns.

[806] Curiosity Meets Cooperation: A Game-Theoretic Approach to Long-Tail Multi-Label Learning

Canran Xiao, Chuangxin Zhao, Zong Ke, Fei Shen

Main category: cs.LG

TL;DR: CD-GTMLL addresses long-tail imbalance in multi-label learning by framing it as a cooperative potential game where players share global accuracy but earn curiosity rewards for rare labels and disagreement, improving rare label performance without manual class weighting.

DetailsMotivation: Long-tail imbalance in multi-label learning causes head labels to dominate gradients while rare labels are ignored, despite their practical importance.

Method: Casts multi-label learning as a cooperative potential game with multiple players sharing global accuracy payoff and earning curiosity rewards based on label rarity and inter-player disagreement.

Result: Achieves state-of-the-art gains with up to +4.3% Rare-F1 and +1.6% P@3 over strongest baselines across conventional benchmarks and extreme-scale datasets.

Conclusion: CD-GTMLL provides a principled, scalable approach for long-tail robustness in multi-label prediction through game-theoretic framework with curiosity-driven rewards.

Abstract: Long-tail imbalance is endemic to multi-label learning: a few head labels dominate the gradient signal, while the many rare labels that matter in practice are silently ignored. We tackle this problem by casting the task as a cooperative potential game. In our Curiosity-Driven Game-Theoretic Multi-Label Learning (CD-GTMLL) framework, the label space is split among several cooperating players that share a global accuracy payoff yet earn additional curiosity rewards that rise with label rarity and inter-player disagreement. These curiosity bonuses inject gradient on under-represented tags without hand-tuned class weights. We prove that gradient best-response updates ascend a differentiable potential and converge to tail-aware stationary points that tighten a lower bound on the expected Rare-F1. Extensive experiments on conventional benchmarks and three extreme-scale datasets show consistent state-of-the-art gains, delivering up to +4.3% Rare-F1 and +1.6% P@3 over the strongest baselines, while ablations reveal emergent division of labour and faster consensus on rare classes. CD-GTMLL thus offers a principled, scalable route to long-tail robustness in multi-label prediction.

[807] Mitigating Clever Hans Strategies in Image Classifiers through Generating Counterexamples

Sidney Bender, Ole Delzer, Jan Herrmann, Heike Antje Marxfeld, Klaus-Robert Müller, Grégoire Montavon

Main category: cs.LG

TL;DR: CFKD addresses Clever Hans predictors in deep learning by generating counterfactuals to correct decision boundaries without needing group labels, overcoming limitations of methods like DFR.

DetailsMotivation: Deep learning models are vulnerable to spurious correlations (Clever Hans predictors), and existing group robustness methods like DFR have limitations: require group labels, low within-group sample sizes, and poor performance with multiple spurious correlations.

Method: Counterfactual Knowledge Distillation (CFKD) generates diverse counterfactuals and uses knowledge distillation to efficiently explore and correct model decision boundaries, enriching underrepresented groups with new data points without requiring confounder labels.

Result: CFKD achieves effective scaling to multiple confounders, yields balanced generalization across groups, and shows strong performance across five datasets, especially in low-data regimes with pronounced spurious correlations.

Conclusion: CFKD provides a robust framework that overcomes key limitations of existing methods by generating counterfactuals for decision boundary correction without requiring group labels, enabling effective handling of multiple spurious correlations.

Abstract: Deep learning models remain vulnerable to spurious correlations, leading to so-called Clever Hans predictors that undermine robustness even in large-scale foundation and self-supervised models. Group distributional robustness methods, such as Deep Feature Reweighting (DFR) rely on explicit group labels to upweight underrepresented subgroups, but face key limitations: (1) group labels are often unavailable, (2) low within-group sample sizes hinder coverage of the subgroup distribution, and (3) performance degrades sharply when multiple spurious correlations fragment the data into even smaller groups. We propose Counterfactual Knowledge Distillation (CFKD), a framework that sidesteps these issues by generating diverse counterfactuals, enabling a human annotator to efficiently explore and correct the model’s decision boundaries through a knowledge distillation step. Unlike DFR, our method not only reweights the undersampled groups, but it also enriches them with new data points. Our method does not require any confounder labels, achieves effective scaling to multiple confounders, and yields balanced generalization across groups. We demonstrate CFKD’s efficacy across five datasets, spanning synthetic tasks to an industrial application, with particularly strong gains in low-data regimes with pronounced spurious correlations. Additionally, we provide an ablation study on the effect of the chosen counterfactual explainer and teacher model, highlighting their impact on robustness.

[808] How Does Label Noise Gradient Descent Improve Generalization in the Low SNR Regime?

Wei Huang, Andi Han, Yujin Song, Yilan Chen, Denny Wu, Difan Zou, Taiji Suzuki

Main category: cs.LG

TL;DR: Adding label noise during gradient descent training helps suppress noise memorization and improves generalization in low signal-to-noise ratio settings, while standard gradient descent tends to overfit to noise.

DetailsMotivation: Deep learning models can overfit to noise in training data, especially in low signal-to-noise ratio (SNR) settings, which harms generalization. Prior observations suggest label noise may provide implicit regularization benefits.

Method: Training a two-layer neural network with label noise gradient descent algorithm in an idealized signal-noise data setting, where label noise is introduced during gradient updates.

Result: Label noise GD suppresses noise memorization, allows rapid signal growth while controlling overfitting, and achieves good generalization in low SNR. Standard GD overfits to noise and has non-vanishing test error lower bound.

Conclusion: Introducing label noise during gradient-based training provides benefits by preventing noise memorization and improving generalization performance in low SNR regimes.

Abstract: The capacity of deep learning models is often large enough to both learn the underlying statistical signal and overfit to noise in the training set. This noise memorization can be harmful especially for data with a low signal-to-noise ratio (SNR), leading to poor generalization. Inspired by prior observations that label noise provides implicit regularization that improves generalization, in this work, we investigate whether introducing label noise to the gradient updates can enhance the test performance of neural network (NN) in the low SNR regime. Specifically, we consider training a two-layer NN with a simple label noise gradient descent (GD) algorithm, in an idealized signal-noise data setting. We prove that adding label noise during training suppresses noise memorization, preventing it from dominating the learning process; consequently, label noise GD enjoys rapid signal growth while the overfitting remains controlled, thereby achieving good generalization despite the low SNR. In contrast, we also show that NN trained with standard GD tends to overfit to noise in the same low SNR setting and establish a non-vanishing lower bound on its test error, thus demonstrating the benefit of introducing label noise in gradient-based training.

[809] Reliable Inference in Edge-Cloud Model Cascades via Conformal Alignment

Jiayi Huang, Sangwoo Park, Nicola Paoletti, Osvaldo Simeone

Main category: cs.LG

TL;DR: A conformal alignment-based cascade method that ensures edge predictions maintain cloud-level conditional coverage while reducing cloud offloading.

DetailsMotivation: Edge intelligence enables low-latency inference but struggles with reliability assurance compared to cloud models. Current approaches lack statistical guarantees for conditional coverage when edge models make predictions.

Method: Proposes CAb (conformal alignment-based) cascading that treats edge-to-cloud escalation as multiple hypothesis testing. Uses conformal alignment to select which inputs can be safely handled at the edge while maintaining statistical guarantees.

Result: Experiments on CIFAR-100 and TeleQnA show the method maintains target conditional coverage for edge predictions while substantially reducing cloud offloading and only modestly increasing prediction set sizes.

Conclusion: The CAb cascade provides statistical guarantees for edge predictions to achieve cloud-level conditional coverage, enabling reliable edge intelligence with controlled trade-offs between coverage, deferral rate, and set size.

Abstract: Edge intelligence enables low-latency inference via compact on-device models, but assuring reliability remains challenging. We study edge-cloud cascades that must preserve conditional coverage: whenever the edge returns a prediction set, it should contain the true label with a user-specified probability, as if produced by the cloud model. We formalize conditional coverage with respect to the cloud predictive distribution, and introduce a conformal alignment-based (CAb) cascading mechanism that certifies this property with user control over the risk level. Our method casts escalation from edge to cloud models as a multiple-hypothesis testing (MHT) problem, tailoring conformal alignment (CA) to select which inputs can be safely handled at the edge. The proposed CAb model cascading method yields statistical guarantees on the average fraction of edge decisions that satisfy cloud-level conditional coverage. The procedure applies to arbitrary edge prediction sets, including variants of conformal prediction (CP), and exposes a tunable trade-off among coverage, deferral rate, and set size. Experiments on CIFAR-100 image classification and the TeleQnA question-answering (QA) benchmark show that the proposed CAb cascade maintains the target conditional coverage for edge predictions while substantially reducing offloading to the cloud and incurring modest increases in prediction-set size.

[810] TrajMamba: An Efficient and Semantic-rich Vehicle Trajectory Pre-training Model

Yichen Liu, Yan Lin, Shengnan Guo, Zeyu Zhou, Youfang Lin, Huaiyu Wan

Main category: cs.LG

TL;DR: TrajMamba is a novel approach for efficient and semantically rich vehicle trajectory learning that addresses computational challenges from textual data and redundant points through a Traj-Mamba Encoder, Travel Purpose-aware Pre-training, and Knowledge Distillation.

DetailsMotivation: To overcome challenges in learning travel semantics from GPS trajectories, including the computational burden from textual addresses/POIs and the negative impact of redundant points on efficiency and embedding quality.

Method: Uses Traj-Mamba Encoder to model GPS and road perspectives jointly, Travel Purpose-aware Pre-training to integrate travel purposes without extra overhead, and Knowledge Distillation Pre-training with learnable mask generator to compress trajectories.

Result: Outperforms state-of-the-art baselines in both efficiency and accuracy on two real-world datasets and three downstream tasks.

Conclusion: TrajMamba provides an effective solution for efficient and semantically rich trajectory learning by addressing key computational and redundancy challenges.

Abstract: Vehicle GPS trajectories record how vehicles move over time, storing valuable travel semantics, including movement patterns and travel purposes. Learning travel semantics effectively and efficiently is crucial for real-world applications of trajectory data, which is hindered by two major challenges. First, travel purposes are tied to the functions of the roads and points-of-interest (POIs) involved in a trip. Such information is encoded in textual addresses and descriptions and introduces heavy computational burden to modeling. Second, real-world trajectories often contain redundant points, which harm both computational efficiency and trajectory embedding quality. To address these challenges, we propose TrajMamba, a novel approach for efficient and semantically rich vehicle trajectory learning. TrajMamba introduces a Traj-Mamba Encoder that captures movement patterns by jointly modeling both GPS and road perspectives of trajectories, enabling robust representations of continuous travel behaviors. It also incorporates a Travel Purpose-aware Pre-training procedure to integrate travel purposes into the learned embeddings without introducing extra overhead to embedding calculation. To reduce redundancy in trajectories, TrajMamba features a Knowledge Distillation Pre-training scheme to identify key trajectory points through a learnable mask generator and obtain effective compressed trajectory embeddings. Extensive experiments on two real-world datasets and three downstream tasks show that TrajMamba outperforms state-of-the-art baselines in both efficiency and accuracy.

[811] ZACH-ViT: A Zero-Token Vision Transformer with ShuffleStrides Data Augmentation for Robust Lung Ultrasound Classification

Athanasios Angelakis, Amne Mousa, Micah L. A. Heldeweg, Laurens A. Biesheuvel, Mark A. Haaksma, Jasper M. Smit, Pieter R. Tuinman, Paul W. G. Elbers

Main category: cs.LG

TL;DR: ZACH-ViT is a compact Vision Transformer that achieves superior performance in classifying cardiogenic pulmonary oedema from non-cardiogenic patterns in lung ultrasound videos, outperforming larger models through permutation-invariant design and specialized data augmentation.

DetailsMotivation: Differentiating cardiogenic pulmonary oedema from non-cardiogenic patterns in lung ultrasound videos is challenging due to high visual variability and overlapping artifacts, which complicates automated classification.

Method: Developed ZACH-ViT, a 0.25M-parameter Vision Transformer variant that removes positional embeddings and CLS token for full permutation-invariance, combined with ShuffleStrides Data Augmentation that permutes probe-view sequences while preserving anatomical validity.

Result: ZACH-ViT achieved highest validation and test ROC-AUC (0.80 and 0.79) with balanced sensitivity (0.60) and specificity (0.91), outperforming 9 state-of-the-art baselines that collapsed to trivial classification. It trains 1.35x faster than Minimal ViT with 2.5x fewer parameters.

Conclusion: Aligning architectural design with data structure can outperform scale in small-data medical imaging, supporting real-time clinical deployment of compact models for lung ultrasound classification.

Abstract: Differentiating cardiogenic pulmonary oedema (CPE) from non-cardiogenic and structurally normal lungs in lung ultrasound (LUS) videos remains challenging due to the high visual variability of non-cardiogenic inflammatory patterns (NCIP/ARDS-like), interstitial lung disease, and healthy lungs. This heterogeneity complicates automated classification as overlapping B-lines and pleural artefacts are common. We introduce ZACH-ViT (Zero-token Adaptive Compact Hierarchical Vision Transformer), a 0.25 M-parameter Vision Transformer variant that removes both positional embeddings and the [CLS] token, making it fully permutation-invariant and suitable for unordered medical image data. To enhance generalization, we propose ShuffleStrides Data Augmentation (SSDA), which permutes probe-view sequences and frame orders while preserving anatomical validity. ZACH-ViT was evaluated on 380 LUS videos from 95 critically ill patients against nine state-of-the-art baselines. Despite the heterogeneity of the non-cardiogenic group, ZACH-ViT achieved the highest validation and test ROC-AUC (0.80 and 0.79) with balanced sensitivity (0.60) and specificity (0.91), while all competing models collapsed to trivial classification. It trains 1.35x faster than Minimal ViT (0.62M parameters) with 2.5x fewer parameters, supporting real-time clinical deployment. These results show that aligning architectural design with data structure can outperform scale in small-data medical imaging.

[812] The Free Transformer

François Fleuret

Main category: cs.LG

TL;DR: Extension of decoder Transformer with unsupervised variational latent variables for improved downstream task performance.

DetailsMotivation: To enhance the generative process of decoder Transformers by incorporating random latent variables that can be learned without supervision.

Method: Proposed a variational procedure to learn random latent variables that condition the generative process of decoder Transformers.

Result: Experimental evaluations demonstrate substantial improvements on downstream tasks when using this conditioning approach.

Conclusion: Conditioning decoder Transformers on unsupervised variational latent variables leads to significant performance gains in downstream applications.

Abstract: We propose an extension of the decoder Transformer that conditions its generative process on random latent variables which are learned without supervision thanks to a variational procedure. Experimental evaluations show that allowing such a conditioning translates into substantial improvements on downstream tasks.

[813] Formally Exploring Time-Series Anomaly Detection Evaluation Metrics

Dennis Wagner, Arjun Nair, Billy Joe Franks, Justus Arweiler, Aparna Muraleedharan, Indra Jungjohann, Fabian Hartung, Mayank C. Ahuja, Andriy Balinskyy, Saurabh Varshneya, Nabeel Hussain Syed, Mayank Nagda, Phillip Liznerski, Steffen Reithermann, Maja Rudolph, Sebastian Vollmer, Ralf Schulz, Torsten Katz, Stephan Mandt, Michael Bortz, Heike Leitte, Daniel Neider, Jakob Burger, Fabian Jirasek, Hans Hasse, Sophie Fellenz, Marius Kloft

Main category: cs.LG

TL;DR: The paper introduces verifiable properties for evaluating time-series anomaly detection metrics, analyzes 37 existing metrics showing none satisfy all properties, and proposes LARM and ALARM metrics that provably meet all requirements.

DetailsMotivation: Current anomaly detection metrics are inadequate and misleading, which can lead to catastrophic failures in safety-critical systems due to undetected anomalies.

Method: Developed a theoretical framework with verifiable properties for evaluating time-series anomaly detection, analyzed 37 existing metrics, and designed new metrics (LARM and ALARM) that provably satisfy all essential properties.

Result: Most existing metrics satisfy only a few properties, none satisfy all, explaining inconsistencies in prior results. The proposed LARM and ALARM metrics provably satisfy all essential evaluation properties.

Conclusion: The paper provides a principled framework for evaluating time-series anomaly detection and introduces LARM/ALARM metrics that overcome limitations of existing metrics, enabling reliable comparisons and better anomaly detection in safety-critical systems.

Abstract: Undetected anomalies in time series can trigger catastrophic failures in safety-critical systems, such as chemical plant explosions or power grid outages. Although many detection methods have been proposed, their performance remains unclear because current metrics capture only narrow aspects of the task and often yield misleading results. We address this issue by introducing verifiable properties that formalize essential requirements for evaluating time-series anomaly detection. These properties enable a theoretical framework that supports principled evaluations and reliable comparisons. Analyzing 37 widely used metrics, we show that most satisfy only a few properties, and none satisfy all, explaining persistent inconsistencies in prior results. To close this gap, we propose LARM, a flexible metric that provably satisfies all properties, and extend it to ALARM, an advanced variant meeting stricter requirements.

[814] Semi-supervised Latent Bayesian Optimization for Designing Antimicrobial Peptides

Jyler Menard, R. A. Mansbach

Main category: cs.LG

TL;DR: This paper investigates using dimensionality reduction on variational autoencoder latent spaces to improve antimicrobial peptide design by enhancing interpretability and optimization efficiency.

DetailsMotivation: Deep generative models for peptide design lack interpretability and rigorous quantification of latent space quality as a search space, making optimization difficult.

Method: The study uses dimensionality reduction on variational autoencoder latent spaces, organizing them with physicochemical properties to improve antimicrobial activity optimization.

Result: Further reduction of latent space via dimensionality reduction is advantageous when organizing with relevant information, improves interpretability, and allows organization with different physicochemical properties even with limited labels.

Conclusion: Dimensionality reduction can enhance latent space interpretability and optimization efficiency in antimicrobial peptide design when combined with physicochemical property organization.

Abstract: Antimicrobial peptides (AMPs) are a promising class of therapeutics to treat bacterial infections. Discovering and designing such peptides is difficult because of the vast number of possible sequences of amino acids. Deep generative models, such as variational autoencoders, have shown value in peptide design due to their ability to model sequence space with a continuous-valued latent space. Although such models have already been used to great effect in biomolecular design, they still suffer from a lack of interpretability and rigorous quantification of latent space quality as a search space. We investigate (1) whether further compression of the design space via dimensionality reduction may facilitate optimization, (2) the interpretability of the spaces, and (3) how organizing latent spaces with physicochemical properties may improve the efficiency of optimizing antimicrobial activity. We find that further reduction of the latent space via dimensionality reduction can be advantageous when organizing the space with more relevant information at data availability, that using the dimensionality reduction search space can be more interpretable, and that we can organize the latent space with different physicochemical properties even at different percentages of available labels.

[815] Handling Extreme Class Imbalance: Using GANs in Data Augmentation for Suicide Prediction

Vaishnavi Visweswaraiah, Tanvi Banerjee, William Romine

Main category: cs.LG

TL;DR: Used machine learning and GAN-based data augmentation to address extreme class imbalance in suicide prediction, achieving high performance metrics with Logistic Regression and Random Forest models.

DetailsMotivation: Suicide prediction is crucial for prevention but faces extreme class imbalance issues due to rare positive cases in real data, making effective modeling challenging.

Method: Applied machine learning models (Logistic Regression, Random Forest, SVM) and used Generative Adversarial Networks (GAN) to generate synthetic data for augmentation from an initial dataset of 656 samples with only 4 positive cases.

Result: Logistic Regression achieved weighted precision 0.99, recall 0.85, F1 0.91; Random Forest: 0.98, 0.99, 0.99; SVM: 0.99, 0.76, 0.86. LR and SVM correctly identified suicide attempts (sensitivity:1.0) while RF had 0 sensitivity but perfect specificity.

Conclusion: Machine learning models demonstrated effectiveness in suicide prediction, with GAN playing a crucial role in addressing class imbalance through synthetic data generation to support prevention efforts.

Abstract: Suicide prediction is the key for prevention, but real data with sufficient positive samples is rare and causes extreme class imbalance. We utilized machine learning (ML) to build the model and deep learning (DL) techniques, like Generative Adversarial Networks (GAN), to generate synthetic data samples to enhance the dataset. The initial dataset contained 656 samples, with only four positive cases, prompting the need for data augmentation. A variety of machine learning models, ranging from interpretable data models to black box algorithmic models, were used. On real test data, Logistic Regression (LR) achieved a weighted precision of 0.99, a weighted recall of 0.85, and a weighted F1 score of 0.91; Random Forest (RF) showed 0.98, 0.99, and 0.99, respectively; and Support Vector Machine (SVM) achieved 0.99, 0.76, and 0.86. LR and SVM correctly identified one suicide attempt case (sensitivity:1.0) and misclassified LR(20) and SVM (31) non-attempts as attempts (specificity: 0.85 & 0.76, respectively). RF identified 0 suicide attempt cases (sensitivity: 0.0) with 0 false positives (specificity: 1.0). These results highlight the models’ effectiveness, with GAN playing a key role in generating synthetic data to support suicide prevention modeling efforts.

[816] Efficient Algorithms for Mitigating Uncertainty and Risk in Reinforcement Learning

Xihong Su

Main category: cs.LG

TL;DR: This dissertation presents three main contributions: CADP algorithm connecting policy gradient and dynamic programming, theoretical analysis of ERM Bellman operators with new algorithms, and model-free Q-learning methods for risk-averse objectives.

DetailsMotivation: To bridge the gap between policy gradient methods and dynamic programming in multi-model MDPs, and to develop effective algorithms for risk-averse reinforcement learning objectives like ERM-TRC and EVaR-TRC.

Method: Developed Coordinate Ascent Dynamic Programming (CADP), exponential value iteration/policy iteration/linear programming algorithms, and model-free Q-learning algorithms with rigorous convergence proofs for risk-averse objectives.

Result: Established conditions for ERM Bellman operator contraction, proved existence of optimal policies for ERM-TRC and EVaR-TRC, and demonstrated convergence of proposed Q-learning algorithms to optimal risk-averse value functions.

Conclusion: The dissertation successfully connects policy gradient with dynamic programming, provides theoretical foundations for risk-averse RL, and delivers practical algorithms for computing optimal policies in uncertain and risk-sensitive environments.

Abstract: This dissertation makes three main contributions. First, We identify a new connection between policy gradient and dynamic programming in MMDPs and propose the Coordinate Ascent Dynamic Programming (CADP) algorithm to compute a Markov policy that maximizes the discounted return averaged over the uncertain models. CADP adjusts model weights iteratively to guarantee monotone policy improvements to a local maximum. Second, We establish sufficient and necessary conditions for the exponential ERM Bellman operator to be a contraction and prove the existence of stationary deterministic optimal policies for ERM-TRC and EVaR-TRC. We also propose exponential value iteration, policy iteration, and linear programming algorithms for computing optimal stationary policies for ERM-TRC and EVaR-TRC. Third, We propose model-free Q-learning algorithms for computing policies with risk-averse objectives: ERM-TRC and EVaR-TRC. The challenge is that Q-learning ERM Bellman may not be a contraction. Instead, we use the monotonicity of Q-learning ERM Bellman operators to derive a rigorous proof that the ERM-TRC and the EVaR-TRC Q-learning algorithms converge to the optimal risk-averse value functions. The proposed Q-learning algorithms compute the optimal stationary policy for ERM-TRC and EVaR-TRC.

[817] Enabling Fine-Grained Operating Points for Black-Box LLMs

Ege Beyazit, KL Navaneet, Prashant Mathur, Roi Blanco, Vidit Bansal, Karim Bouyarmane

Main category: cs.LG

TL;DR: Black-box LLMs have low numerical output cardinality, limiting operational granularity for applications requiring specific metric constraints. The paper proposes efficient methods to increase operating points without performance loss.

DetailsMotivation: Black-box LLMs are practical for decision-making but have limited control over operating points due to low numerical output cardinality, preventing fine-grained adjustment for applications with specific metric constraints.

Method: Investigates reasons for low-cardinality outputs, experiments with standard techniques (prompt engineering, uncertainty estimation, confidence elicitation), and proposes efficient approaches to increase operating point diversity.

Result: Proposed approaches provide finer-grained operating points and achieve comparable or better performance than benchmark methods across 11 datasets and 3 LLMs.

Conclusion: Efficient methods can significantly improve operational granularity of black-box LLMs without sacrificing performance or increasing inference cost.

Abstract: Black-box Large Language Models (LLMs) provide practical and accessible alternatives to other machine learning methods, as they require minimal labeled data and machine learning expertise to develop solutions for various decision making problems. However, for applications that need operating with constraints on specific metrics (e.g., precision $\geq$ 95%), decision making with black-box LLMs remains unfavorable, due to their low numerical output cardinalities. This results in limited control over their operating points, preventing fine-grained adjustment of their decision making behavior. In this paper, we study using black-box LLMs as classifiers, focusing on efficiently improving their operational granularity without performance loss. Specifically, we first investigate the reasons behind their low-cardinality numerical outputs and show that they are biased towards generating rounded but informative verbalized probabilities. Then, we experiment with standard prompt engineering, uncertainty estimation and confidence elicitation techniques, and observe that they do not effectively improve operational granularity without sacrificing performance or increasing inference cost. Finally, we propose efficient approaches to significantly increase the number and diversity of available operating points. Our proposed approaches provide finer-grained operating points and achieve comparable to or better performance than the benchmark methods across 11 datasets and 3 LLMs.

[818] Atlas-based Manifold Representations for Interpretable Riemannian Machine Learning

Ryan A. Robinett, Sophia A. Madejski, Kyle Ruark, Samantha J. Riesenfeld, Lorenzo Orecchia

Main category: cs.LG

TL;DR: The paper proposes a differentiable atlas-based method for machine learning directly on data manifolds, addressing limitations of traditional manifold learning that lose manifold features during dimensionality reduction.

DetailsMotivation: Current manifold-learning methods primarily perform dimensionality reduction into Euclidean space, losing key manifold features when embedding dimension approaches the intrinsic dimension. Methods that directly learn latent manifolds as differentiable atlases have been underexplored.

Method: Implemented a generic data structure to maintain a differentiable atlas enabling Riemannian optimization over the manifold, complemented by an unsupervised heuristic that learns a differentiable atlas from point cloud data.

Result: The approach demonstrated advantages in efficiency and accuracy in selected settings. In supervised classification over the Klein bottle and RNA velocity analysis of hematopoietic data, it showed improved interpretability and robustness.

Conclusion: Atlas-based methods provide an effective approach for machine learning directly on data manifolds, offering better preservation of manifold structure and improved performance in various applications.

Abstract: Despite the popularity of the manifold hypothesis, current manifold-learning methods do not support machine learning directly on the latent $d$-dimensional data manifold, as they primarily aim to perform dimensionality reduction into $\mathbb{R}^D$, losing key manifold features when the embedding dimension $D$ approaches $d$. On the other hand, methods that directly learn the latent manifold as a differentiable atlas have been relatively underexplored. In this paper, we aim to give a proof of concept of the effectiveness and potential of atlas-based methods. To this end, we implement a generic data structure to maintain a differentiable atlas that enables Riemannian optimization over the manifold. We complement this with an unsupervised heuristic that learns a differentiable atlas from point cloud data. We experimentally demonstrate that this approach has advantages in terms of efficiency and accuracy in selected settings. Moreover, in a supervised classification task over the Klein bottle and in RNA velocity analysis of hematopoietic data, we showcase the improved interpretability and robustness of our approach.

[819] Inference-Time Compute Scaling For Flow Matching

Adam Stecklov, Noah El Rimawi-Fine, Mathieu Blanchette

Main category: cs.LG

TL;DR: The paper introduces novel inference-time scaling procedures for Flow Matching that preserve linear interpolants during sampling, showing consistent quality improvements in image and protein generation tasks.

DetailsMotivation: While inference-time computation scaling has improved sample quality in other models, Flow Matching lacks effective scaling methods that maintain its efficient sampling properties, and existing approaches sacrifice these benefits.

Method: Developed new inference-time scaling procedures for Flow Matching that preserve the linear interpolant during sampling, unlike previous methods that used non-linear interpolants.

Result: Evaluations on image generation and unconditional protein generation show consistent quality improvements with increased inference compute, demonstrating applicability to scientific domains.

Conclusion: The proposed inference-time scaling methods for Flow Matching successfully preserve efficient sampling while improving quality across domains, including scientific applications like protein generation.

Abstract: Allocating extra computation at inference time has recently improved sample quality in large language models and diffusion-based image generation. In parallel, Flow Matching (FM) has gained traction in language, vision, and scientific domains, but inference-time scaling methods for it remain under-explored. Concurrently, Kim et al., 2025 approach this problem but replace the linear interpolant with a non-linear variance-preserving (VP) interpolant at inference, sacrificing FM’s efficient and straight sampling. Additionally, inference-time compute scaling for flow matching has only been applied to visual tasks, like image generation. We introduce novel inference-time scaling procedures for FM that preserve the linear interpolant during sampling. Evaluations of our method on image generation, and for the first time (to the best of our knowledge), unconditional protein generation, show that I) sample quality consistently improves as inference compute increases, and II) flow matching inference-time scaling can be applied to scientific domains.

[820] Functional Distribution Networks (FDN)

Omer Haq

Main category: cs.LG

TL;DR: Functional Distribution Networks (FDN) is a method that creates input-conditioned weight distributions to produce adaptive predictive mixtures, trained with beta-ELBO and Monte Carlo sampling, improving calibration under distribution shift.

DetailsMotivation: Modern probabilistic regressors often remain overconfident under distribution shift, necessitating methods that can adapt predictive uncertainty to input conditions.

Method: FDN uses input-conditioned distributions over network weights to induce predictive mixtures with adaptive dispersion, trained via beta-ELBO and Monte Carlo sampling.

Result: FDN benchmarks against Bayesian, ensemble, dropout, and hypernetwork baselines under matched budgets, showing improved accuracy, calibration, and shift-awareness in standard regression tasks.

Conclusion: The framework and evaluation protocol aim to make OOD-aware, well-calibrated neural regression practical and modular.

Abstract: Modern probabilistic regressors often remain overconfident under distribution shift. We present Functional Distribution Networks (FDN), an input-conditioned distribution over network weights that induces predictive mixtures whose dispersion adapts to the input. FDN is trained with a beta-ELBO and Monte Carlo sampling. We further propose an evaluation protocol that cleanly separates interpolation from extrapolation and stresses OOD sanity checks (e.g., that predictive likelihood degrades under shift while in-distribution accuracy and calibration are maintained). On standard regression tasks, we benchmark against strong Bayesian, ensemble, dropout, and hypernetwork baselines under matched parameter and update budgets, and assess accuracy, calibration, and shift-awareness with standard diagnostics. Together, the framework and protocol aim to make OOD-aware, well-calibrated neural regression practical and modular.

[821] Parameter Efficient Fine-tuning via Explained Variance Adaptation

Fabian Paischer, Lukas Hauzenberger, Thomas Schmied, Benedikt Alkin, Marc Peter Deisenroth, Sepp Hochreiter

Main category: cs.LG

TL;DR: EVA is a new LoRA initialization method that uses directions capturing the most activation variance to maximize gradient signal and accelerate fine-tuning of foundation models.

DetailsMotivation: Existing LoRA initialization strategies don't provably maximize expected gradient signal, which is critical for fast adaptation during fine-tuning.

Method: EVA performs incremental SVD on minibatches of activation vectors and selects converged right-singular vectors for initialization. It selects directions capturing most activation variance for given rank budget, enabling adaptive ranks.

Result: EVA shows faster convergence than competitors and achieves highest average scores across language generation/understanding, image classification, and reinforcement learning tasks while reducing trainable parameters through rank redistribution.

Conclusion: EVA establishes a new Pareto frontier compared to existing LoRA initialization schemes in both accuracy and efficiency.

Abstract: Foundation models (FMs) are pre-trained on large-scale datasets and then fine-tuned for a specific downstream task. The most common fine-tuning method is to update pretrained weights via low-rank adaptation (LoRA). Existing initialization strategies for LoRA often rely on singular value decompositions (SVD) of gradients or weight matrices. However, they do not provably maximize the expected gradient signal, which is critical for fast adaptation. To this end, we introduce Explained Variance Adaptation (EVA), an initialization scheme that uses the directions capturing the most activation variance, provably maximizing the expected gradient signal and accelerating fine-tuning. EVA performs incremental SVD on minibatches of activation vectors and selects the right-singular vectors for initialization once they converged. Further, by selecting the directions that capture the most activation-variance for a given rank budget, EVA accommodates adaptive ranks that reduce the number of trainable parameters. We apply EVA to a variety of fine-tuning tasks as language generation and understanding, image classification, and reinforcement learning. EVA exhibits faster convergence than competitors and achieves the highest average score across a multitude of tasks per domain while reducing the number of trainable parameters through rank redistribution. In summary, EVA establishes a new Pareto frontier compared to existing LoRA initialization schemes in both accuracy and efficiency.

[822] Graph Neural Networks for the Offline Nanosatellite Task Scheduling Problem

Bruno Machado Pacheco, Laio Oriel Seman, Cezar Antonio Rigo, Eduardo Camponogara, Eduardo Augusto Bezerra, Leandro dos Santos Coelho

Main category: cs.LG

TL;DR: This paper proposes using Graph Neural Networks (GNNs) to improve nanosatellite task scheduling by learning problem structure and providing better heuristic solutions compared to traditional solvers.

DetailsMotivation: Traditional mathematical methods for the Offline Nanosatellite Task Scheduling (ONTS) problem have limited applicability to challenging cases, while GNNs have shown success in other optimization problems.

Method: The study investigates whether GNNs can learn the complex structure of ONTS problems regarding feasibility and optimality, and evaluates GNN-based heuristic solutions to improve optimization performance.

Result: GNNs successfully learned feasibility and optimality for ONTS instances and generalized to harder cases. GNN-based heuristics improved expected objective value by 45% and reduced time to find feasible solutions by 35% compared to SCIP solver.

Conclusion: GNNs are effective for nanosatellite task scheduling, demonstrating strong learning capabilities and significant performance improvements over traditional optimization methods.

Abstract: This study investigates how to schedule nanosatellite tasks more efficiently using Graph Neural Networks (GNNs). In the Offline Nanosatellite Task Scheduling (ONTS) problem, the goal is to find the optimal schedule for tasks to be carried out in orbit while taking into account Quality-of-Service (QoS) considerations such as priority, minimum and maximum activation events, execution time-frames, periods, and execution windows, as well as constraints on the satellite’s power resources and the complexity of energy harvesting and management. The ONTS problem has been approached using conventional mathematical formulations and exact methods, but their applicability to challenging cases of the problem is limited. This study examines the use of GNNs in this context, which has been effectively applied to optimization problems such as the traveling salesman, scheduling, and facility placement problems. More specifically, we investigate whether GNNs can learn the complex structure of the ONTS problem with respect to feasibility and optimality of candidate solutions. Furthermore, we evaluate using GNN-based heuristic solutions to provide better solutions (w.r.t. the objective value) to the ONTS problem and reduce the optimization cost. Our experiments show that GNNs are not only able to learn feasibility and optimality for instances of the ONTS problem, but they can generalize to harder instances than those seen during training. Furthermore, the GNN-based heuristics improved the expected objective value of the best solution found under the time limit in 45%, and reduced the expected time to find a feasible solution in 35%, when compared to the SCIP (Solving Constraint Integer Programs) solver in its off-the-shelf configuration

[823] Membership Privacy Risks of Sharpness Aware Minimization

Young In Kim, Andrea Agiollo, Pratiksha Agrawal, Johannes O. Royset, Rajiv Khanna

Main category: cs.LG

TL;DR: SAM optimization improves generalization but increases membership inference attack vulnerability compared to SGD, challenging conventional privacy-generalization trade-off beliefs.

DetailsMotivation: To investigate whether optimization algorithms that find flatter minima (like SAM) impact membership privacy, given their known benefits for generalization.

Method: Empirical evaluation across multiple datasets and attack methods, analysis of memorization and influence scores, and theoretical analysis of minority subclass feature learning.

Result: SAM is more vulnerable to membership inference attacks than SGD despite achieving lower test error, suggesting it memorizes atypical subpatterns more effectively.

Conclusion: Models that better capture minority subclass features can simultaneously improve generalization and increase membership privacy risk, revealing a counter-intuitive relationship between privacy and generalization.

Abstract: Optimization algorithms that seek flatter minima such as Sharpness-Aware Minimization (SAM) are widely credited with improved generalization. We ask whether such gains impact membership privacy. Surprisingly, we find that SAM is more prone to membership inference attacks than classical SGD across multiple datasets and attack methods, despite achieving lower test error. This is an intriguing phenomenon as conventional belief posits that higher membership privacy risk is associated with poor generalization. We conjecture that SAM is capable of memorizing atypical subpatterns more, leading to better generalization but higher privacy risk. We empirically validate our hypothesis by running extensive analysis on memorization and influence scores. Finally, we theoretically show how a model that captures minority subclass features more can effectively generalize better \emph{and} have higher membership privacy risk.

[824] Harmony in Divergence: Towards Fast, Accurate, and Memory-efficient Zeroth-order LLM Fine-tuning

Qitao Tan, Jun Liu, Zheng Zhan, Caiwei Ding, Yanzhi Wang, Xiaolong Ma, Jaewoo Lee, Jin Lu, Geng Yuan

Main category: cs.LG

TL;DR: DiZO is a novel zeroth-order optimization method that uses layer-wise divergence analysis to improve convergence speed and accuracy, reducing training GPU hours by up to 48% while sometimes outperforming memory-intensive first-order fine-tuning.

DetailsMotivation: Standard first-order fine-tuning requires significant memory, limiting real-world deployment. Zeroth-order optimization is memory-efficient but lags behind in convergence speed and accuracy compared to first-order methods.

Method: DiZO conducts divergence-driven layer adaptation by incorporating projections to ZO updates, generating diverse-magnitude updates precisely scaled to layer-wise individual optimization needs based on layer-wise divergence analysis.

Result: DiZO significantly reduces needed iterations for convergence without sacrificing throughput, cutting training GPU hours by up to 48% on various datasets. It consistently outperforms ZO baselines and sometimes surpasses memory-intensive FO fine-tuning on RoBERTa-large, OPT-series, and Llama-series.

Conclusion: DiZO bridges the performance gap between zeroth-order and first-order optimization, offering a memory-efficient training paradigm that maintains high accuracy while significantly reducing computational requirements.

Abstract: Large language models (LLMs) excel across various tasks, but standard first-order (FO) fine-tuning demands considerable memory, significantly limiting real-world deployment. Recently, zeroth-order (ZO) optimization stood out as a promising memory-efficient training paradigm, avoiding backward passes and relying solely on forward passes for gradient estimation, making it attractive for resource-constrained scenarios. However, ZO method lags far behind FO method in both convergence speed and accuracy. To bridge the gap, we introduce a novel layer-wise divergence analysis that uncovers the distinct update pattern of FO and ZO optimization. Aiming to resemble the learning capacity of FO method from the findings, we propose Divergence-driven Zeroth-Order (DiZO) optimization. DiZO conducts divergence-driven layer adaptation by incorporating projections to ZO updates, generating diverse-magnitude updates precisely scaled to layer-wise individual optimization needs. Our results demonstrate that DiZO significantly reduces the needed iterations for convergence without sacrificing throughput, cutting training GPU hours by up to 48% on various datasets. Moreover, DiZO consistently outperforms the representative ZO baselines in fine-tuning RoBERTa-large, OPT-series, and Llama-series on downstream tasks and, in some cases, even surpasses memory-intensive FO fine-tuning. Our code is released at https://github.com/Skilteee/DiZO.

[825] Diffusion Models as Constrained Samplers for Optimization with Unknown Constraints

Lingkai Kong, Yuanqi Du, Wenhao Mu, Kirill Neklyudov, Valentin De Bortoli, Dongxia Wu, Haorui Wang, Aaron Ferber, Yi-An Ma, Carla P. Gomes, Chao Zhang

Main category: cs.LG

TL;DR: The paper proposes using diffusion models to handle optimization problems with unknown constraints by reformulating optimization as sampling from the product of Boltzmann distribution and learned data distribution, with different methods for differentiable vs non-differentiable objectives.

DetailsMotivation: Real-world optimization often lacks analytic constraints, leading to unrealistic solutions. While unknown objectives have been studied, unknown constraints remain under-explored despite their practical importance.

Method: Reformulate optimization as sampling from Boltzmann distribution × data distribution. For differentiable objectives: two-stage framework with guided diffusion + Langevin dynamics. For non-differentiable: iterative importance sampling using diffusion model as proposal distribution.

Result: Comprehensive experiments on synthetic, real-world black-box optimization, and multi-objective molecule optimization datasets show better or comparable performance with state-of-the-art baselines.

Conclusion: The proposed diffusion-based approach effectively handles unknown constraints in optimization problems and achieves strong performance across diverse domains.

Abstract: Addressing real-world optimization problems becomes particularly challenging when analytic objective functions or constraints are unavailable. While numerous studies have addressed the issue of unknown objectives, limited research has focused on scenarios where feasibility constraints are not given explicitly. Overlooking these constraints can lead to spurious solutions that are unrealistic in practice. To deal with such unknown constraints, we propose to perform optimization within the data manifold using diffusion models. To constrain the optimization process to the data manifold, we reformulate the original optimization problem as a sampling problem from the product of the Boltzmann distribution defined by the objective function and the data distribution learned by the diffusion model. Depending on the differentiability of the objective function, we propose two different sampling methods. For differentiable objectives, we propose a two-stage framework that begins with a guided diffusion process for warm-up, followed by a Langevin dynamics stage for further correction. For non-differentiable objectives, we propose an iterative importance sampling strategy using the diffusion model as the proposal distribution. Comprehensive experiments on a synthetic dataset, six real-world black-box optimization datasets, and a multi-objective molecule optimization dataset show that our method achieves better or comparable performance with previous state-of-the-art baselines.

[826] LinkedIn Post Embeddings: Industrial Scale Embedding Generation and Usage across LinkedIn

Sudarshan Srinivasa Ramanujam, Akanksha Bindal, Yu Jiang, Timothy J. Hazen, David Golland, Fengyu Zhang, Daqi Sun, Wanning Li, Birjodh Singh Tiwana, Siddharth Dangi, Peng Yan

Main category: cs.LG

TL;DR: LinkedIn developed post embeddings using fine-tuned LLMs with multi-task learning, achieving superior performance over baselines and OpenAI embeddings, with successful production deployment powering multiple products for over two years.

DetailsMotivation: To create effective post embeddings for LinkedIn's retrieval and ranking systems (feed, video tab) that capture semantic meaning and outperform existing solutions.

Method: Fine-tuned pre-trained transformer-based LLM using multi-task learning across diverse semantic labeling tasks, deployed to near-line infrastructure for real-time availability.

Result: Positive transfer across all tasks, outperforming baseline models in zero-shot learning and OpenAI’s ADA embeddings on LinkedIn-specific datasets, with proven real-world impact in production.

Conclusion: Multi-task fine-tuned LLM embeddings provide superior performance for LinkedIn’s ranking and retrieval systems, demonstrating broad applicability and sustained production success.

Abstract: A post embedding (representation of text in embedding space that effectively captures semantic meaning) is a foundational component of LinkedIn that is consumed by product surfaces in retrieval and ranking (e.g., ranking posts in the feed or video tab). This paper presents the post embeddings used at LinkedIn, where a pre-trained transformer-based large language model (LLM) is taken as input and fine-tuned using multi-task learning across a diverse set of semantic labeling tasks. We observe positive transfer, leading to improved performance across all tasks, compared to training them independently. The generated post embeddings outperform baseline models in zero-shot learning, demonstrating its potential for broader applicability. Furthermore, the generated post embeddings’ performance surpasses that of OpenAI’s ADA-001 and ADA-002 embeddings on LinkedIn specific datasets and tasks. We also describe the offline evaluation methodology and the deployment to our near-line infrastructure, which makes the post embedding available for use within minutes of post creation for any downstream application. We present how the embeddings were applied in the Feed product surface, in both ranking and retrieval stages, and showcase the real world online impact to demonstrate the superior performance of these embeddings. Finally, we also share the results of applying the embeddings to the retrieval system of our video ranking product surface in LinkedIn. These embeddings have been battle-tested in production at LinkedIn for over two years, consistently powering multiple products.

[827] Large Language Models are Powerful Electronic Health Record Encoders

Stefan Hegselmann, Georg von Arnim, Tillmann Rheude, Noel Kronenberg, David Sontag, Gerhard Hindricks, Roland Eils, Benjamin Wild

Main category: cs.LG

TL;DR: Using general-purpose LLMs to encode EHR data into text-based representations achieves comparable or better performance than specialized EHR foundation models, while offering better generalization across different healthcare systems and coding standards.

DetailsMotivation: EHR foundation models face limitations due to restricted access to diverse datasets and inconsistencies in coding standards. General-purpose LLMs offer an alternative with better semantic understanding and generalization capabilities.

Method: Convert structured EHR data into Markdown-formatted plain-text documents by replacing medical codes with natural language descriptions, then use LLMs to encode these documents for clinical prediction tasks.

Result: LLM-based embeddings match or surpass specialized EHR foundation model (CLMBR-T-Base) across 15 clinical tasks, and show superior performance on out-of-domain UK Biobank data for disease onset, hospitalization, and mortality prediction.

Conclusion: LLM-based EHR encoding provides a flexible, generalizable alternative to specialized foundation models, requiring no institution-specific training and handling diverse medical codes through text descriptions.

Abstract: Electronic Health Records (EHRs) offer considerable potential for clinical prediction, but their complexity and heterogeneity present significant challenges for traditional machine learning methods. Recently, domain-specific EHR foundation models trained on large volumes of unlabeled EHR data have shown improved predictive accuracy and generalization. However, their development is constrained by limited access to diverse, high-quality datasets, and inconsistencies in coding standards and clinical practices. In this study, we explore the use of general-purpose Large Language Models (LLMs) to encode EHR into high-dimensional representations for downstream clinical prediction tasks. We convert structured EHR data into Markdown-formatted plain-text documents by replacing medical codes with natural language descriptions. This enables the use of LLMs and their extensive semantic understanding and generalization capabilities as effective encoders of EHRs without requiring access to private medical training data. We show that LLM-based embeddings can often match or even surpass the performance of a specialized EHR foundation model, CLMBR-T-Base, across 15 diverse clinical tasks from the EHRSHOT benchmark. Critically, our approach requires no institution-specific training and can incorporate any medical code with a text description, whereas existing EHR foundation models operate on fixed vocabularies and can only process codes seen during pretraining. To demonstrate generalizability, we further evaluate the approach on the UK Biobank (UKB) cohort, out-of-domain for CLMBR-T-Base, whose fixed vocabulary covers only 16% of UKB codes. Notably, an LLM-based model achieves superior performance for prediction of disease onset, hospitalization, and mortality, indicating robustness to population and coding shifts.

[828] Target Networks and Over-parameterization Stabilize Off-policy Bootstrapping with Function Approximation

Fengdi Che, Chenjun Xiao, Jincheng Mei, Bo Dai, Ramki Gummadi, Oscar A Ramirez, Christopher K Harris, A. Rupam Mahmood, Dale Schuurmans

Main category: cs.LG

TL;DR: Target networks combined with over-parameterized linear function approximation enable convergence in bootstrapped value estimation with off-policy data, even with truncated trajectories.

DetailsMotivation: To establish convergence guarantees for bootstrapped value estimation in reinforcement learning, particularly with off-policy data and linear function approximation, where standard methods may fail.

Method: Combining target networks with over-parameterized linear function approximation, analyzing temporal difference estimation for prediction and extending to Q-learning for control.

Result: Proves convergence under weaker conditions, provides high-probability value estimation error bounds, and validates empirically on Baird’s counterexample and Four-room task.

Conclusion: The synergy of target networks and over-parameterization ensures convergence in challenging scenarios like off-policy learning and truncated trajectories, with extensions to control settings.

Abstract: We prove that the combination of a target network and over-parameterized linear function approximation establishes a weaker convergence condition for bootstrapped value estimation in certain cases, even with off-policy data. Our condition is naturally satisfied for expected updates over the entire state-action space or learning with a batch of complete trajectories from episodic Markov decision processes. Notably, using only a target network or an over-parameterized model does not provide such a convergence guarantee. Additionally, we extend our results to learning with truncated trajectories, showing that convergence is achievable for all tasks with minor modifications, akin to value truncation for the final states in trajectories. Our primary result focuses on temporal difference estimation for prediction, providing high-probability value estimation error bounds and empirical analysis on Baird’s counterexample and a Four-room task. Furthermore, we explore the control setting, demonstrating that similar convergence conditions apply to Q-learning.

[829] Hallucination Detection in LLMs Using Spectral Features of Attention Maps

Jakub Binkowski, Denis Janiak, Albert Sawczyn, Bogdan Gabrys, Tomasz Kajdanowicz

Main category: cs.LG

TL;DR: The paper proposes LapEigvals, a method that uses spectral features from attention maps (interpreted as graph adjacency matrices) to detect hallucinations in Large Language Models, achieving state-of-the-art performance.

DetailsMotivation: LLMs are prone to hallucinations which is problematic for safety-critical applications. Existing attention-based detection methods have limited effectiveness, so better approaches are needed.

Method: Interpret attention maps as adjacency matrices of graph structures and use the top-k eigenvalues of the Laplacian matrix derived from these attention maps as input to hallucination detection probes.

Result: Empirical evaluations show the approach achieves state-of-the-art hallucination detection performance among attention-based methods, with robustness and generalization demonstrated through ablation studies.

Conclusion: The LapEigvals method effectively leverages spectral features of attention maps for hallucination detection, paving the way for future advancements in this domain.

Abstract: Large Language Models (LLMs) have demonstrated remarkable performance across various tasks but remain prone to hallucinations. Detecting hallucinations is essential for safety-critical applications, and recent methods leverage attention map properties to this end, though their effectiveness remains limited. In this work, we investigate the spectral features of attention maps by interpreting them as adjacency matrices of graph structures. We propose the $\text{LapEigvals}$ method, which utilises the top-$k$ eigenvalues of the Laplacian matrix derived from the attention maps as an input to hallucination detection probes. Empirical evaluations demonstrate that our approach achieves state-of-the-art hallucination detection performance among attention-based methods. Extensive ablation studies further highlight the robustness and generalisation of $\text{LapEigvals}$, paving the way for future advancements in the hallucination detection domain.

[830] MoFO: Momentum-Filtered Optimizer for Mitigating Forgetting in LLM Fine-Tuning

Yupeng Chen, Senmiao Wang, Yushun Zhang, Zhihang Lin, Haozhe Zhang, Weijian Sun, Tian Ding, Ruoyu Sun

Main category: cs.LG

TL;DR: MoFO is a new fine-tuning algorithm that mitigates knowledge forgetting in LLMs by only updating parameters with largest momentum magnitudes, achieving similar performance to standard fine-tuning without requiring pre-training data.

DetailsMotivation: To address the problem of knowledge forgetting during LLM fine-tuning when pre-training data is unavailable, particularly for checkpoint-only open-source models.

Method: Extension of greedy block coordinate descent methods - in each iteration, only updates model parameters with largest momentum magnitudes while keeping others fixed.

Result: Achieves similar fine-tuning performance to default algorithms while effectively mitigating knowledge forgetting, validated through convergence analysis and experiments.

Conclusion: MoFO provides an effective solution for mitigating forgetting during LLM fine-tuning without access to pre-training data.

Abstract: Large language models (LLMs) have demonstrated remarkable capabilities across a wide range of tasks. Typically, LLMs are first pre-trained on large corpora and subsequently fine-tuned on task-specific datasets. However, during fine-tuning, LLMs may forget some knowledge acquired in the pre-training stage, leading to a decline in general capabilities. Existing approaches to mitigate forgetting often rely on access to pre-training data, which may be unavailable in many real-world scenarios–such as fine-tuning checkpoint-only open-source LLMs. To address this challenge, we propose a new fine-tuning algorithm termed Momentum-Filtered Optimizer (MoFO). MoFO is an extension of greedy block coordinate descent (BCD) methods: in each iteration, MoFO only updates the model parameters with the largest momentum magnitudes, while keeping all other parameters fixed. MoFO achieves similar fine-tuning performance to the default fine-tuning algorithm while effectively mitigating knowledge forgetting. We validate MoFO through rigorous convergence analysis and extensive experiments, demonstrating its effectiveness in mitigating forgetting without pre-training data.

[831] $Q\sharp$: Provably Optimal Distributional RL for LLM Post-Training

Jin Peng Zhou, Kaiwen Wang, Jonathan Chang, Zhaolin Gao, Nathan Kallus, Kilian Q. Weinberger, Kianté Brantley, Wen Sun

Main category: cs.LG

TL;DR: Q♯ is a value-based algorithm for KL-regularized RL that learns the optimal Q function using distributional RL, outperforming prior methods in math reasoning benchmarks while maintaining smaller KL divergence to the reference policy.

DetailsMotivation: Existing policy-based RL methods like PPO and DPO often fail to fix shortcuts inherited from pre-training in LLM alignment and reasoning tasks.

Method: Proposes Q♯, a value-based algorithm that guides the reference policy using the optimal regularized Q function learned through distributional RL on an aggregated online dataset.

Result: Empirically outperforms prior baselines in math reasoning benchmarks while maintaining smaller KL divergence to the reference policy. Theoretically establishes reduction from KL-regularized RL to no-regret online learning with variance-dependent bounds.

Conclusion: Q♯ is an effective approach for post-training LLMs that offers both improved performance and theoretical guarantees.

Abstract: Reinforcement learning (RL) post-training is crucial for LLM alignment and reasoning, but existing policy-based methods, such as PPO and DPO, can fall short of fixing shortcuts inherited from pre-training. In this work, we introduce $Q\sharp$, a value-based algorithm for KL-regularized RL that guides the reference policy using the optimal regularized $Q$ function. We propose to learn the optimal $Q$ function using distributional RL on an aggregated online dataset. Unlike prior value-based baselines that guide the model using unregularized $Q$-values, our method is theoretically principled and provably learns the optimal policy for the KL-regularized RL problem. Empirically, $Q\sharp$ outperforms prior baselines in math reasoning benchmarks while maintaining a smaller KL divergence to the reference policy. Theoretically, we establish a reduction from KL-regularized RL to no-regret online learning, providing the first bounds for deterministic MDPs under only realizability. Thanks to distributional RL, our bounds are also variance-dependent and converge faster when the reference policy has small variance. In sum, our results highlight $Q\sharp$ as an effective approach for post-training LLMs, offering both improved performance and theoretical guarantees. The code can be found at https://github.com/jinpz/q_sharp.

[832] LLM as GNN: Graph Vocabulary Learning for Text-Attributed Graph Foundation Models

Xi Zhu, Haochen Xue, Ziwei Zhao, Wujiang Xu, Jingyuan Huang, Minghao Guo, Qifan Wang, Kaixiong Zhou, Imran Razzak, Yongfeng Zhang

Main category: cs.LG

TL;DR: PromptGFM is a Graph Foundation Model that integrates LLMs and GNNs through graph vocabulary learning, enabling cross-graph and cross-task transferability for Text-Attributed Graphs.

DetailsMotivation: Existing approaches for Text-Attributed Graphs suffer from decoupled LLM-GNN architectures with two-stage alignment, OOV token issues causing graph-specific semantics, and incompatibility with prompt templates, limiting cross-graph transferability.

Method: Proposes PromptGFM with two components: (1) Graph Understanding Module that prompts LLMs to replicate GNN workflow in text space for seamless integration, and (2) Graph Inference Module that establishes language-based graph vocabulary for expressiveness and transferability.

Result: Extensive experiments demonstrate superiority and transferability across diverse graphs and tasks.

Conclusion: PromptGFM provides an effective solution for building versatile Graph Foundation Models that generalize across different graphs and tasks through elegant graph-text alignment and language-based graph vocabulary.

Abstract: Text-Attributed Graphs (TAGs), where each node is associated with text descriptions, are ubiquitous in real-world scenarios. They typically exhibit distinctive structure and domain-specific knowledge, motivating the development of a Graph Foundation Model (GFM) that generalizes across diverse graphs and tasks. Despite large efforts to integrate Large Language Models (LLMs) and Graph Neural Networks (GNNs) for TAGs, existing approaches suffer from decoupled architectures with two-stage alignment, limiting their synergistic potential. Even worse, existing methods assign out-of-vocabulary (OOV) tokens to graph nodes, leading to graph-specific semantics, token explosion, and incompatibility with task-oriented prompt templates, which hinders cross-graph and cross-task transferability. To address these challenges, we propose PromptGFM, a versatile GFM for TAGs grounded in graph vocabulary learning. PromptGFM comprises two key components: (1) Graph Understanding Module, which explicitly prompts LLMs to replicate the finest GNN workflow within the text space, facilitating seamless GNN-LLM integration and elegant graph-text alignment; (2) Graph Inference Module, which establishes a language-based graph vocabulary ensuring expressiveness, transferability, and scalability, enabling readable instructions for LLM fine-tuning. Extensive experiments demonstrate our superiority and transferability across diverse graphs and tasks. The code is available at this: https://github.com/agiresearch/PromptGFM.

[833] A Prospect-Theoretic Policy Gradient Framework for Behaviorally Nuanced Reinforcement Learning

Olivier Lepel, Anas Barakat

Main category: cs.LG

TL;DR: This paper introduces a novel policy gradient theorem and algorithm for Cumulative Prospect Theory (CPT) in reinforcement learning, addressing limitations of standard expected utility theory in modeling human decision-making.

DetailsMotivation: Standard RL assumes rational decision-making based on expected utility theory, which doesn't align with actual human preferences. CPT provides a more accurate model of human decision-making that captures diverse risk attitudes and perceptions of gains/losses.

Method: Derived a novel policy gradient theorem for CPT objectives, designed a model-free policy gradient algorithm, analyzed the gradient estimator, and proved asymptotic convergence to first-order stationary points.

Result: The proposed first-order policy gradient algorithm scales better than existing zeroth-order methods to larger state spaces and was validated through simulations.

Conclusion: The theoretical framework offers more flexibility to advance the integration of behavioral decision-making into RL, bridging the gap between standard RL and human-like decision models.

Abstract: Classical reinforcement learning (RL) typically assumes rational decision-making based on expected utility theory. However, this model has been shown to be empirically inconsistent with actual human preferences, as evidenced in psychology and behavioral economics. Cumulative Prospect Theory (CPT) provides a more nuanced model for human-based decision-making, capturing diverse attitudes and perceptions toward risk, gains, and losses. While prior work has integrated CPT with RL to solve CPT policy optimization problems, the understanding and impact of this formulation remain limited. Our contributions are as follows: (a) we derive a novel policy gradient theorem for CPT objectives, generalizing the foundational result in standard RL, (b) we design a model-free policy gradient algorithm for solving the CPT-RL problem, (c) we analyze our policy gradient estimator and prove asymptotic convergence of the algorithm to first-order stationary points, and (d) test its performance through simulations. Notably, our first-order policy gradient algorithm scales better than existing zeroth-order methods to larger state spaces. Our theoretical framework offers more flexibility to advance the integration of behavioral decision-making into RL.

[834] HardNet: Hard-Constrained Neural Networks with Universal Approximation Guarantees

Youngjae Min, Navid Azizan

Main category: cs.LG

TL;DR: HardNet is a framework for building neural networks that inherently satisfy hard constraints without sacrificing model capacity, enabling end-to-end training with guaranteed constraint satisfaction.

DetailsMotivation: Existing approaches use soft constraints through regularization, which offers no guarantee of constraint satisfaction, especially on inputs far from training data - a critical requirement for safety-critical applications.

Method: HardNet appends a differentiable closed-form enforcement layer to the network’s output, allowing unconstrained optimization using standard algorithms while enforcing multiple input-dependent inequality constraints.

Result: HardNet enables efficient and differentiable enforcement of multiple input-dependent inequality constraints, retains universal approximation capabilities, and demonstrates effectiveness in learning with constraints, optimization solvers, and safety-critical control policies.

Conclusion: HardNet provides a practical solution for constructing neural networks that satisfy hard constraints without compromising performance, making it suitable for safety-critical applications where constraint guarantees are essential.

Abstract: Incorporating prior knowledge or specifications of input-output relationships into machine learning models has attracted significant attention, as it enhances generalization from limited data and yields conforming outputs. However, most existing approaches use soft constraints by penalizing violations through regularization, which offers no guarantee of constraint satisfaction, especially on inputs far from the training distribution–an essential requirement in safety-critical applications. On the other hand, imposing hard constraints on neural networks may hinder their representational power, adversely affecting performance. To address this, we propose HardNet, a practical framework for constructing neural networks that inherently satisfy hard constraints without sacrificing model capacity. Unlike approaches that modify outputs only at inference time, HardNet enables end-to-end training with hard constraint guarantees, leading to improved performance. To the best of our knowledge, HardNet is the first method that enables efficient and differentiable enforcement of more than one input-dependent inequality constraint. It allows unconstrained optimization of the network parameters using standard algorithms by appending a differentiable closed-form enforcement layer to the network’s output. Furthermore, we show that HardNet retains neural networks’ universal approximation capabilities. We demonstrate its versatility and effectiveness across various applications: learning with piecewise constraints, learning optimization solvers with guaranteed feasibility, and optimizing control policies in safety-critical systems.

[835] Diffusion Transformers as Open-World Spatiotemporal Foundation Models

Yuan Yuan, Chonghua Han, Jingtao Ding, Guozhen Zhang, Depeng Jin, Yong Li

Main category: cs.LG

TL;DR: UrbanDiT is a diffusion transformer foundation model for urban spatio-temporal learning that unifies diverse data types and tasks, featuring adaptive prompt learning and strong zero-shot generalization capabilities.

DetailsMotivation: Urban environments have complex spatio-temporal dynamics from human activities that need effective modeling for urban system understanding and optimization. Current approaches lack unified handling of diverse data sources and tasks across different cities.

Method: Uses diffusion transformers with an elaborated prompt learning framework that generates both data-driven and task-specific prompts. Unifies grid-based and graph-based data into sequential format and supports multi-task learning across various urban applications.

Result: UrbanDiT demonstrates superior performance across various urban applications, with powerful zero-shot capabilities outperforming nearly all baselines that require training data. It effectively supports bi-directional spatio-temporal prediction, temporal interpolation, spatial extrapolation, and spatio-temporal imputation.

Conclusion: UrbanDiT sets a new benchmark for foundation models in urban spatio-temporal domain by successfully scaling diffusion transformers, unifying diverse data and tasks, and showing strong generalization to open-world scenarios with zero-shot capabilities.

Abstract: The urban environment is characterized by complex spatio-temporal dynamics arising from diverse human activities and interactions. Effectively modeling these dynamics is essential for understanding and optimizing urban systems. In this work, we introduce UrbanDiT, a foundation model for open-world urban spatio-temporal learning that successfully scales up diffusion transformers in this field. UrbanDiT pioneers a unified model that integrates diverse data sources and types while learning universal spatio-temporal patterns across different cities and scenarios. This allows the model to unify both multi-data and multi-task learning, and effectively support a wide range of spatio-temporal applications. Its key innovation lies in the elaborated prompt learning framework, which adaptively generates both data-driven and task-specific prompts, guiding the model to deliver superior performance across various urban applications. UrbanDiT offers three advantages: 1) It unifies diverse data types, such as grid-based and graph-based data, into a sequential format; 2) With task-specific prompts, it supports a wide range of tasks, including bi-directional spatio-temporal prediction, temporal interpolation, spatial extrapolation, and spatio-temporal imputation; and 3) It generalizes effectively to open-world scenarios, with its powerful zero-shot capabilities outperforming nearly all baselines with training data. UrbanDiT sets up a new benchmark for foundation models in the urban spatio-temporal domain. Code and datasets are publicly available at https://github.com/tsinghua-fib-lab/UrbanDiT.

[836] UFT: Unifying Supervised and Reinforcement Fine-Tuning

Mingyang Liu, Gabriele Farina, Asuman Ozdaglar

Main category: cs.LG

TL;DR: UFT is a unified post-training method that combines SFT and RFT, overcoming their limitations and achieving better performance across model sizes while breaking RFT’s exponential sample complexity bottleneck.

DetailsMotivation: To address limitations of existing post-training methods: SFT causes overfitting and limits reasoning in large models, while RFT depends heavily on base model strength and has exponential sample complexity.

Method: Proposed Unified Fine-Tuning (UFT) that integrates SFT and RFT into a single process, enabling exploration while incorporating supervision signals.

Result: UFT outperforms both SFT and RFT across different model sizes and theoretically breaks RFT’s exponential sample complexity bottleneck.

Conclusion: UFT provides a unified approach that bridges memorization and thinking, accelerating convergence on long-horizon reasoning tasks.

Abstract: Post-training has demonstrated its importance in enhancing the reasoning capabilities of large language models (LLMs). The primary post-training methods can be categorized into supervised fine-tuning (SFT) and reinforcement fine-tuning (RFT). SFT is efficient and well-suited for small language models, but it may lead to overfitting and limit the reasoning abilities of larger models. In contrast, RFT generally yields better generalization but depends heavily on the strength of the base model. To address the limitations of SFT and RFT, we propose Unified Fine-Tuning (UFT), a novel post-training paradigm that unifies SFT and RFT into a single, integrated process. UFT enables the model to effectively explore solutions while incorporating informative supervision signals, bridging the gap between memorizing and thinking underlying existing methods. Notably, UFT outperforms both SFT and RFT in general, regardless of model sizes. Furthermore, we theoretically prove that UFT breaks RFT’s inherent exponential sample complexity bottleneck, showing for the first time that unified training can exponentially accelerate convergence on long-horizon reasoning tasks.

[837] Enhancing Efficiency and Exploration in Reinforcement Learning for LLMs

Mengqi Liao, Xiangyu Xi, Ruinian Chen, Jia Leng, Yangen Hu, Ke Zeng, Shuai Liu, Huaiyu Wan

Main category: cs.LG

TL;DR: The paper proposes E3-RL4LLMs, a method that dynamically allocates rollout budgets based on question difficulty and uses adaptive temperature adjustment to maintain exploration during reinforcement learning for large language models.

DetailsMotivation: Existing RL approaches for LLMs inefficiently allocate equal rollouts to all questions, limiting gains on simple problems while providing insufficient training for challenging ones. RL also reduces model exploration ability, potentially capping performance below the base model.

Method: Dynamic rollout budget allocation based on question difficulty, and adaptive dynamic temperature adjustment strategy to maintain stable entropy levels and encourage exploration.

Result: The proposed approach enables more efficient RL training by focusing resources on challenging questions while preserving the model’s exploratory ability to discover correct pathways.

Conclusion: E3-RL4LLMs improves response precision while maintaining exploration capability, addressing inefficiencies in current RL methods for LLMs through dynamic resource allocation and entropy control.

Abstract: Reasoning large language models (LLMs) excel in complex tasks, which has drawn significant attention to reinforcement learning (RL) for LLMs. However, existing approaches allocate an equal number of rollouts to all questions during the RL process, which is inefficient. This inefficiency stems from the fact that training on simple questions yields limited gains, whereas more rollouts are needed for challenging questions to sample correct answers. Furthermore, while RL improves response precision, it limits the model’s exploration ability, potentially resulting in a performance cap below that of the base model prior to RL. To address these issues, we propose a mechanism for dynamically allocating rollout budgets based on the difficulty of the problems, enabling more efficient RL training. Additionally, we introduce an adaptive dynamic temperature adjustment strategy to maintain the entropy at a stable level, thereby encouraging sufficient exploration. This enables LLMs to improve response precision while preserving their exploratory ability to uncover potential correct pathways. The code and data is available on: https://github.com/LiaoMengqi/E3-RL4LLMs

[838] Towards Principled Unsupervised Multi-Agent Reinforcement Learning

Riccardo Zamboni, Mirco Mutti, Marcello Restelli

Main category: cs.LG

TL;DR: This paper addresses unsupervised pre-training in multi-agent reinforcement learning via task-agnostic exploration, proposing a scalable decentralized algorithm and identifying mixture entropy as an optimal objective.

DetailsMotivation: While unsupervised pre-training is well-studied in single-agent RL, little is known about multi-agent settings despite their real-world prevalence. The paper aims to understand alternative problem formulations, theoretical hardness, and practical solutions for multi-agent task-agnostic exploration.

Method: The authors characterize alternative problem formulations, highlight practical challenges, and present a scalable, decentralized, trust-region policy search algorithm for multi-agent task-agnostic exploration.

Result: Numerical validations corroborate theoretical findings and demonstrate that optimizing for mixture entropy provides an excellent trade-off between tractability and performance in challenging domains.

Conclusion: The paper establishes foundations for unsupervised multi-agent reinforcement learning through task-agnostic exploration, showing that mixture entropy optimization offers a practical and effective approach for pre-training policies without task specifications.

Abstract: In reinforcement learning, we typically refer to unsupervised pre-training when we aim to pre-train a policy without a priori access to the task specification, i.e. rewards, to be later employed for efficient learning of downstream tasks. In single-agent settings, the problem has been extensively studied and mostly understood. A popular approach, called task-agnostic exploration, casts the unsupervised objective as maximizing the entropy of the state distribution induced by the agent’s policy, from which principles and methods follow. In contrast, little is known about it in multi-agent settings, which are ubiquitous in the real world. What are the pros and cons of alternative problem formulations in this setting? How hard is the problem in theory, how can we solve it in practice? In this paper, we address these questions by first characterizing those alternative formulations and highlighting how the problem, even when tractable in theory, is non-trivial in practice. Then, we present a scalable, decentralized, trust-region policy search algorithm to address the problem in practical settings. Finally, we provide numerical validations to both corroborate the theoretical findings and pave the way for unsupervised multi-agent reinforcement learning via task-agnostic exploration in challenging domains, showing that optimizing for a specific objective, namely mixture entropy, provides an excellent trade-off between tractability and performances.

[839] DOGe: Defensive Output Generation for LLM Protection Against Knowledge Distillation

Pingzhi Li, Zhen Tan, Mohan Zhang, Huaizhi Qu, Huan Liu, Tianlong Chen

Main category: cs.LG

TL;DR: DOGe is a defense strategy that protects LLMs from knowledge distillation by adversarially fine-tuning the final layer to generate outputs that are useful for legitimate users but misleading for distillation, significantly reducing imitation performance.

DetailsMotivation: LLMs are valuable investments but vulnerable to imitation through knowledge distillation from publicly accessible outputs. Existing defenses like watermarking only detect imitation after the fact or assume access to internal logits, making them ineffective against output-only distillation.

Method: Fine-tune only the final linear layer of the teacher LLM with adversarial loss to subtly modify output behavior. This creates outputs that are accurate for legitimate users but designed to mislead distillation attempts during inference.

Result: While preserving teacher model performance, student models distilled from defensively generated outputs show catastrophically reduced performance, demonstrating effective protection against KD-based imitation.

Conclusion: DOGe provides a practical and efficient safeguard against model imitation through knowledge distillation, operating effectively within realistic API-based access constraints without compromising legitimate user experience.

Abstract: Large Language Models (LLMs) represent substantial intellectual and economic investments, yet their effectiveness can inadvertently facilitate model imitation via knowledge distillation (KD). In practical scenarios, competitors can distill proprietary LLM capabilities by simply observing publicly accessible outputs, akin to reverse-engineering a complex performance by observation alone. Existing protective methods like watermarking only identify imitation post-hoc, while other defenses assume the student model mimics the teacher’s internal logits, rendering them ineffective against distillation purely from observed output text. This paper confronts the challenge of actively protecting LLMs within the realistic constraints of API-based access. We introduce an effective and efficient Defensive Output Generation (DOGe) strategy that subtly modifies the output behavior of an LLM. Its outputs are accurate and useful for legitimate users, yet are designed to be misleading for distillation, significantly undermining imitation attempts. We achieve this by fine-tuning only the final linear layer of the teacher LLM with an adversarial loss. This targeted training approach anticipates and disrupts distillation attempts during inference time. Our experiments show that, while preserving the performance of the teacher model, student models distilled from the defensively generated outputs demonstrate catastrophically reduced performance, demonstrating DOGe as a practical safeguard against KD-based model imitation.

[840] REASONING GYM: Reasoning Environments for Reinforcement Learning with Verifiable Rewards

Zafir Stojanovski, Oliver Stanley, Joe Sharratt, Richard Jones, Abdulhakeem Adefioye, Jean Kaddour, Andreas Köpf

Main category: cs.LG

TL;DR: Reasoning Gym (RG) is a library of reasoning environments for reinforcement learning with verifiable rewards, featuring over 100 data generators across multiple domains and procedural generation for infinite training data with adjustable complexity.

DetailsMotivation: To address the limitation of fixed reasoning datasets by providing a platform that can generate virtually infinite training data with adjustable complexity for continuous evaluation across varying difficulty levels.

Method: Developed Reasoning Gym (RG) with over 100 data generators and verifiers spanning domains including algebra, arithmetic, computation, cognition, geometry, graph theory, logic, and games, using procedural generation approach.

Result: Experimental results demonstrate the efficacy of RG in both evaluating and reinforcement learning of reasoning models.

Conclusion: RG successfully provides a scalable framework for reinforcement learning with verifiable rewards across multiple reasoning domains through procedural data generation.

Abstract: We introduce Reasoning Gym (RG), a library of reasoning environments for reinforcement learning with verifiable rewards. It provides over 100 data generators and verifiers spanning multiple domains including algebra, arithmetic, computation, cognition, geometry, graph theory, logic, and various common games. Its key innovation is the ability to generate virtually infinite training data with adjustable complexity, unlike most previous reasoning datasets, which are typically fixed. This procedural generation approach allows for continuous evaluation across varying difficulty levels. Our experimental results demonstrate the efficacy of RG in both evaluating and reinforcement learning of reasoning models.

[841] Cross-Domain Graph Anomaly Detection via Test-Time Training with Homophily-Guided Self-Supervision

Delaram Pirhayati, Arlei Silva

Main category: cs.LG

TL;DR: GADT3 is a test-time training framework for cross-domain graph anomaly detection that combines supervised and self-supervised learning during training, and adapts to new domains during testing using self-supervised learning with homophily-based affinity scores.

DetailsMotivation: Existing supervised GAD approaches are ineffective across graph domains due to distribution shifts and heterogeneous feature spaces, while labeled anomalies are often scarce in emerging applications.

Method: Combines supervised and self-supervised learning during training; uses test-time adaptation with self-supervised learning leveraging homophily-based affinity scores; includes attention-based edge importance weights, domain-specific encoders, and class-aware regularization.

Result: Significantly outperforms existing approaches across multiple cross-domain settings, achieving average improvements of over 8.2% in AUROC and AUPRC compared to the best competing model.

Conclusion: GADT3 provides an effective framework for cross-domain graph anomaly detection that addresses domain shifts and heterogeneous features through test-time training and domain-invariant anomaly properties.

Abstract: Graph Anomaly Detection (GAD) has demonstrated great effectiveness in identifying unusual patterns within graph-structured data. However, while labeled anomalies are often scarce in emerging applications, existing supervised GAD approaches are either ineffective or not applicable when moved across graph domains due to distribution shifts and heterogeneous feature spaces. To address these challenges, we present GADT3, a novel test-time training framework for cross-domain GAD. GADT3 combines supervised and self-supervised learning during training while adapting to a new domain during test time using only self-supervised learning by leveraging a homophily-based affinity score that captures domain-invariant properties of anomalies. Our framework introduces four key innovations to cross-domain GAD: an effective self-supervision scheme, an attention-based mechanism that dynamically learns edge importance weights during message passing, domain-specific encoders for handling heterogeneous features, and class-aware regularization to address imbalance. Experiments across multiple cross-domain settings demonstrate that GADT3 significantly outperforms existing approaches, achieving average improvements of over 8.2% in AUROC and AUPRC compared to the best competing model.

[842] The Shape of Attraction in UMAP: Exploring the Embedding Forces in Dimensionality Reduction

Mohammad Tariqul Islam, Jason W. Fleischer

Main category: cs.LG

TL;DR: Analysis of UMAP’s attractive and repulsive forces reveals their effects on cluster formation and visualization, leading to improved consistency and interpretability.

DetailsMotivation: To understand how UMAP's attractive and repulsive forces affect cluster formations and visualization, and to make embedding methods more interpretable and robust.

Method: Analyzed UMAP’s force dynamics, compared to contemporary methods, and modified attraction to improve cluster consistency under random initialization.

Result: Repulsion emphasizes differences and controls cluster boundaries, while attractive tension can manifest as both attraction and repulsion in lower dimensions. Modified attraction improved cluster consistency.

Conclusion: The analysis makes UMAP and similar embedding methods more interpretable, robust, and accurate by revealing force dynamics and improving initialization consistency.

Abstract: Uniform manifold approximation and projection (UMAP) is among the most popular neighbor embedding methods. The method relies on attractive and repulsive forces among high-dimensional data points to obtain a low-dimensional embedding. In this paper, we analyze the forces to reveal their effects on cluster formations and visualization and compare UMAP to its contemporaries. Repulsion emphasizes differences, controlling cluster boundaries and inter-cluster distance. Attraction is more subtle, as attractive tension between points can manifest simultaneously as attraction and repulsion in the lower-dimensional mapping. This explains the need for learning rate annealing and motivates the different treatments between attractive and repulsive terms. Moreover, by modifying attraction, we improve the consistency of cluster formation under random initialization. Overall, our analysis makes UMAP and similar embedding methods more interpretable, more robust, and more accurate.

[843] DeepSeek-Inspired Exploration of RL-based LLMs and Synergy with Wireless Networks: A Survey

Yu Qiao, Phuong-Nam Tran, Ji Su Yoon, Loc X. Nguyen, Eui-Nam Huh, Dusit Niyato, Choong Seon Hong

Main category: cs.LG

TL;DR: This survey explores the integration of RL-based LLMs (like DeepSeek) with wireless networks, highlighting mutual benefits: LLMs enhance network optimization while wireless infrastructure enables broad model deployment.

DetailsMotivation: The convergence of RL-based LLMs' strong reasoning capabilities with the growing demand for AI-enabled wireless networks creates synergistic opportunities for mutual enhancement.

Method: The survey reviews network optimization techniques, examines RL-based LLM advancements using DeepSeek as an example, explores domain synergy, and identifies emerging integration directions.

Result: The analysis demonstrates how DeepSeek-style LLMs can enhance wireless network optimization through reasoning and decision-making, while wireless infrastructure supports broad deployment of these models.

Conclusion: The interplay between DeepSeek-style LLMs and wireless networks creates a mutually beneficial relationship that drives innovation in both domains, with emerging directions including quantum, on-device, and neural-symbolic models.

Abstract: Reinforcement learning (RL)-based large language models (LLMs), such as ChatGPT, DeepSeek, and Grok-3, have attracted widespread attention for their remarkable capabilities in multimodal data understanding. Meanwhile, the rapid expansion of information services has led to a growing demand for AI-enabled wireless networks. The open-source DeepSeek models are famous for their innovative designs, such as large-scale pure RL and cost-efficient training, which make them well-suited for practical deployment in wireless networks. By integrating DeepSeek-style LLMs with wireless infrastructures, a synergistic opportunity arises: the DeepSeek-style LLMs enhance network optimization with strong reasoning and decision-making abilities, while wireless infrastructure enables the broad deployment of these models. Motivated by this convergence, this survey presents a comprehensive DeepSeek-inspired exploration of RL-based LLMs in the context of wireless networks. We begin by reviewing key techniques behind network optimization to establish a foundation for understanding DeepSeek-style LLM integration. Next, we examine recent advancements in RL-based LLMs, using DeepSeek models as a representative example. Building on this, we explore the synergy between the two domains, highlighting motivations, challenges, and potential solutions. Finally, we highlight emerging directions for integrating LLMs with wireless networks, such as quantum, on-device, and neural-symbolic LLM models, as well as embodied AI agents. Overall, this survey offers a comprehensive examination of the interplay between DeepSeek-style LLMs and wireless networks, demonstrating how these domains can mutually enhance each other to drive innovation.

[844] Watch the Weights: Unsupervised monitoring and control of fine-tuned LLMs

Ziqian Zhong, Aditi Raghunathan

Main category: cs.LG

TL;DR: A method that analyzes weight differences between fine-tuned and base LLMs to detect backdoors, monitor behavior changes, and audit models without needing training data.

DetailsMotivation: Existing interpretability methods require distributionally similar data, which is unavailable when training data is not released, making it hard to detect novel threats like backdoors.

Method: Analyze top singular vectors of weight differences between fine-tuned and base models, then monitor cosine similarity of activations along these directions.

Result: Detects backdoored models with 100% attack prevention and <1.2% false positive rate; detects unlearned topics with 95.42% accuracy; can uncover fine-tuning focus in commercial models.

Conclusion: Weight-based analysis effectively monitors and controls fine-tuned LLMs without requiring training data, enabling detection of backdoors, unlearning, and model auditing.

Abstract: The releases of powerful open-weight large language models (LLMs) are often not accompanied by access to their full training data. Existing interpretability methods, particularly those based on activations, often require or assume distributionally similar data. This is a significant limitation when detecting and defending against novel potential threats like backdoors, which are by definition out-of-distribution. In this work, we introduce a new method for understanding, monitoring and controlling fine-tuned LLMs that interprets weights, rather than activations, thereby side stepping the need for data that is distributionally similar to the unknown training data. We demonstrate that the top singular vectors of the weight difference between a fine-tuned model and its base model correspond to newly acquired behaviors. By monitoring the cosine similarity of activations along these directions, we can detect salient behaviors introduced during fine-tuning with high precision. For backdoored models that bypasses safety mechanisms when a secret trigger is present, our method stops up to 100% of attacks with a false positive rate below 1.2%. For models that have undergone unlearning, we detect inference on erased topics with accuracy up to 95.42% and can even steer the model to recover “unlearned” information. Besides monitoring, our method also shows potential for pre-deployment model auditing: by analyzing commercial instruction-tuned models (OLMo, Llama, Qwen), we are able to uncover model-specific fine-tuning focus including marketing strategies and Midjourney prompt generation. Our implementation can be found at https://github.com/fjzzq2002/WeightWatch.

[845] Provably Efficient Reward Transfer in Reinforcement Learning with Discrete Markov Decision Processes

Kevin Vora, Yu Zhang

Main category: cs.LG

TL;DR: Q-Manipulation (Q-M) is a new method for reward adaptation in RL that uses Q-function bounds to enable action pruning before learning, improving efficiency when adapting to target rewards using existing source behaviors.

DetailsMotivation: Learning target behaviors from scratch in reinforcement learning is inefficient when source behaviors already exist under the same domain dynamics but different reward functions.

Method: Proposes Q-Manipulation (Q-M) which computes bounds on Q-functions and uses an iterative process (similar to value iteration) to tighten these bounds, enabling action pruning before learning starts. Requires a lite-model that is easy to provide or learn.

Result: Formally proven that Q-M does not affect optimality of returned policy in discrete domains and is provably efficient in sample complexity. Evaluated in synthetic and simulation domains showing effectiveness, generalizability, and practicality.

Conclusion: Q-Manipulation provides an efficient approach to reward adaptation by leveraging existing source behaviors through Q-function manipulation and action pruning, with proven optimality guarantees and practical effectiveness.

Abstract: In this paper, we propose a new solution to reward adaptation (RA) in reinforcement learning, where the agent adapts to a target reward function based on one or more existing source behaviors learned a priori under the same domain dynamics but different reward functions. While learning the target behavior from scratch is possible, it is often inefficient given the available source behaviors. Our work introduces a new approach to RA through the manipulation of Q-functions. Assuming the target reward function is a known function of the source reward functions, we compute bounds on the Q-function and present an iterative process (akin to value iteration) to tighten these bounds. Such bounds enable action pruning in the target domain before learning even starts. We refer to this method as “Q-Manipulation” (Q-M). The iteration process assumes access to a lite-model, which is easy to provide or learn. We formally prove that Q-M, under discrete domains, does not affect the optimality of the returned policy and show that it is provably efficient in terms of sample complexity in a probabilistic sense. Q-M is evaluated in a variety of synthetic and simulation domains to demonstrate its effectiveness, generalizability, and practicality.

[846] CAPO: Towards Enhancing LLM Reasoning through Generative Credit Assignment

Guofu Xie, Yunsheng Shi, Hongtao Tian, Ting Yao, Xiao Zhang

Main category: cs.LG

TL;DR: CAPO introduces a novel credit assignment method for RL with LLMs that uses off-the-shelf LLMs as generative process reward models to provide deterministic token-level feedback, overcoming limitations of existing RLVR methods.

DetailsMotivation: Current RLVR methods assign same reward to every token, hampering precise credit assignment. Existing methods like PPO have inaccurate signals, while Process Reward Models require costly supervision and are unreliable.

Method: CAPO uses an off-the-shelf LLM as a Generative Process Reward Model (LLM-as-GenPRM) to generate step-wise critiques in one pass, providing deterministic token-level credits. Voting mechanisms enhance accuracy and robustness.

Result: CAPO consistently outperforms supervised learning and RL-based fine-tuning methods across four mathematical benchmarks and three out-of-domain benchmarks, helping models learn correct reasoning pathways.

Conclusion: CAPO provides an efficient and effective solution for precise credit assignment in RL with LLMs, enabling better reasoning pathway learning without requiring expensive process supervision.

Abstract: Reinforcement Learning with Verifiable Rewards (RLVR) has improved the reasoning abilities of Large Language Models (LLMs) by using rule-based binary feedback. However, current RLVR methods typically assign the same reward to every token. This coarse-grained feedback hampers precise credit assignment, making it hard for models to identify which reasoning steps lead to success or failure, and often results in suboptimal policies. Methods like PPO provide credit assignment by value estimation, but yield inaccurate and unverifiable signals due to limited sampling. On the other hand, methods using Process Reward Models can provide step-wise rewards but suffer from several key limitations: they require high-quality process supervision labels, the feedback is unreliable due to probabilistic reward modeling, and their application in online reinforcement learning (RL) is time-consuming. To overcome these limitations, we introduce a simple but efficient method-Credit Assignment Policy Optimization (CAPO). Instead of training auxiliary models, CAPO directly leverages an off-the-shelf, general-purpose LLM as a Generative Process Reward Model (LLM-as-GenPRM) to generate all step-wise critique by one pass only based on the correctness of the step itself, providing deterministic token-level credits to refine the tokens that were originally assigned identical rule-based rewards. To further enhance the accuracy and robustness, we employ voting mechanisms that scale with the number of generated critiques. Extensive experiments on various backbones like Llama and Qwen models show that CAPO consistently outperforms supervised learning-based and RL-based fine-tuning methods across four challenging mathematical benchmarks and three out-of-domain benchmarks. Further analysis shows that CAPO can help the model to foster the learning of correct reasoning pathways leading to correct answers.

Mingchen Li, Di Zhuang, Keyu Chen, Dumindu Samaraweera, Morris Chang

Main category: cs.LG

TL;DR: This paper introduces a meta-learning-based unweighted graph poisoning attack that significantly degrades GNN link prediction performance, outperforming existing methods.

DetailsMotivation: GNN models are vulnerable to adversarial attacks, but most research focuses on node classification robustness while link prediction robustness has been neglected. This gap needs to be addressed to ensure stable GNN applications.

Method: Proposed an unweighted graph poisoning attack using meta-learning with weighted scheme strategies to target GNN link prediction models.

Result: Comprehensive experiments across diverse datasets show the approach significantly reduces link prediction performance and consistently outperforms state-of-the-art baselines.

Conclusion: The proposed meta-learning-based poisoning attack effectively compromises GNN link prediction, highlighting the need for improved robustness in this area.

Abstract: Link prediction in graph data uses various algorithms and Graph Nerual Network (GNN) models to predict potential relationships between graph nodes. These techniques have found widespread use in numerous real-world applications, including recommendation systems, community/social networks, and biological structures. However, recent research has highlighted the vulnerability of GNN models to adversarial attacks, such as poisoning and evasion attacks. Addressing the vulnerability of GNN models is crucial to ensure stable and robust performance in GNN applications. Although many works have focused on enhancing the robustness of node classification on GNN models, the robustness of link prediction has received less attention. To bridge this gap, this article introduces an unweighted graph poisoning attack that leverages meta-learning with weighted scheme strategies to degrade the link prediction performance of GNNs. We conducted comprehensive experiments on diverse datasets across multiple link prediction applications to evaluate the proposed method and its parameters, comparing it with existing approaches under similar conditions. Our results demonstrate that our approach significantly reduces link prediction performance and consistently outperforms other state-of-the-art baselines.

[848] Limitations of Normalization in Attention Mechanism

Timur Mudarisov, Mikhail Burtsev, Tatiana Petrova, Radu State

Main category: cs.LG

TL;DR: This paper analyzes limitations of softmax normalization in attention mechanisms, showing that increasing selected tokens reduces model’s ability to distinguish informative tokens and gradient sensitivity poses training challenges.

DetailsMotivation: To understand the theoretical limitations and practical challenges of softmax-based normalization in attention mechanisms, particularly regarding token selection ability and training stability.

Method: Developed theoretical framework for analyzing token selection and geometric separation, conducted experiments with pre-trained GPT-2 model to empirically validate theoretical findings.

Result: As number of selected tokens increases, model’s ability to distinguish informative tokens declines, converging toward uniform selection; gradient sensitivity under softmax normalization causes training challenges, especially at low temperatures.

Conclusion: Softmax-based attention mechanisms have inherent limitations in token selection and training stability, motivating the need for more robust normalization and selection strategies in future attention architectures.

Abstract: This paper investigates the limitations of the normalization in attention mechanisms. We begin with a theoretical framework that enables the identification of the model’s selective ability and the geometric separation involved in token selection. Our analysis includes explicit bounds on distances and separation criteria for token vectors under softmax scaling. Through experiments with pre-trained GPT-2 model, we empirically validate our theoretical results and analyze key behaviors of the attention mechanism. Notably, we demonstrate that as the number of selected tokens increases, the model’s ability to distinguish informative tokens declines, often converging toward a uniform selection pattern. We also show that gradient sensitivity under softmax normalization presents challenges during training, especially at low temperature settings. These findings advance current understanding of softmax-based attention mechanism and motivate the need for more robust normalization and selection strategies in future attention architectures.

[849] Error Broadcast and Decorrelation as a Potential Artificial and Natural Learning Mechanism

Mete Erdogan, Cengiz Pehlevan, Alper T. Erdogan

Main category: cs.LG

TL;DR: EBD is a novel neural network training framework that broadcasts output errors directly to layers using stochastic orthogonality principles, avoiding backpropagation’s weight transport problem while achieving competitive performance.

DetailsMotivation: To address the credit assignment problem in neural networks and circumvent the biologically implausible weight transport requirement of backpropagation.

Method: Uses error broadcasting based on Minimum Mean Square Error estimators’ stochastic orthogonality property, defining layerwise loss functions that penalize correlations between layer activations and output errors.

Result: EBD demonstrates competitive or better performance compared to other error-broadcast methods on benchmark datasets.

Conclusion: EBD provides an efficient, biologically plausible, and principled alternative to backpropagation for neural network training.

Abstract: We introduce Error Broadcast and Decorrelation (EBD), a novel learning framework for neural networks that addresses credit assignment by directly broadcasting output errors to individual layers, circumventing weight transport of backpropagation. EBD is rigorously grounded in the stochastic orthogonality property of Minimum Mean Square Error estimators. This fundamental principle states that the error of an optimal estimator is orthogonal to functions of the input. Guided by this insight, EBD defines layerwise loss functions that directly penalize correlations between layer activations and output errors, thereby establishing a principled foundation for error broadcasting. This theoretically sound mechanism naturally leads to the experimentally observed three-factor learning rule and integrates with biologically plausible frameworks to enhance performance and plausibility. Numerical experiments demonstrate EBD’s competitive or better performance against other error-broadcast methods on benchmark datasets. Our findings establish EBD as an efficient, biologically plausible, and principled alternative for neural network training. The implementation is available at: https://github.com/meterdogan07/error-broadcast-decorrelation.

[850] SpikingBrain: Spiking Brain-inspired Large Models

Yuqi Pan, Yupeng Feng, Jinghao Zhuang, Siyu Ding, Han Xu, Zehao Liu, Bohan Sun, Yuhong Chou, Xuerui Qiu, Anlin Deng, Anjie Hu, Peng Zhou, Man Yao, Jibin Wu, Jian Yang, Guoliang Sun, Bo Xu, Guoqi Li

Main category: cs.LG

TL;DR: SpikingBrain is a family of brain-inspired models that address efficiency bottlenecks in Transformer-based LLMs through linear/hybrid-linear attention with spiking neurons, achieving comparable performance to Transformers with 150B tokens while enabling efficient long-context processing with constant memory inference.

DetailsMotivation: To overcome the quadratic training computation scaling and linear inference memory growth limitations of Transformer-based LLMs, especially for long-context processing, and to enable stable large-scale model development on non-NVIDIA platforms.

Method: Three-pronged approach: (1) Linear and hybrid-linear attention architectures with adaptive spiking neurons; (2) Efficient conversion-based training pipeline and spike coding framework; (3) Customized training frameworks, operator libraries, and parallelism strategies optimized for MetaX hardware.

Result: Developed SpikingBrain-7B (linear LLM) and SpikingBrain-76B (hybrid-linear MoE LLM) that achieve comparable performance to open-source Transformers using only 150B tokens. SpikingBrain-7B achieves 100x speedup in Time to First Token for 4M-token sequences, 23.4% Model FLOPs Utilization, 69.15% sparsity for low-power operation, and stable training on MetaX C550 GPUs.

Conclusion: Brain-inspired mechanisms show strong potential for driving efficient and scalable large model design, demonstrating feasibility of large-scale LLM development on non-NVIDIA platforms with significant improvements in long-sequence training efficiency and constant-memory inference.

Abstract: Mainstream Transformer-based large language models face major efficiency bottlenecks: training computation scales quadratically with sequence length, and inference memory grows linearly, limiting long-context processing. Building large models on non-NVIDIA platforms also poses challenges for stable and efficient training. To address this, we introduce SpikingBrain, a family of brain-inspired models designed for efficient long-context training and inference. SpikingBrain leverages the MetaX GPU cluster and focuses on three aspects: (1) Model Architecture: linear and hybrid-linear attention architectures with adaptive spiking neurons; (2) Algorithmic Optimizations: an efficient, conversion-based training pipeline and a dedicated spike coding framework; (3) System Engineering: customized training frameworks, operator libraries, and parallelism strategies tailored to MetaX hardware. Using these techniques, we develop two models: SpikingBrain-7B, a linear LLM, and SpikingBrain-76B, a hybrid-linear MoE LLM. These models demonstrate the feasibility of large-scale LLM development on non-NVIDIA platforms. SpikingBrain achieves performance comparable to open-source Transformer baselines while using only about 150B tokens for continual pre-training. Our models significantly improve long-sequence training efficiency and deliver inference with (partially) constant memory and event-driven spiking behavior. For example, SpikingBrain-7B attains over 100x speedup in Time to First Token for 4M-token sequences. Training remains stable for weeks on hundreds of MetaX C550 GPUs, with the 7B model reaching a Model FLOPs Utilization of 23.4 percent. The proposed spiking scheme achieves 69.15 percent sparsity, enabling low-power operation. Overall, this work demonstrates the potential of brain-inspired mechanisms to drive the next generation of efficient and scalable large model design.

[851] RHYTHM: Reasoning with Hierarchical Temporal Tokenization for Human Mobility

Haoyu He, Haozheng Luo, Yan Chen, Qi R. Wang

Main category: cs.LG

TL;DR: RHYTHM is a human mobility prediction framework that uses LLMs with temporal tokenization to handle complex dependencies and periodic behaviors, achieving better accuracy with faster training.

DetailsMotivation: Human mobility prediction is challenging due to complex long-range dependencies and multi-scale periodic behaviors that existing methods struggle to capture effectively.

Method: Uses temporal tokenization to partition trajectories into daily segments, encodes them as discrete tokens with hierarchical attention, and enriches tokens with pre-computed prompt embeddings from frozen LLMs to capture interdependencies.

Result: Achieves 2.4% improvement in overall accuracy, 5.0% increase on weekends, and 24.6% reduction in training time compared to state-of-the-art methods on three real-world datasets.

Conclusion: RHYTHM demonstrates that LLMs can serve as effective spatio-temporal predictors for human mobility, offering improved accuracy and computational efficiency through temporal tokenization and frozen backbone architecture.

Abstract: Predicting human mobility is inherently challenging due to complex long-range dependencies and multi-scale periodic behaviors. To address this, we introduce RHYTHM (Reasoning with Hierarchical Temporal Tokenization for Human Mobility), a unified framework that leverages large language models (LLMs) as general-purpose spatio-temporal predictors and trajectory reasoners. Methodologically, RHYTHM employs temporal tokenization to partition each trajectory into daily segments and encode them as discrete tokens with hierarchical attention that captures both daily and weekly dependencies, thereby quadratically reducing the sequence length while preserving cyclical information. Additionally, we enrich token representations by adding pre-computed prompt embeddings for trajectory segments and prediction targets via a frozen LLM, and feeding these combined embeddings back into the LLM backbone to capture complex interdependencies. Computationally, RHYTHM keeps the pretrained LLM backbone frozen, yielding faster training and lower memory usage. We evaluate our model against state-of-the-art methods using three real-world datasets. Notably, RHYTHM achieves a 2.4% improvement in overall accuracy, a 5.0% increase on weekends, and a 24.6% reduction in training time. Code is publicly available at https://github.com/he-h/rhythm.

[852] Improving Coverage in Combined Prediction Sets with Weighted p-values

Gina Wong, Drew Prinster, Suchi Saria, Rama Chellappa, Anqi Liu

Main category: cs.LG

TL;DR: A framework for weighted aggregation of conformal prediction sets that maintains finite-sample validity and achieves tighter coverage bounds than naive aggregation, even with data-dependent weights.

DetailsMotivation: Aggregating multiple conformal prediction sets with individual 1-α coverage weakens the overall guarantee to 1-2α worst-case coverage. There's a need for methods that can aggregate prediction sets more effectively while maintaining validity.

Method: Proposed a weighted aggregation framework where weights are assigned to each prediction set based on their contribution. The method generalizes to data-dependent weights and maintains finite-sample validity through a derived procedure.

Result: The framework achieves tighter coverage bounds that interpolate between the 1-2α guarantee of combined models and the 1-α guarantee of individual models, depending on weight distribution. Experiments in mixture-of-experts settings show adaptive coverage.

Conclusion: The proposed weighted aggregation framework provides flexible control over prediction set aggregation while maintaining validity, making it broadly applicable to settings where weights are learned, such as mixture-of-experts models.

Abstract: Conformal prediction quantifies the uncertainty of machine learning models by augmenting point predictions with valid prediction sets. For complex scenarios involving multiple trials, models, or data sources, conformal prediction sets can be aggregated to create a prediction set that captures the overall uncertainty, often improving precision. However, aggregating multiple prediction sets with individual $1-\alpha$ coverage inevitably weakens the overall guarantee, typically resulting in $1-2\alpha$ worst-case coverage. In this work, we propose a framework for the weighted aggregation of prediction sets, where weights are assigned to each prediction set based on their contribution. Our framework offers flexible control over how the sets are aggregated, achieving tighter coverage bounds that interpolate between the $1-2\alpha$ guarantee of the combined models and the $1-\alpha$ guarantee of an individual model depending on the distribution of weights. Importantly, our framework generalizes to data-dependent weights, as we derive a procedure for weighted aggregation that maintains finite-sample validity even when the weights depend on the data. This extension makes our framework broadly applicable to settings where weights are learned, such as mixture-of-experts (MoE), and we demonstrate through experiments in the MoE setting that our methods achieve adaptive coverage.

[853] Max It or Miss It: Benchmarking LLM On Solving Extremal Problems

Binxin Gao, Jingjun Han

Main category: cs.LG

TL;DR: The paper introduces ExtremBench, a benchmark for evaluating LLMs’ optimization reasoning capabilities through mathematical extremal problems, revealing discrepancies with existing mathematical benchmarks.

DetailsMotivation: To understand the sources and mechanisms of LLMs' reasoning capabilities, particularly in optimization reasoning which underpins critical applications like planning and resource allocation.

Method: Created ExtremBench - 93 standardized extrema-finding problems curated from Chinese Mathematical Olympiad inequality exercises, and evaluated various state-of-the-art open-source LLMs including Qwen3, GPT-OSS, and DeepSeek.

Result: LLMs’ extremal-solving reasoning capabilities don’t align with performance on current mathematical benchmarks (AIME25, MATH-500), with some models showing strong general mathematical reasoning but poor extremal-solving skills, and vice versa.

Conclusion: Existing benchmarks may not comprehensively capture the full spectrum of mathematical reasoning abilities, highlighting a critical gap in current evaluation practices.

Abstract: Test-time scaling has enabled Large Language Models (LLMs) with remarkable reasoning capabilities, particularly in mathematical domains, through intermediate chain-of-thought (CoT) reasoning before generating final answers. However, the specific sources and mechanisms underlying these reasoning capabilities remain insufficiently understood. Optimization reasoning, i.e. finding extrema under constraints, represents a fundamental abstraction that underpins critical applications in planning, control, resource allocation, and prompt search. To systematically evaluate this capability, we introduce ExtremBench, a benchmark dataset for solving mathematical extremal problems, curated from inequality exercises used for Chinese Mathematical Olympiad and transformed into $93$ standardized extrema-finding problems. We conduct extensive evaluations across various state-of-the-art open-source model families, including the Qwen3, GPT-OSS, and DeepSeek. Our results reveal that LLMs’ extremal-solving reasoning capabilities do not always align with those of current mathematical benchmarks such as AIME25 and MATH-500, with some models showing strong general mathematical reasoning but poor extremal-solving skills, and vice versa. This discrepancy highlights a critical gap in current evaluation practices and suggests that existing benchmarks may not comprehensively capture the full spectrum of mathematical reasoning abilities.

[854] When majority rules, minority loses: bias amplification of gradient descent

François Bachoc, Jérôme Bolte, Ryan Boustany, Jean-Michel Loubes

Main category: cs.LG

TL;DR: Theoretical analysis shows standard ML training amplifies bias by favoring majority groups, creating stereotypical predictors that neglect minority features due to population and variance imbalances.

DetailsMotivation: Despite empirical evidence of bias amplification in machine learning, there's poor theoretical understanding of how standard training systematically favors majority groups over minorities.

Method: Developed a formal framework for majority-minority learning tasks, analyzing how population and variance imbalances lead to stereotypical predictors that neglect minority-specific features.

Result: Three key findings: (1) close proximity between ‘full-data’ and stereotypical predictors, (2) dominance of region where training learns only majority traits, (3) lower bound on additional training required to overcome bias.

Conclusion: Standard ML training inherently amplifies bias through systematic preference for majority groups, requiring substantial additional training to learn minority-specific features, as demonstrated in deep learning experiments for tabular and image classification.

Abstract: Despite growing empirical evidence of bias amplification in machine learning, its theoretical foundations remain poorly understood. We develop a formal framework for majority-minority learning tasks, showing how standard training can favor majority groups and produce stereotypical predictors that neglect minority-specific features. Assuming population and variance imbalance, our analysis reveals three key findings: (i) the close proximity between ``full-data’’ and stereotypical predictors, (ii) the dominance of a region where training the entire model tends to merely learn the majority traits, and (iii) a lower bound on the additional training required. Our results are illustrated through experiments in deep learning for tabular and image classification tasks.

[855] Incentivizing Truthful Language Models via Peer Elicitation Games

Baiting Chen, Tong Zhu, Jiale Han, Lexin Li, Gang Li, Xiaowu Dai

Main category: cs.LG

TL;DR: PEG is a training-free, game-theoretic framework that uses peer elicitation between generator and discriminator LLMs to improve factual accuracy and reduce hallucinations without supervision.

DetailsMotivation: LLMs have strong generative capabilities but suffer from inconsistencies and hallucinations, requiring methods to improve factual accuracy without extensive supervision or fine-tuning.

Method: Peer Elicitation Games framework with generator and multiple discriminators from distinct base models, using determinant-based mutual information scoring that incentivizes truthful reporting without ground-truth labels.

Result: Theoretical guarantees show sublinear regret and convergence to truthful Nash equilibrium. Empirical evaluations demonstrate significant improvements in factual accuracy across multiple benchmarks.

Conclusion: PEG provides a practical approach for eliciting truthful behavior from LLMs without supervision or fine-tuning, with proven theoretical guarantees and empirical effectiveness.

Abstract: Large Language Models (LLMs) have demonstrated strong generative capabilities but remain prone to inconsistencies and hallucinations. We introduce Peer Elicitation Games (PEG), a training-free, game-theoretic framework for aligning LLMs through a peer elicitation mechanism involving a generator and multiple discriminators instantiated from distinct base models. Discriminators interact in a peer evaluation setting, where utilities are computed using a determinant-based mutual information score that provably incentivizes truthful reporting without requiring ground-truth labels. We establish theoretical guarantees showing that each agent, via online learning, achieves sublinear regret in the sense their cumulative performance approaches that of the best fixed truthful strategy in hindsight. Moreover, we prove last-iterate convergence to a truthful Nash equilibrium, ensuring that the actual policies used by agents converge to stable and truthful behavior over time. Empirical evaluations across multiple benchmarks demonstrate significant improvements in factual accuracy. These results position PEG as a practical approach for eliciting truthful behavior from LLMs without supervision or fine-tuning.

[856] Understanding Prompt Tuning and In-Context Learning via Meta-Learning

Tim Genewein, Li Kevin Wenliang, Jordi Grau-Moya, Anian Ruoss, Laurent Orseau, Marcus Hutter

Main category: cs.LG

TL;DR: The paper presents a Bayesian theoretical framework for understanding prompt optimization, revealing fundamental limitations of prompting that can only be overcome by weight tuning, and demonstrates that soft prefixes can create effective prompts by manipulating activations.

DetailsMotivation: To develop a conceptual understanding of prompting beyond empirical methods, using Bayesian theory to explain how optimal prompting works and its fundamental limitations.

Method: Uses Bayesian theory to model meta-trained neural networks as Bayesian predictors, studies optimal prompting as conditioning these predictors, and conducts experiments with LSTMs and Transformers comparing prefix-tuning and weight-tuning methods.

Result: The theory provides criteria for when optimal prompting is possible, and experiments show that soft prefixes (real-valued vectors) can create highly effective prompts by manipulating activations in ways hard tokens cannot.

Conclusion: Prompting has fundamental limitations that require weight tuning to overcome, and soft prefixes offer a powerful mechanistic approach beyond the conceptual Bayesian theory for effective prompt optimization.

Abstract: Prompting is one of the main ways to adapt a pretrained model to target tasks. Besides manually constructing prompts, many prompt optimization methods have been proposed in the literature. Method development is mainly empirically driven, with less emphasis on a conceptual understanding of prompting. In this paper we discuss how optimal prompting can be understood through a Bayesian view, which also implies some fundamental limitations of prompting that can only be overcome by tuning weights. The paper explains in detail how meta-trained neural networks behave as Bayesian predictors over the pretraining distribution, whose hallmark feature is rapid in-context adaptation. Optimal prompting can be studied formally as conditioning these Bayesian predictors, yielding criteria for target tasks where optimal prompting is and is not possible. We support the theory with educational experiments on LSTMs and Transformers, where we compare different versions of prefix-tuning and different weight-tuning methods. We also confirm that soft prefixes, which are sequences of real-valued vectors outside the token alphabet, can lead to very effective prompts for trained and even untrained networks by manipulating activations in ways that are not achievable by hard tokens. This adds an important mechanistic aspect beyond the conceptual Bayesian theory.

[857] CLIMB: Class-imbalanced Learning Benchmark on Tabular Data

Zhining Liu, Zihao Li, Ze Yang, Tianxin Wei, Jian Kang, Yada Zhu, Hendrik Hamann, Jingrui He, Hanghang Tong

Main category: cs.LG

TL;DR: CLIMB is a comprehensive benchmark for class-imbalanced learning on tabular data, featuring 73 datasets and 29 CIL algorithms with unified implementations.

DetailsMotivation: Class-imbalanced learning on tabular data is crucial for real-world applications where minority classes represent critical but rare outcomes, but existing benchmarks lack comprehensive coverage and standardized implementations.

Method: Developed a high-quality Python package with unified API designs, detailed documentation, and rigorous code quality controls, implementing 29 representative CIL algorithms across 73 real-world datasets with diverse domains and imbalance levels.

Result: Extensive experiments revealed practical insights including limitations of naive rebalancing, effectiveness of ensemble methods, and the importance of data quality in class-imbalanced learning.

Conclusion: CLIMB provides a standardized benchmark that enables easy implementation and comparison of CIL algorithms, offering valuable insights for researchers and practitioners working with imbalanced tabular data.

Abstract: Class-imbalanced learning (CIL) on tabular data is important in many real-world applications where the minority class holds the critical but rare outcomes. In this paper, we present CLIMB, a comprehensive benchmark for class-imbalanced learning on tabular data. CLIMB includes 73 real-world datasets across diverse domains and imbalance levels, along with unified implementations of 29 representative CIL algorithms. Built on a high-quality open-source Python package with unified API designs, detailed documentation, and rigorous code quality controls, CLIMB supports easy implementation and comparison between different CIL algorithms. Through extensive experiments, we provide practical insights on method accuracy and efficiency, highlighting the limitations of naive rebalancing, the effectiveness of ensembles, and the importance of data quality. Our code, documentation, and examples are available at https://github.com/ZhiningLiu1998/imbalanced-ensemble.

[858] DISCOVER: Automated Curricula for Sparse-Reward Reinforcement Learning

Leander Diaz-Bone, Marco Bagatella, Jonas Hübotter, Andreas Krause

Main category: cs.LG

TL;DR: DISCOVER is a method for directed sparse-reward goal-conditioned RL that selects exploratory goals in the direction of the target task, enabling efficient exploration in high-dimensional, long-horizon environments without prior information.

DetailsMotivation: Sparse-reward RL faces challenges in exploration and long-horizon credit assignment. Prior methods struggle with individual high-dimensional, long-horizon tasks because they explore many unrelated tasks. Solving challenging tasks requires solving simpler relevant tasks that teach necessary skills.

Method: DISCOVER extracts direction from existing RL algorithms to select exploratory goals that lead toward the target task. It connects to principled exploration in bandits and formally bounds the time until target task achievability based on initial distance rather than task space volume.

Result: DISCOVER solves exploration problems beyond the reach of prior state-of-the-art RL exploration methods in high-dimensional environments, demonstrating superior performance in directed sparse-reward goal-conditioned very long-horizon RL.

Conclusion: Directed goal selection through DISCOVER provides effective exploration for challenging sparse-reward tasks by focusing on relevant sub-tasks that build skills toward the target, overcoming limitations of prior exploration methods.

Abstract: Sparse-reward reinforcement learning (RL) can model a wide range of highly complex tasks. Solving sparse-reward tasks is RL’s core premise, requiring efficient exploration coupled with long-horizon credit assignment, and overcoming these challenges is key for building self-improving agents with superhuman ability. Prior work commonly explores with the objective of solving many sparse-reward tasks, making exploration of individual high-dimensional, long-horizon tasks intractable. We argue that solving such challenging tasks requires solving simpler tasks that are relevant to the target task, i.e., whose achieval will teach the agent skills required for solving the target task. We demonstrate that this sense of direction, necessary for effective exploration, can be extracted from existing RL algorithms, without leveraging any prior information. To this end, we propose a method for directed sparse-reward goal-conditioned very long-horizon RL (DISCOVER), which selects exploratory goals in the direction of the target task. We connect DISCOVER to principled exploration in bandits, formally bounding the time until the target task becomes achievable in terms of the agent’s initial distance to the target, but independent of the volume of the space of all tasks. We then perform a thorough evaluation in high-dimensional environments. We find that the directed goal selection of DISCOVER solves exploration problems that are beyond the reach of prior state-of-the-art exploration methods in RL.

[859] Efficient Large Language Model Inference with Neural Block Linearization

Mete Erdogan, Francesco Tonin, Volkan Cevher

Main category: cs.LG

TL;DR: Neural Block Linearization (NBL) accelerates transformer LLM inference by replacing self-attention layers with linear approximations using LMMSE estimators, achieving 32% speedup with <1% accuracy loss.

DetailsMotivation: Transformer-based LLMs have high inference demands that pose deployment challenges, requiring efficient acceleration methods.

Method: NBL replaces self-attention layers with linear approximations derived from Linear Minimum Mean Squared Error estimators, uses Canonical Correlation Analysis to compute approximation error bounds, and selects layers with lowest linearization error for substitution without fine-tuning.

Result: Achieved 32% inference speed increase on DeepSeek-R1-Distill-Llama-8B by applying NBL to 12 self-attention layers with less than 1% accuracy trade-off, while maintaining competitive accuracy on multiple reasoning benchmarks.

Conclusion: NBL is a flexible and promising solution that can be efficiently applied to pre-trained LLMs without fine-tuning to significantly improve inference efficiency.

Abstract: The high inference demands of transformer-based Large Language Models (LLMs) pose substantial challenges in their deployment. To this end, we introduce Neural Block Linearization (NBL), a novel framework for accelerating transformer model inference by replacing self-attention layers with linear approximations derived from Linear Minimum Mean Squared Error estimators. NBL leverages Canonical Correlation Analysis to compute a theoretical upper bound on the approximation error. Then, we use this bound as a criterion for substitution, selecting the LLM layers with the lowest linearization error. NBL can be efficiently applied to pre-trained LLMs without the need for fine-tuning. In experiments, NBL achieves notable computational speed-ups while preserving competitive accuracy on multiple reasoning benchmarks. For instance, applying NBL to 12 self-attention layers in DeepSeek-R1-Distill-Llama-8B increases the inference speed by 32% with less than 1% accuracy trade-off, making it a flexible and promising solution to improve the inference efficiency of LLMs. The implementation is available at: https://github.com/LIONS-EPFL/NBL.

[860] The quest for the GRAph Level autoEncoder (GRALE)

Paul Krzakala, Gabriel Melo, Charlotte Laclau, Florence d’Alché-Buc, Rémi Flamary

Main category: cs.LG

TL;DR: GRALE is a novel graph autoencoder that encodes and decodes graphs of varying sizes using Optimal Transport-inspired loss and differentiable node matching, enabling general pre-training for diverse graph tasks.

DetailsMotivation: Graph representation learning remains challenging but crucial for applications in chemistry and biology, requiring methods that can handle graphs of varying sizes and support diverse downstream tasks.

Method: Proposes GRALE graph autoencoder with attention-based architecture extending Evoformer from AlphaFold, using Optimal Transport-inspired loss and differentiable node matching trained jointly with encoder/decoder.

Result: GRALE enables highly general pre-training applicable to various downstream tasks including classification, regression, graph interpolation, editing, matching, and prediction on both simulated and molecular data.

Conclusion: GRALE provides a powerful framework for graph representation learning that supports diverse applications through its flexible architecture and training approach.

Abstract: Although graph-based learning has attracted a lot of attention, graph representation learning is still a challenging task whose resolution may impact key application fields such as chemistry or biology. To this end, we introduce GRALE, a novel graph autoencoder that encodes and decodes graphs of varying sizes into a shared embedding space. GRALE is trained using an Optimal Transport-inspired loss that compares the original and reconstructed graphs and leverages a differentiable node matching module, which is trained jointly with the encoder and decoder. The proposed attention-based architecture relies on Evoformer, the core component of AlphaFold, which we extend to support both graph encoding and decoding. We show, in numerical experiments on simulated and molecular data, that GRALE enables a highly general form of pre-training, applicable to a wide range of downstream tasks, from classification and regression to more complex tasks such as graph interpolation, editing, matching, and prediction.

[861] RocqStar: Leveraging Similarity-driven Retrieval and Agentic Systems for Rocq generation

Andrei Kozyrev, Nikita Khramov, Gleb Solovev, Anton Podkopaev

Main category: cs.LG

TL;DR: The paper presents a multi-agent system for generating Rocq proofs, featuring a novel premise selection method using self-attentive embedding and multi-agent debate, achieving up to 28% performance improvement.

DetailsMotivation: To improve Rocq proof generation by addressing the challenge of thorough premise selection and leveraging multi-agent collaboration for formal verification.

Method: Proposes a multi-stage agentic system with self-attentive embedder for retrieval-based premise selection, multi-agent debate during planning, and reflection mechanisms for enhanced stability.

Result: Achieved up to 28% relative performance increase, with multi-agent debate boosting proof success rate by 20% overall and nearly doubling it for complex theorems.

Conclusion: The proposed multi-agent approach with premise selection and debate mechanisms significantly enhances Rocq proof generation effectiveness, particularly for complex theorems.

Abstract: Interactive Theorem Proving was repeatedly shown to be fruitful combined with Generative Artificial Intelligence. This paper assesses multiple approaches to Rocq generation and illuminates potential avenues for improvement. We highlight the importance of thorough premise selection for generating Rocq proofs and propose a novel approach, leveraging retrieval via a self-attentive embedder model. The evaluation of the designed approach shows up to 28% relative increase of the generator’s performance. We tackle the problem of writing Rocq proofs using a multi-stage agentic system, tailored for formal verification, and demonstrate its high effectiveness. We conduct an ablation study and demonstrate shows that incorporating multi-agent debate during the planning stage increases the proof success rate by 20% overall and nearly doubles it for complex theorems, while the reflection mechanism further enhances stability and consistency.

[862] HUMAP: Hierarchical Uniform Manifold Approximation and Projection

Wilson E. Marcílio-Jr, Danilo M. Eler, Fernando V. Paulovich, Rafael M. Martins

Main category: cs.LG

TL;DR: HUMAP is a novel hierarchical dimensionality reduction technique that preserves both local and global structures while maintaining mental map consistency during hierarchical exploration, outperforming current approaches.

DetailsMotivation: Hierarchical DR techniques are needed for datasets with multiple granularities and to follow the information visualization mantra of presenting major structures first and details on demand.

Method: HUMAP is a flexible hierarchical dimensionality reduction technique designed to preserve local and global structures and maintain mental map consistency throughout hierarchical exploration.

Result: Empirical evidence shows HUMAP’s superiority over current hierarchical approaches, and a case study demonstrates its application for dataset labelling.

Conclusion: HUMAP is an effective hierarchical DR technique that successfully preserves structural information at multiple scales while maintaining exploration consistency.

Abstract: Dimensionality reduction (DR) techniques help analysts to understand patterns in high-dimensional spaces. These techniques, often represented by scatter plots, are employed in diverse science domains and facilitate similarity analysis among clusters and data samples. For datasets containing many granularities or when analysis follows the information visualization mantra, hierarchical DR techniques are the most suitable approach since they present major structures beforehand and details on demand. This work presents HUMAP, a novel hierarchical dimensionality reduction technique designed to be flexible on preserving local and global structures and preserve the mental map throughout hierarchical exploration. We provide empirical evidence of our technique’s superiority compared with current hierarchical approaches and show a case study applying HUMAP for dataset labelling.

[863] VERINA: Benchmarking Verifiable Code Generation

Zhe Ye, Zhengxu Yan, Jingxuan He, Timothe Kasriel, Kaiyu Yang, Dawn Song

Main category: cs.LG

TL;DR: Verina is a benchmark for verifiable code generation that evaluates code, specifications, and proofs together, revealing significant challenges in proof generation for LLMs.

DetailsMotivation: Current benchmarks don't provide holistic evaluation of verifiable code generation (code + specifications + proofs), limiting progress in ensuring LLM-generated code correctness.

Method: Created Verina benchmark with 189 manually curated Lean coding tasks, including problem descriptions, reference implementations, formal specifications, and test suites for comprehensive evaluation.

Result: Best model (OpenAI o4-mini) achieved 61.4% code correctness, 51.0% specification quality, and only 3.6% proof success rate, showing major gaps in proof generation.

Conclusion: Verina enables rigorous evaluation of verifiable code generation and highlights the need for improved LLM-based theorem provers in verification domains.

Abstract: Large language models (LLMs) are increasingly integrated in software development, but ensuring correctness in LLM-generated code remains challenging and often requires costly manual review. Verifiable code generation – jointly generating code, specifications, and proofs of code-specification alignment – offers a promising path to address this limitation and further unleash LLMs’ benefits in coding. Yet, there exists a significant gap in evaluation: current benchmarks often focus on only individual components rather than providing a holistic evaluation framework of all tasks. In this paper, we introduce Verina (Verifiable Code Generation Arena), a high-quality benchmark enabling a comprehensive and modular evaluation of code, specification, and proof generation as well as their compositions. Verina consists of 189 manually curated coding tasks in Lean, with detailed problem descriptions, reference implementations, formal specifications, and extensive test suites. Our extensive evaluation of state-of-the-art LLMs reveals significant challenges in verifiable code generation, especially in proof generation, underscoring the need for improving LLM-based theorem provers in verification domains. The best model, OpenAI o4-mini, achieves a 61.4% code correctness rate, 51.0% for specification soundness and completeness, and a mere 3.6% proof success rate (based on one trial per task). We hope Verina will catalyze progress in verifiable code generation by providing a rigorous and comprehensive benchmark. We release our dataset on https://huggingface.co/datasets/sunblaze-ucb/verina and our evaluation code on https://github.com/sunblaze-ucb/verina.

[864] Identification and Adaptive Control of Markov Jump Systems: Sample Complexity and Regret Bounds

Yahya Sattar, Zhe Du, Davoud Ataee Tarzanagh, Laura Balzano, Necmiye Ozay, Samet Oymak

Main category: cs.LG

TL;DR: This paper presents an identification-based adaptive control scheme for unknown Markov jump linear systems (MJS) that achieves sublinear regret through episodic learning and certainty equivalent control.

DetailsMotivation: Controlling unknown dynamical systems with time-varying dynamics is crucial for autonomous systems, particularly when dealing with Markov jump linear systems where underlying dynamics change over time.

Method: Proposes a two-phase approach: (1) system identification algorithm to learn dynamics in each mode and Markov transition matrix from a single trajectory, (2) episodic adaptive control scheme combining system identification with certainty equivalent control.

Result: Achieves O(√T) regret with appropriate episode lengths, which improves to O(polylog(T)) with partial system knowledge. Sample complexity of identification is O(1/√T).

Conclusion: The approach effectively handles Markovian jumps and weaker stability notions in MJSs, providing insights into system theoretic quantities affecting learning accuracy and control performance.

Abstract: Learning how to effectively control unknown dynamical systems is crucial for intelligent autonomous systems. This task becomes a significant challenge when the underlying dynamics are changing with time. Motivated by this challenge, this paper considers the problem of controlling an unknown Markov jump linear system (MJS) to optimize a quadratic objective. By taking a model-based perspective, we consider identification-based adaptive control of MJSs. We first provide a system identification algorithm for MJS to learn the dynamics in each mode as well as the Markov transition matrix, underlying the evolution of the mode switches, from a single trajectory of the system states, inputs, and modes. Through martingale-based arguments, sample complexity of this algorithm is shown to be $\mathcal{O}(1/\sqrt{T})$. We then propose an adaptive control scheme that performs system identification together with certainty equivalent control to adapt the controllers in an episodic fashion. Combining our sample complexity results with recent perturbation results for certainty equivalent control, we prove that when the episode lengths are appropriately chosen, the proposed adaptive control scheme achieves $\mathcal{O}(\sqrt{T})$ regret, which can be improved to $\mathcal{O}(polylog(T))$ with partial knowledge of the system. Our proof strategy introduces innovations to handle Markovian jumps and a weaker notion of stability common in MJSs. Our analysis provides insights into system theoretic quantities that affect learning accuracy and control performance. Numerical simulations are presented to further reinforce these insights.

[865] Transfer Q-learning

Elynn Chen, Sai Li, Michael I. Jordan

Main category: cs.LG

TL;DR: This paper proposes transfer learning algorithms for time-inhomogeneous finite-horizon MDPs that enable knowledge transfer from source tasks to target tasks, featuring a novel re-targeting step for cross-stage transfer and demonstrating improved convergence rates and lower regret bounds.

DetailsMotivation: Address challenges in high-dimensional state spaces, time-inhomogeneity, and limited sample availability in healthcare and business applications by leveraging data from related source tasks to improve decision-making in target reinforcement learning tasks.

Method: Developed transfer learning algorithms for both batch and online Q-learning with a novel re-targeting step that enables cross-stage transfer along multiple stages, in addition to cross-task transfer. The approach integrates insights from offline source studies.

Result: Established theoretical justifications showing faster convergence rates for Q*-function estimation in offline RL transfer and lower regret bounds in offline-to-online RL transfer under stage-wise reward similarity and mild design similarity across tasks. Empirical evaluation on synthetic and real datasets supports the theoretical results.

Conclusion: The proposed transfer Q-learning algorithm effectively addresses the challenges of time-inhomogeneous finite-horizon MDPs by enabling knowledge transfer across tasks and stages, with theoretical guarantees and empirical validation demonstrating improved performance in reinforcement learning scenarios with limited data.

Abstract: Time-inhomogeneous finite-horizon Markov decision processes (MDP) are frequently employed to model decision-making in dynamic treatment regimes and other statistical reinforcement learning (RL) scenarios. These fields, especially healthcare and business, often face challenges such as high-dimensional state spaces and time-inhomogeneity of the MDP process, compounded by insufficient sample availability which complicates informed decision-making. To overcome these challenges, we investigate knowledge transfer within time-inhomogeneous finite-horizon MDP by leveraging data from both a target RL task and several related source tasks. We have developed transfer learning (TL) algorithms that are adaptable for both batch and online $Q$-learning, integrating valuable insights from offline source studies. The proposed transfer $Q$-learning algorithm contains a novel {\em re-targeting} step that enables {\em cross-stage transfer} along multiple stages in an RL task, besides the usual {\em cross-task transfer} for supervised learning. We establish the first theoretical justifications of TL in RL tasks by showing a faster rate of convergence of the $Q^*$-function estimation in the offline RL transfer, and a lower regret bound in the offline-to-online RL transfer under stage-wise reward similarity and mild design similarity across tasks. Empirical evidence from both synthetic and real datasets is presented to evaluate the proposed algorithm and support our theoretical results.

[866] Denoising the Future: Top-p Distributions for Moving Through Time

Florian Andreas Marwitz, Ralf Möller, Magnus Bender, Marcel Gehrke

Main category: cs.LG

TL;DR: The paper proposes using only the top-p most probable states in Hidden Markov Models to speed up inference and reduce noise, with bounded error guarantees.

DetailsMotivation: Dynamic probabilistic model inference is computationally expensive, requiring enumeration of all states including those with negligible probabilities, leading to inefficiency and noise propagation.

Method: Use only the top-p states (most probable states with accumulated probability p) for inference, with error bounds related to p and the model’s minimal mixing rate.

Result: Empirical evaluation shows speedups of at least an order of magnitude while maintaining total variation distance error below 0.09.

Conclusion: The top-p states approach effectively denoises future predictions and accelerates inference in Hidden Markov Models with provable error bounds.

Abstract: Inference in dynamic probabilistic models is a complex task involving expensive operations. In particular, for Hidden Markov Models, the whole state space has to be enumerated for advancing in time. Even states with negligible probabilities are considered, resulting in computational inefficiency and increased noise due to the propagation of unlikely probability mass. We propose to denoise the future and speed up inference by using only the top-p states, i.e., the most probable states with accumulated probability p. We show that the error introduced by using only the top-p states is bound by p and the so-called minimal mixing rate of the underlying model. Moreover, in our empirical evaluation, we show that we can expect speedups of at least an order of magnitude, while the error in terms of total variation distance is below 0.09.

[867] What is Memory? A Homological Perspective

Xin Li

Main category: cs.LG

TL;DR: The paper introduces a delta-homology model of memory where recall, learning, and prediction emerge from cycle closure in the brain’s latent manifold, with memory traces as nontrivial homology generators.

DetailsMotivation: To provide a unified framework that explains memory as a topological process rather than static attractors, connecting neural dynamics with mathematical structures like homology and cycle closure.

Method: Represent spike-timing dynamics as spatiotemporal complexes organized into cell posets, formalizing learning and recall through cycle closure under contextual modulation via the Context-Content Uncertainty Principle (CCUP).

Result: Memory is conceptualized as a process where inference trajectories stabilize into nontrivial homology classes when both local synchrony (context) and global recurrence (content) conditions are satisfied.

Conclusion: The delta-homology model offers a topological interpretation of memory where cognition minimizes joint uncertainty between context and content variables through synchronization and recurrence mechanisms.

Abstract: We introduce the delta-homology model of memory, a unified framework in which recall, learning, and prediction emerge from cycle closure, the completion of topologically constrained trajectories within the brain’s latent manifold. A Dirac-like memory trace corresponds to a nontrivial homology generator, representing a sparse, irreducible attractor that reactivates only when inference trajectories close upon themselves. In this view, memory is not a static attractor landscape but a topological process of recurrence, where structure arises through the stabilization of closed loops. Building on this principle, we represent spike-timing dynamics as spatiotemporal complexes, in which temporally consistent transitions among neurons form chain complexes supporting persistent activation cycles. These cycles are organized into cell posets, compact causal representations that encode overlapping and compositional memory traces. Within this construction, learning and recall correspond to cycle closure under contextual modulation: inference trajectories stabilize into nontrivial homology classes when both local synchrony (context) and global recurrence (content) are satisfied. We formalize this mechanism through the Context-Content Uncertainty Principle (CCUP), which states that cognition minimizes joint uncertainty between a high-entropy context variable and a low-entropy content variable. Synchronization acts as a context filter selecting coherent subnetworks, while recurrence acts as a content filter validating nontrivial cycles.

[868] PhysioWave: A Multi-Scale Wavelet-Transformer for Physiological Signal Representation

Yanlong Chen, Mattia Orlandi, Pierangelo Maria Rapa, Simone Benatti, Luca Benini, Yawei Li

Main category: cs.LG

TL;DR: A wavelet-based approach for physiological signal analysis that captures multi-scale time-frequency features, with pretrained models for EMG and ECG, and a unified multi-modal framework integrating EEG for superior performance in downstream tasks.

DetailsMotivation: Physiological signals are often corrupted by motion artifacts, baseline drift, and low-SNR disturbances, with strong non-stationarity and abrupt changes that traditional methods struggle to represent effectively.

Method: Novel wavelet-based approach for multi-scale time-frequency feature extraction, large-scale pretrained models for EMG and ECG, and unified multi-modal framework with dedicated branches for each modality and learnable weighted fusion.

Result: Achieves superior performance and sets new baselines in downstream tasks, effectively addresses challenges like low SNR, high inter-subject variability, and device mismatch, outperforming existing methods on multi-modal tasks.

Conclusion: The wavelet-based architecture provides a solid foundation for analyzing diverse physiological signals, while the multi-modal design points to next-generation physiological signal processing with potential impact on wearable health monitoring, clinical diagnostics, and biomedical applications.

Abstract: Physiological signals are often corrupted by motion artifacts, baseline drift, and other low-SNR disturbances, which pose significant challenges for analysis. Additionally, these signals exhibit strong non-stationarity, with sharp peaks and abrupt changes that evolve continuously, making them difficult to represent using traditional time-domain or filtering methods. To address these issues, a novel wavelet-based approach for physiological signal analysis is presented, aiming to capture multi-scale time-frequency features in various physiological signals. Leveraging this technique, two large-scale pretrained models specific to EMG and ECG are introduced for the first time, achieving superior performance and setting new baselines in downstream tasks. Additionally, a unified multi-modal framework is constructed by integrating pretrained EEG model, where each modality is guided through its dedicated branch and fused via learnable weighted fusion. This design effectively addresses challenges such as low signal-to-noise ratio, high inter-subject variability, and device mismatch, outperforming existing methods on multi-modal tasks. The proposed wavelet-based architecture lays a solid foundation for analysis of diverse physiological signals, while the multi-modal design points to next-generation physiological signal processing with potential impact on wearable health monitoring, clinical diagnostics, and broader biomedical applications. Code and data are available at: github.com/ForeverBlue816/PhysioWave

[869] UniCrossFi: A Unified Framework For Cross-Domain Wi-Fi-based Gesture Recognition

Ke Xu, Zhiyong Zheng, Hongyuan Zhu, Lei Wang, Jiangtao Wang

Main category: cs.LG

TL;DR: UniCrossFi is a unified framework for Wi-Fi sensing that addresses cross-domain problems through semi-supervised domain generalization and physics-informed data augmentation using antenna response consistency.

DetailsMotivation: Wi-Fi sensing systems face severe performance degradation in unseen real-world environments due to domain shift, and existing methods require extensive labeled data which is impractical in real scenarios.

Method: Proposes UniCrossFi with: 1) Semi-Supervised Domain Generalization (SSDG) for limited labeled source data, 2) Antenna Response Consistency (ARC) data augmentation using multi-antenna spatial diversity, 3) Unified Contrastive Objective to prevent separating same-class samples from different domains.

Result: UniCrossFi establishes new state-of-the-art performance, significantly outperforming existing methods across unsupervised domain adaptation, domain generalization, and SSDG benchmarks on Widar and CSIDA datasets.

Conclusion: UniCrossFi provides a principled and practical solution to domain shift challenges, advancing robust Wi-Fi sensing systems that operate effectively with limited labeled data in real-world deployments.

Abstract: Wi-Fi sensing systems are severely hindered by cross domain problem when deployed in unseen real-world environments. Existing methods typically design separate frameworks for either domain adaptation or domain generalization, often relying on extensive labeled data. Existing methods that designed for domain generalization is often relying on extensive labeled data. However, real-world scenarios are far more complex, where the deployed model must be capable of handling generalization under limited labeled source data. To this end, we propose UniCrossFi, a unified framework designed to mitigate performance drop in CSI-based sensing across diverse deployment settings. Our framework not only extends conventional Domain Generalization (DG) to a more practical Semi-Supervised Domain Generalization (SSDG) setting, where only partially labeled source data are available, but also introduces a physics-informed data augmentation strategy, Antenna Response Consistency (ARC). ARC mitigates the risk of learning superficial shortcuts by exploiting the intrinsic spatial diversity of multi-antenna systems, treating signals from different antennas as naturally augmented views of the same event. In addition, we design a Unified Contrastive Objective to prevent conventional contrastive learning from pushing apart samples from different domains that share the same class. We conduct extensive experiments on the public Widar and CSIDA datasets. The results demonstrate that UniCrossFi consistently establishes a new state-of-the-art, significantly outperforming existing methods across all unsupervised domain adaptation, DG, and SSDG benchmarks. UniCrossFi provides a principled and practical solution to the domain shift challenge, advancing the feasibility of robust, real-world Wi-Fi sensing systems that can operate effectively with limited labeled data.

[870] Every Rollout Counts: Optimal Resource Allocation for Efficient Test-Time Scaling

Xinglin Wang, Yiwei Li, Shaoxiong Feng, Peiwen Yuan, Yueqi Zhang, Jiayi Shi, Chuyi Tan, Boyuan Pan, Yao Hu, Kan Li

Main category: cs.LG

TL;DR: DORA is a provably optimal resource allocation method for test-time search that addresses bias in existing methods by allocating resources at the direction level rather than solution level, achieving state-of-the-art performance on mathematical reasoning benchmarks.

DetailsMotivation: Existing test-time search methods inefficiently use compute by favoring reasoning directions with more candidates, leading to suboptimal resource allocation under fixed rollout budgets.

Method: Formulate test-time search as resource allocation problem and propose Direction-Oriented Resource Allocation (DORA) that decouples direction quality from candidate count, allocating resources at direction level.

Result: DORA consistently outperforms strong baselines on MATH500, AIME2024, and AIME2025 benchmarks with comparable computational cost, achieving state-of-the-art accuracy.

Conclusion: DORA provides an optimal resource allocation strategy for test-time scaling that improves LLM performance efficiently, contributing to better understanding of optimal TTS methods.

Abstract: Test-Time Scaling (TTS) improves the performance of Large Language Models (LLMs) by using additional inference-time computation to explore multiple reasoning paths through search. Yet how to allocate a fixed rollout budget most effectively during search remains underexplored, often resulting in inefficient use of compute at test time. To bridge this gap, we formulate test-time search as a resource allocation problem and derive the optimal allocation strategy that maximizes the probability of obtaining a correct solution under a fixed rollout budget. Within this formulation, we reveal a core limitation of existing search methods: solution-level allocation tends to favor reasoning directions with more candidates, leading to theoretically suboptimal and inefficient use of compute. To address this, we propose Direction-Oriented Resource Allocation (DORA), a provably optimal method that mitigates this bias by decoupling direction quality from candidate count and allocating resources at the direction level. To demonstrate DORA’s effectiveness, we conduct extensive experiments on challenging mathematical reasoning benchmarks including MATH500, AIME2024, and AIME2025. The empirical results show that DORA consistently outperforms strong baselines with comparable computational cost, achieving state-of-the-art accuracy. We hope our findings contribute to a broader understanding of optimal TTS for LLMs.

[871] Direct Preference Optimization With Unobserved Preference Heterogeneity: The Necessity of Ternary Preferences

Keertana Chidambaram, Karthik Vinay Seetharaman, Vasilis Syrgkanis

Main category: cs.LG

TL;DR: The paper addresses limitations in RLHF and DPO by showing binary comparisons are insufficient for identifying user preferences and proposing methods to incorporate heterogeneous preferences through EM adaptation and fairness-aware aggregation.

DetailsMotivation: Current RLHF and DPO approaches assume uniform annotator preferences and rely on binary comparisons, overlooking human diversity and the limitations of pairwise feedback.

Method: 1) Theoretical connection to econometrics showing rankings over ≥3 responses ensure identifiability; 2) EM adaptation of DPO to discover latent annotator types and train mixture models; 3) Min-max regret fairness aggregation for equitable generative policies.

Result: Establishes theoretical framework showing binary comparisons are insufficient for preference identification, while rankings provide identifiability. Develops practical algorithms for handling heterogeneous preferences.

Conclusion: Provides a comprehensive theoretical and algorithmic framework for fairness and personalization in generative model alignment, addressing diversity of human preferences through improved preference modeling and equitable aggregation methods.

Abstract: Reinforcement Learning from Human Feedback (RLHF) has become central to aligning large language models with human values, typically by first learning a reward model from preference data which is then used to update the model with reinforcement learning. Recent alternatives such as Direct Preference Optimization (DPO) simplify this pipeline by directly optimizing on preferences. However, both approaches often assume uniform annotator preferences and rely on binary comparisons, overlooking two key limitations: the diversity of human evaluators and the limitations of pairwise feedback. In this work, we address both these issues. First, we connect preference learning in RLHF with the econometrics literature and show that binary comparisons are insufficient for identifying latent user preferences from finite user data and infinite users, while (even incomplete) rankings over three or more responses ensure identifiability. Second, we introduce methods to incorporate heterogeneous preferences into alignment algorithms. We develop an Expectation-Maximization adaptation of DPO that discovers latent annotator types and trains a mixture of LLMs accordingly. Then we propose an aggregation algorithm using a min-max regret fairness criterion to produce a single generative policy with equitable performance guarantees. Together, these contributions establish a theoretical and algorithmic framework for fairness and personalization for diverse users in generative model alignment.

[872] Neural Green’s Operators for Parametric Partial Differential Equations

Hugo Melchers, Joost Prins, Michael Abdelmalik

Main category: cs.LG

TL;DR: This paper introduces Neural Green’s Operators (NGOs) - parametric neural operators derived from finite-dimensional representations of Green’s operators for linear PDEs, which preserve linear action while approximating nonlinear dependence on PDE coefficients using neural networks.

DetailsMotivation: To develop efficient neural operators that reduce complexity from learning entire solution operators to only learning Green's functions, enable resolution of multiple scales, and embed desirable mathematical properties like symmetry and conservation.

Method: Construct NGOs by preserving linear action of Green’s operators on inhomogeneity fields while approximating nonlinear dependence of Green’s function on PDE coefficients using neural networks that take weighted averages of coefficients as input.

Result: NGOs achieve comparable or superior accuracy to other operator networks (DeepONet, VMO, FNO) with similar parameters, generalize better on out-of-distribution data, produce accurate dynamics for time-dependent PDEs, and enable construction of effective matrix preconditioners.

Conclusion: NGOs provide an effective framework for learning solution operators of linear PDEs by explicitly representing Green’s functions, offering improved generalization, mathematical property preservation, and practical applications like preconditioner construction.

Abstract: This work introduces a paradigm for constructing parametric neural operators that are derived from finite-dimensional representations of Green’s operators, with learnable Green’s functions, for linear partial differential equations (PDEs). We refer to such neural operators as Neural Green’s Operators (NGOs). Our construction of NGOs preserves the linear action of Green’s operators on the inhomogeneity fields, while approximating the nonlinear dependence of the Green’s function on the coefficients of the PDE using neural networks that take weighted averages of such coefficients as input. This construction reduces the complexity of the problem from learning the entire solution operator and its dependence on all parameters to only learning the Green’s function and its dependence on the PDE coefficients. Moreover, taking weighted averages, rather than point samples, of input functions decouples the network size from the number of sampling points, enabling efficient resolution of multiple scales in the input fields. Furthermore, we show that our explicit representation of Green’s functions enables the embedding of desirable mathematical attributes in our NGO architectures, such as symmetry, spectral, and conservation properties. Through numerical benchmarks on canonical PDEs, we demonstrate that NGOs achieve comparable or superior accuracy to deep operator networks, variationally mimetic operator networks, and Fourier neural operators with similar parameter counts, while generalizing significantly better when tested on out-of-distribution data. For time-dependent PDEs, we show that NGOs can produce pointwise-accurate dynamics in an auto-regressive manner when trained on a single time step. Finally, we show that we can leverage the explicit representation of Green’s functions returned by NGOs to construct effective matrix preconditioners that accelerate iterative solvers for PDEs.

[873] Absolute abstraction: a renormalisation group approach

Carlo Orientale Caputo, Elias Seiffert, Enrico Frausin, Matteo Marsili

Main category: cs.LG

TL;DR: The paper argues that depth alone is insufficient for developing truly abstract representations in neural networks; the breadth of the training data is equally crucial. Using a renormalisation group approach, they identify the Hierarchical Feature Model as an absolutely abstract representation and validate this through experiments with Deep Belief Networks and auto-encoders.

DetailsMotivation: To challenge the notion that depth alone drives abstraction in neural networks, emphasizing the importance of data breadth in developing truly abstract representations.

Method: The authors use a renormalisation group approach to expand representations across broader data sets, identifying the Hierarchical Feature Model as a fixed point. They test this theory with numerical experiments using Deep Belief Networks and auto-encoders trained on data of varying breadth.

Result: Experiments show that neural network representations approach the Hierarchical Feature Model as data breadth increases and depth grows, aligning with theoretical predictions.

Conclusion: Both depth and data breadth are essential for achieving truly abstract representations in neural networks, with the Hierarchical Feature Model serving as a benchmark for absolute abstraction.

Abstract: Abstraction is the process of extracting the essential features from raw data while ignoring irrelevant details. It is well known that abstraction emerges with depth in neural networks, where deep layers capture abstract characteristics of data by combining lower level features encoded in shallow layers (e.g. edges). Yet we argue that depth alone is not enough to develop truly abstract representations. We advocate that the level of abstraction crucially depends on how broad the training set is. We address the issue within a renormalisation group approach where a representation is expanded to encompass a broader set of data. We take the unique fixed point of this transformation – the Hierarchical Feature Model – as a candidate for a representation which is absolutely abstract. This theoretical picture is tested in numerical experiments based on Deep Belief Networks and auto-encoders trained on data of different breadth. These show that representations in neural networks approach the Hierarchical Feature Model as the data get broader and as depth increases, in agreement with theoretical predictions.

[874] Client Clustering Meets Knowledge Sharing: Enhancing Privacy and Robustness in Personalized Peer-to-Peer Learning

Mohammad Mahdi Maheri, Denys Herasymuk, Hamed Haddadi

Main category: cs.LG

TL;DR: P4 is a decentralized method for personalized learning in IoT that enables private client similarity detection, collaborative group formation, and differentially private knowledge distillation to achieve high accuracy while being robust to poisoning attacks.

DetailsMotivation: The growing adoption of AI in IoT ecosystems requires personalized learning methods that can operate efficiently and privately across resource-constrained devices, while addressing challenges like knowledge transfer, data privacy protection, and resilience against poisoning attacks.

Method: P4 employs a lightweight, fully decentralized algorithm to privately detect client similarity and form collaborative groups. Within groups, clients use differentially private knowledge distillation to co-train their models while maintaining robustness against malicious clients.

Result: P4 achieves 5% to 30% higher accuracy than leading differentially private peer-to-peer approaches and maintains robustness with up to 30% malicious clients. Deployment on resource-constrained devices shows only ~7 seconds overhead for collaborative training between two clients.

Conclusion: P4 successfully addresses the challenges of personalized learning in decentralized IoT settings by providing an efficient, private, and robust solution that outperforms existing approaches while maintaining practical deployment feasibility.

Abstract: The growing adoption of Artificial Intelligence (AI) in Internet of Things (IoT) ecosystems has intensified the need for personalized learning methods that can operate efficiently and privately across heterogeneous, resource-constrained devices. However, enabling effective personalized learning in decentralized settings introduces several challenges, including efficient knowledge transfer between clients, protection of data privacy, and resilience against poisoning attacks. In this paper, we address these challenges by developing P4 (Personalized, Private, Peer-to-Peer) – a method designed to deliver personalized models for resource-constrained IoT devices while ensuring differential privacy and robustness against poisoning attacks. Our solution employs a lightweight, fully decentralized algorithm to privately detect client similarity and form collaborative groups. Within each group, clients leverage differentially private knowledge distillation to co-train their models, maintaining high accuracy while ensuring robustness to the presence of malicious clients. We evaluate P4 on popular benchmark datasets using both linear and CNN-based architectures across various heterogeneity settings and attack scenarios. Experimental results show that P4 achieves 5% to 30% higher accuracy than leading differentially private peer-to-peer approaches and maintains robustness with up to 30% malicious clients. Additionally, we demonstrate its practicality by deploying it on resource-constrained devices, where collaborative training between two clients adds only ~7 seconds of overhead.

[875] Identifiable Latent Bandits: Leveraging observational data for personalized decision-making

Ahmet Zahid Balcıoğlu, Newton Mwai, Emil Carlsson, Fredrik D. Johansson

Main category: cs.LG

TL;DR: Proposes an identifiable latent bandit framework that uses nonlinear independent component analysis to learn representations from historical data, enabling faster exploration and personalization than classical bandits.

DetailsMotivation: Sequential decision-making algorithms like bandits are sample-hungry, making them impractical for personalized medicine where training from scratch for each patient is infeasible due to limited decision points per patient.

Method: Uses nonlinear independent component analysis to provably identify representations from observational data (historical records of decisions and outcomes) that are sufficient to infer optimal actions in new bandit instances.

Result: Substantial improvement over online and offline learning baselines in simulated and semi-synthetic environments when identifying conditions are satisfied, with shorter exploration time than classical bandits.

Conclusion: The identifiable latent bandit framework enables optimal decision-making with reduced exploration time by leveraging historical data through provably identifiable representations.

Abstract: Sequential decision-making algorithms such as multi-armed bandits can find optimal personalized decisions, but are notoriously sample-hungry. In personalized medicine, for example, training a bandit from scratch for every patient is typically infeasible, as the number of trials required is much larger than the number of decision points for a single patient. To combat this, latent bandits offer rapid exploration and personalization beyond what context variables alone can offer, provided that a latent variable model of problem instances can be learned consistently. However, existing works give no guidance as to how such a model can be found. In this work, we propose an identifiable latent bandit framework that leads to optimal decision-making with a shorter exploration time than classical bandits by learning from historical records of decisions and outcomes. Our method is based on nonlinear independent component analysis that provably identifies representations from observational data sufficient to infer optimal actions in new bandit instances. We verify this strategy in simulated and semi-synthetic environments, showing substantial improvement over online and offline learning baselines when identifying conditions are satisfied.

[876] Adaptive Policy Synchronization for Scalable Reinforcement Learning

Rodney Lafuente-Mercado

Main category: cs.LG

TL;DR: ClusterEnv is a lightweight distributed RL framework using the DETACH pattern and Adaptive Policy Synchronization to reduce synchronization overhead while maintaining performance.

DetailsMotivation: Existing RL frameworks tie simulation, training, and infrastructure into rigid systems, making distributed environment execution complex and inefficient.

Method: Uses DETACH pattern to move environment operations to remote workers while keeping learning centralized, with Adaptive Policy Synchronization (APS) that updates policies only when divergence grows too large.

Result: Experiments on discrete control tasks show APS maintains performance while cutting synchronization overhead, with efficient cluster execution.

Conclusion: ClusterEnv provides a flexible distributed RL solution that integrates easily with existing code and reduces communication costs through adaptive synchronization.

Abstract: Scaling reinforcement learning (RL) often requires running environments across many machines, but most frameworks tie simulation, training, and infrastructure into rigid systems. We introduce ClusterEnv, a lightweight interface for distributed environment execution that preserves the familiar Gymnasium API. ClusterEnv uses the DETACH pattern, which moves environment reset() and step() operations to remote workers while keeping learning centralized. To reduce policy staleness without heavy communication, we propose Adaptive Policy Synchronization (APS), where workers request updates only when divergence from the central learner grows too large. ClusterEnv supports both on- and off-policy methods, integrates into existing training code with minimal changes, and runs efficiently on clusters. Experiments on discrete control tasks show that APS maintains performance while cutting synchronization overhead. Source code is available at https://github.com/rodlaf/ClusterEnv.

[877] Navigating Uncertainties in Machine Learning for Structural Dynamics: A Comprehensive Survey of Probabilistic and Non-Probabilistic Approaches in Forward and Inverse Problems

Wang-Ji Yan, Lin-Feng Mei, Jiang Mo, Costas Papadimitriou, Ka-Veng Yuen, Michael Beer

Main category: cs.LG

TL;DR: This paper provides a comprehensive review of uncertainty-aware machine learning approaches for structural dynamics, categorizing methods into probabilistic (Bayesian and frequentist) and non-probabilistic (interval learning, fuzzy learning) techniques.

DetailsMotivation: Machine learning has become powerful in structural dynamics but uncertainties like measurement noise and modeling errors can compromise prediction reliability, highlighting the need for effective uncertainty awareness to enhance robustness.

Method: The review categorizes uncertainty-aware approaches into probabilistic methods (Bayesian and frequentist perspectives) and non-probabilistic methods (interval learning and fuzzy learning), with emphasis on Bayesian neural networks for their uncertainty quantification capabilities.

Result: The paper examines strengths and limitations of each approach and their applications in structural dynamic problems including forward problems (response prediction, sensitivity assessment, reliability analysis) and inverse problems (system identification, model updating, damage identification).

Conclusion: The review identifies research gaps and suggests future directions, providing comprehensive insights to help researchers and practitioners make informed decisions when using ML techniques to address uncertainties in structural dynamic problems.

Abstract: In the era of big data, machine learning (ML) has become a powerful tool in various fields, notably impacting structural dynamics. ML algorithms offer advantages by modeling physical phenomena based on data, even in the absence of underlying mechanisms. However, uncertainties such as measurement noise and modeling errors can compromise the reliability of ML predictions, highlighting the need for effective uncertainty awareness to enhance prediction robustness. This paper presents a comprehensive review on navigating uncertainties in ML, categorizing uncertainty-aware approaches into probabilistic methods (including Bayesian and frequentist perspectives) and non-probabilistic methods (such as interval learning and fuzzy learning). Bayesian neural networks, known for their uncertainty quantification and nonlinear mapping capabilities, are emphasized for their superior performance and potential. The review covers various techniques and methodologies for addressing uncertainties in ML, discussing fundamentals and implementation procedures of each method. While providing a concise overview of fundamental concepts, the paper refrains from in-depth critical explanations. Strengths and limitations of each approach are examined, along with their applications in structural dynamic forward problems like response prediction, sensitivity assessment, and reliability analysis, and inverse problems like system identification, model updating, and damage identification. Additionally, the review identifies research gaps and suggests future directions for investigations, aiming to provide comprehensive insights to the research community. By offering an extensive overview of both probabilistic and non-probabilistic approaches, this review aims to assist researchers and practitioners in making informed decisions when utilizing ML techniques to address uncertainties in structural dynamic problems.

[878] ReDi: Rectified Discrete Flow

Jaehoon Yoo, Wonjung Kim, Seunghoon Hong

Main category: cs.LG

TL;DR: ReDi is a novel iterative method that reduces factorization error in Discrete Flow-based Models by rectifying couplings between distributions, enabling efficient few-step generation with theoretical guarantees of convergence.

DetailsMotivation: Discrete Flow-based Models suffer from slow sampling speeds due to iterative decoding processes caused by factorization approximation errors in high-dimensional data handling.

Method: Proposed Rectified Discrete Flow (ReDi) method that reduces factorization error by rectifying couplings between source and target distributions, with theoretical proof of monotonic decreasing Conditional Total Correlation.

Result: ReDi significantly reduces Conditional TC and enables few-step generation, with rectified couplings being suitable for training efficient one-step models on image generation.

Conclusion: ReDi provides a simple and theoretically grounded approach for efficient discrete data synthesis, addressing the few-step generation challenge in DFMs.

Abstract: Discrete Flow-based Models (DFMs) are powerful generative models for high-quality discrete data but typically suffer from slow sampling speeds due to their reliance on iterative decoding processes. This reliance on a multi-step process originates from the factorization approximation of DFMs, which is necessary for handling high-dimensional data. In this paper, we analyze the factorization approximation error using Conditional Total Correlation (TC), and reveal its dependence on the coupling. To address the challenge of efficient few-step generation, we propose Rectified Discrete Flow (ReDi), a novel iterative method that reduces the underlying factorization error (measured as Conditional TC) by rectifying the coupling between source and target distributions. We theoretically prove that each ReDi step guarantees a monotonic decreasing Conditional TC, ensuring its convergence. Empirically, ReDi significantly reduces Conditional TC and enables few-step generation. Moreover, we demonstrate that the rectified couplings are well-suited for training efficient one-step models on image generation. ReDi offers a simple and theoretically grounded approach for tackling the few-step challenge, providing a new perspective on efficient discrete data synthesis. Code is available at https://github.com/Ugness/ReDi_discrete.

[879] Solving Oscillator Ordinary Differential Equations in the Time Domain with High Performance via Soft-constrained Physics-informed Neural Network with Small Data

Kai-liang Lu

Main category: cs.LG

TL;DR: Soft-constrained PINN method effectively solves ODEs with minimal data (1-2 training points) and handles noise, achieving comparable precision to classical methods while incorporating physical constraints.

DetailsMotivation: Address the challenge of sparse and noisy data in scientific applications by leveraging physics-informed neural networks to achieve strong generalization with minimal labeled data.

Method: Soft-constrained PINN approach using DeepXDE framework, incorporating physical laws as regularization terms in loss function, with collocation points that don’t require labels.

Result: PINN requires only 1-2 training data points for first/second-order ODEs respectively, achieves equivalent precision to classical methods, trains quickly (seconds for scalar ODEs), and handles nonlinear systems like Duffing oscillators effectively.

Conclusion: PINN provides a computationally efficient alternative to classical ODE solvers with minimal data requirements, easily extensible to PDEs and suitable for Digital Twin applications.

Abstract: In many scientific and engineering (e.g., physical, biochemical, medical) practices, data generated through expensive experiments or large-scale simulations, are often sparse and noisy. Physics-informed neural network (PINN) incorporates physical information and knowledge into network topology or computational processes as model priors, with the unique advantage of achieving strong generalization with small data. This study aims to investigate the performance characteristics of the soft-constrained PINN method to solving typical linear and nonlinear ordinary differential equations (ODEs) such as primer, Van der Pol and Duffing oscillators, especially the effectiveness, efficiency, and robustness to noise with minimal data. It is verified that the soft-constrained PINN significantly reduces the need for labeled data. With the aid of appropriate collocation points no need to be labeled, it can predict and also extrapolate with minimal data. First-order and second-order ODEs, no matter linear or nonlinear oscillators, require only one and two training data (containing initial values) respectively, just like classical analytic or Runge-Kutta methods, and with equivalent precision and comparable efficiency (fast training in seconds for scalar ODEs). Furthermore, it can conveniently impose a physical law (e.g., conservation of energy) constraint by adding a regularization term to the total loss function, improving the performance to deal with various complexities such as nonlinearity like Duffing. The DeepXDE-based PINN implementation is light code and can be efficiently trained on both GPU and CPU platforms. The mathematical and computational framework of this alternative and feasible PINN method to ODEs, can be easily extended to PDEs, etc., and is becoming a favorable catalyst for the era of Digital Twins.

[880] Channel Matters: Estimating Channel Influence for Multivariate Time Series

Muyao Wang, Zeke Xie, Bo Chen, Hongwei Liu, James Kwok

Main category: cs.LG

TL;DR: Proposes Channel-wise Influence (ChInf) method - the first to estimate influence of different channels in Multivariate Time Series (MTS), enabling improved model performance and interpretability without retraining.

DetailsMotivation: Channel information is critical for MTS tasks but channel-centric methods are under-explored, with no previous work studying counterfactual effects between channels and model performance.

Method: Developed ChInf method that estimates influence of different channels in MTS, and derived two channel-wise algorithms by incorporating ChInf into classic MTS tasks.

Result: ChInf-based methods rank top-1 in MTS anomaly detection and data pruning tasks, outperforming previous influence functions that don’t perform well on MTS problems.

Conclusion: ChInf demonstrates superiority and necessity for MTS analysis, providing effective channel-wise influence estimation for improved performance and interpretability.

Abstract: The influence function serves as an efficient post-hoc interpretability tool that quantifies the impact of training data modifications on model parameters, enabling enhanced model performance, improved generalization, and interpretability insights without the need for expensive retraining processes. Recently, Multivariate Time Series (MTS) analysis has become an important yet challenging task, attracting significant attention. While channel extremely matters to MTS tasks, channel-centric methods are still largely under-explored for MTS. Particularly, no previous work studied the effects of channel information of MTS in order to explore counterfactual effects between these channels and model performance. To fill this gap, we propose a novel Channel-wise Influence (ChInf) method that is the first to estimate the influence of different channels in MTS. Based on ChInf,we naturally derived two channel-wise algorithms by incorporating ChInf into classic MTS tasks. Extensive experiments demonstrate the effectiveness of ChInf and ChInf-based methods in critical MTS analysis tasks, such as MTS anomaly detection and MTS data pruning. Specifically, our ChInf-based methods rank top-1 among all methods for comparison, while previous influence functions do not perform well on MTS anomaly detection tasks and MTS data pruning problem. This fully supports the superiority and necessity of ChInf.

[881] Riemannian Federated Learning via Averaging Gradient Streams

Zhenwei Huang, Wen Huang, Pratik Jawanpuria, Bamdev Mishra

Main category: cs.LG

TL;DR: This paper introduces RFedAGS, a Riemannian federated learning algorithm that handles partial participation and data heterogeneity using gradient stream aggregation, with proven convergence guarantees.

DetailsMotivation: Federated learning has been well-studied in Euclidean settings but lacks investigation in Riemannian settings, particularly for challenges like partial participation and data heterogeneity among agents.

Method: Proposes RFedAGS algorithm based on averaging gradient streams for server aggregation, designed to handle partial participation and data heterogeneity in Riemannian federated learning.

Result: Theoretical analysis shows RFedAGS has global convergence with sublinear rate under decaying step sizes, and converges to neighborhood of stationary point/solution under fixed step sizes. Experiments demonstrate good performance on synthetic and real-world data.

Conclusion: RFedAGS successfully addresses Riemannian federated learning challenges with proven convergence properties and empirical validation.

Abstract: Federated learning (FL) as a distributed learning paradigm has a significant advantage in addressing large-scale machine learning tasks. In the Euclidean setting, FL algorithms have been extensively studied with both theoretical and empirical success. However, there exist few works that investigate federated learning algorithms in the Riemannian setting. In particular, critical challenges such as partial participation and data heterogeneity among agents are not explored in the Riemannian federated setting. This paper presents and analyzes a Riemannian FL algorithm, called RFedAGS, based on a new efficient server aggregation – averaging gradient streams, which can simultaneously handle partial participation and data heterogeneity. We theoretically show that the proposed RFedAGS has global convergence and sublinear convergence rate under decaying step sizes cases; and converges sublinearly/linearly to a neighborhood of a stationary point/solution under fixed step sizes cases. These analyses are based on a vital and non-trivial assumption induced by partial participation, which is shown to hold with high probability. Extensive experiments conducted on synthetic and real-world data demonstrate the good performance of RFedAGS.

[882] Intrinsic Dimensionality of Fermi-Pasta-Ulam-Tsingou High-Dimensional Trajectories Through Manifold Learning: A Linear Approach

Gionni Marchetti

Main category: cs.LG

TL;DR: A data-driven approach using unsupervised machine learning (PCA) reveals that the intrinsic dimension of FPUT model trajectories increases with nonlinear strength, and suggests quasi-periodic motion on low-dimensional manifolds explains energy recurrences.

DetailsMotivation: To understand the intrinsic dimensionality of high-dimensional trajectories in the Fermi-Pasta-Ulam-Tsingou model and its relationship to nonlinear strength and energy recurrences.

Method: Applied principal component analysis (PCA) to trajectory data with 4,000,000 datapoints from the FPUT β model with 32 coupled oscillators, using multiple methods (participation ratio, Kaiser rule, Kneedle algorithm) to estimate intrinsic dimension.

Result: Intrinsic dimension m* increases with model nonlinearity. In weakly nonlinear regime with first mode excitation, participation ratio estimates m* = 2, 3, suggesting quasi-periodic motion on low-dimensional Riemannian manifolds.

Conclusion: The characteristic energy recurrences in the FPUT model can be explained by quasi-periodic motion on low-dimensional manifolds, with intrinsic dimension increasing with nonlinear strength.

Abstract: A data-driven approach based on unsupervised machine learning is proposed to infer the intrinsic dimension $m^{\ast}$ of the high-dimensional trajectories of the Fermi-Pasta-Ulam-Tsingou (FPUT) model. Principal component analysis (PCA) is applied to trajectory data consisting of $n_s = 4,000,000$ datapoints, of the FPUT $\beta$ model with $N = 32$ coupled oscillators, revealing a critical relationship between $m^{\ast}$ and the model’s nonlinear strength. By estimating the intrinsic dimension $m^{\ast}$ using multiple methods (participation ratio, Kaiser rule, and the Kneedle algorithm), it is found that $m^{\ast}$ increases with the model nonlinearity. Interestingly, in the weakly nonlinear regime, for trajectories initialized by exciting the first mode, the participation ratio estimates $m^{\ast} = 2, 3$, strongly suggesting that quasi-periodic motion on a low-dimensional Riemannian manifold underlies the characteristic energy recurrences observed in the FPUT model.

[883] FGBench: A Dataset and Benchmark for Molecular Property Reasoning at Functional Group-Level in Large Language Models

Xuan Liu, Siru Ouyang, Xianrui Zhong, Jiawei Han, Huimin Zhao

Main category: cs.LG

TL;DR: FGBench is a dataset of 625K molecular property reasoning problems with fine-grained functional group annotations, designed to enhance LLMs’ understanding of structure-property relationships in chemistry.

DetailsMotivation: Existing datasets focus on molecular-level properties but overlook functional group information, which provides valuable prior knowledge linking molecular structures with textual descriptions for building more interpretable, structure-aware LLMs.

Method: Created FGBench dataset with 625K problems featuring precisely annotated and localized functional groups across 245 different FGs in three categories: single FG impacts, multiple FG interactions, and direct molecular comparisons.

Result: Benchmarking state-of-the-art LLMs on 7K curated data shows current models struggle with FG-level property reasoning, highlighting the need to enhance reasoning capabilities for chemistry tasks.

Conclusion: FGBench provides a foundational framework for generating datasets with functional group-level information to help LLMs better understand fine-grained molecular structure-property relationships, advancing molecular design and drug discovery.

Abstract: Large language models (LLMs) have gained significant attention in chemistry. However, most existing datasets center on molecular-level property prediction and overlook the role of fine-grained functional group (FG) information. Incorporating FG-level data can provide valuable prior knowledge that links molecular structures with textual descriptions, which can be used to build more interpretable, structure-aware LLMs for reasoning on molecule-related tasks. Moreover, LLMs can learn from such fine-grained information to uncover hidden relationships between specific functional groups and molecular properties, thereby advancing molecular design and drug discovery. Here, we introduce FGBench, a dataset comprising 625K molecular property reasoning problems with functional group information. Functional groups are precisely annotated and localized within the molecule, which ensures the dataset’s interoperability thereby facilitating further multimodal applications. FGBench includes both regression and classification tasks on 245 different functional groups across three categories for molecular property reasoning: (1) single functional group impacts, (2) multiple functional group interactions, and (3) direct molecular comparisons. In the benchmark of state-of-the-art LLMs on 7K curated data, the results indicate that current LLMs struggle with FG-level property reasoning, highlighting the need to enhance reasoning capabilities in LLMs for chemistry tasks. We anticipate that the methodology employed in FGBench to construct datasets with functional group-level information will serve as a foundational framework for generating new question-answer pairs, enabling LLMs to better understand fine-grained molecular structure-property relationships. The dataset and evaluation code are available at https://github.com/xuanliugit/FGBench.

[884] OneProt: Towards Multi-Modal Protein Foundation Models

Klemens Flöge, Srisruthi Udayakumar, Johanna Sommer, Marie Piraud, Stefan Kesselheim, Vincent Fortuin, Stephan Günneman, Karel J van der Weg, Holger Gohlke, Erinc Merdivan, Alina Bazarova

Main category: cs.LG

TL;DR: OneProt is a multi-modal AI system for proteins that integrates structural, sequence, text, and binding site data using ImageBind framework with lightweight fine-tuning. It uses Graph Neural Networks and transformers to achieve strong performance in protein retrieval and downstream tasks.

DetailsMotivation: To extend multi-modal AI systems beyond text and vision to proteins, enabling better integration of diverse protein information spaces including structure, sequence, text descriptions, and binding sites.

Method: Uses ImageBind framework to align latent spaces of protein modality encoders with lightweight fine-tuning. Employs Graph Neural Networks and transformer architectures, focusing on pairwise alignment with sequence data rather than requiring full matches.

Result: Demonstrates strong performance in retrieval tasks and downstream baselines including enzyme function prediction and binding site analysis. Enables transfer of representational information between encoders, improving distinction of evolutionarily related sequences.

Conclusion: Expands horizons of multi-modal protein models, paving way for transformative applications in drug discovery, biocatalytic reaction planning, and protein engineering. Binding site encoder identified as particularly significant contributor to performance.

Abstract: Recent advances in Artificial Intelligence have enabled multi-modal systems to model and translate diverse information spaces. Extending beyond text and vision, we introduce OneProt, a multi-modal AI for proteins that integrates structural, sequence, text, and binding site data. Using the ImageBind framework, OneProt aligns the latent spaces of protein modality encoders in a lightweight fine-tuning scheme that focuses on pairwise alignment with sequence data rather than requiring full matches. This novel approach comprises a mix of Graph Neural Networks and transformer architectures. It demonstrates strong performance in retrieval tasks and showcases the efficacy of multi-modal systems in Protein Machine Learning through a broad spectrum of downstream baselines, including enzyme function prediction and binding site analysis. Furthermore, OneProt enables the transfer of representational information from specialized encoders to the sequence encoder, enhancing capabilities for distinguishing evolutionarily related and unrelated sequences and exhibiting representational properties where evolutionarily related proteins align in similar directions within the latent space. In addition, we extensively investigate modality ablations to identify the encoders that contribute most to predictive performance, highlighting the significance of the binding site encoder, which has not been used in similar models previously. This work expands the horizons of multi-modal protein models, paving the way for transformative applications in drug discovery, biocatalytic reaction planning, and protein engineering.

[885] SAFES: Sequential Privacy and Fairness Enhancing Data Synthesis for Responsible AI

Spencer Giddens, Xiaon Lang, Fang Liu

Main category: cs.LG

TL;DR: SAFES is a sequential procedure that combines differential privacy data synthesis with fairness-aware preprocessing to address both privacy and fairness concerns simultaneously in synthetic data generation.

DetailsMotivation: Most prior work treats data privacy and decision fairness separately, and existing approaches that consider both are limited to specific learning tasks, lacking generalizability.

Method: SAFES sequentially combines DP data synthesis with fairness-aware data preprocessing, allowing flexible navigation of privacy-fairness-utility trade-offs using different DP synthesizers and fairness preprocessing methods.

Result: Empirical evaluations on multiple real datasets show that for reasonable privacy loss, SAFES-generated synthetic data achieves significantly improved fairness metrics with relatively low utility loss.

Conclusion: SAFES provides an effective framework for simultaneously addressing privacy and fairness concerns in synthetic data generation, demonstrating practical trade-offs between privacy, fairness, and utility.

Abstract: As data-driven and AI-based decision making gains widespread adoption across disciplines, it is crucial that both data privacy and decision fairness are appropriately addressed. Although differential privacy (DP) provides a robust framework for guaranteeing privacy and methods are available to improve fairness, most prior work treats the two concerns separately. Even though there are existing approaches that consider privacy and fairness simultaneously, they typically focus on a single specific learning task, limiting their generalizability. In response, we introduce SAFES, a Sequential PrivAcy and Fairness Enhancing data Synthesis procedure that sequentially combines DP data synthesis with a fairness-aware data preprocessing step. SAFES allows users flexibility in navigating the privacy-fairness-utility trade-offs. We illustrate SAFES with different DP synthesizers and fairness-aware data preprocessing methods and run extensive experiments on multiple real datasets to examine the privacy-fairness-utility trade-offs of synthetic data generated by SAFES. Empirical evaluations demonstrate that for reasonable privacy loss, SAFES-generated synthetic data can achieve significantly improved fairness metrics with relatively low utility loss.

[886] Breaking the Exploration Bottleneck: Rubric-Scaffolded Reinforcement Learning for General LLM Reasoning

Yang Zhou, Sunzhu Li, Shunyu Liu, Wenkai Fang, Kongcheng Zhang, Jiale Zhao, Jingwen Yang, Yihe Zhou, Jianwei Lv, Tongya Zheng, Hengtong Lu, Wei Chen, Yan Xie, Mingli Song

Main category: cs.LG

TL;DR: RuscaRL is a novel RL framework that uses checklist-style rubrics to break the exploration bottleneck in LLM reasoning by providing explicit scaffolding during rollout generation and verifiable rewards during training.

DetailsMotivation: Current RL approaches for LLMs face a dilemma where improvement requires high-quality samples, but exploration is limited by the LLMs' inherent capabilities, creating a cycle where unexplored patterns cannot be learned.

Method: RuscaRL introduces rubric scaffolding: (1) checklist-style rubrics guide diverse high-quality responses during rollout, with gradual decay to encourage internalization; (2) rubrics serve as references for robust LLM-as-a-Judge scoring to enable effective RL training.

Result: RuscaRL significantly boosts performance across benchmarks, increasing Qwen2.5-7B-Instruct from 23.6 to 50.3 on HealthBench-500 (surpassing GPT-4.1), and achieving 61.1 with Qwen3-30B-A3B-Instruct, outperforming OpenAI-o3.

Conclusion: The proposed RuscaRL framework effectively expands reasoning boundaries by breaking the exploration bottleneck through instructional scaffolding, demonstrating superior performance on general reasoning tasks.

Abstract: Recent advances in Large Language Models (LLMs) have underscored the potential of Reinforcement Learning (RL) to facilitate the emergence of reasoning capabilities. Despite the encouraging results, a fundamental dilemma persists as RL improvement relies on learning from high-quality samples, yet the exploration for such samples remains bounded by the inherent limitations of LLMs. This, in effect, creates an undesirable cycle in which what cannot be explored cannot be learned. In this work, we propose Rubric-Scaffolded Reinforcement Learning (RuscaRL), a novel instructional scaffolding framework designed to break the exploration bottleneck for general LLM reasoning. Specifically, RuscaRL introduces checklist-style rubrics as (1) explicit scaffolding for exploration during rollout generation, where different rubrics are provided as external guidance within task instructions to steer diverse high-quality responses. This guidance is gradually decayed over time, encouraging the model to internalize the underlying reasoning patterns; (2) verifiable rewards for exploitation during model training, where we can obtain robust LLM-as-a-Judge scores using rubrics as references, enabling effective RL on general reasoning tasks. Extensive experiments demonstrate the superiority of the proposed RuscaRL across various benchmarks, effectively expanding reasoning boundaries under the Best-of-N evaluation. Notably, RuscaRL significantly boosts Qwen2.5-7B-Instruct from 23.6 to 50.3 on HealthBench-500, surpassing GPT-4.1. Furthermore, our fine-tuned variant on Qwen3-30B-A3B-Instruct achieves 61.1 on HealthBench-500, outperforming leading LLMs including OpenAI-o3. Our code is available at https://github.com/IANNXANG/RuscaRL.

[887] Understanding Generalization of Federated Learning: the Trade-off between Model Stability and Optimization

Dun Zeng, Zheshun Wu, Shiyu Liu, Yu Pan, Xiaoying Tang, Zenglin Xu

Main category: cs.LG

TL;DR: The paper introduces Libra, a generalization dynamics analysis framework for Federated Learning that analyzes the trade-off between model stability and gradient norms to improve excess risk minimization.

DetailsMotivation: Existing FL approaches struggle with data heterogeneity causing inconsistent local optima, and current analysis methods (convergence analysis and algorithmic stability) don't adequately capture generalization performance for non-convex neural networks.

Method: Proposes Libra framework for algorithm-dependent excess risk minimization, analyzing trade-offs between model stability and gradient norms. Applies to standard federated optimization and variants with server momentum.

Result: Shows that larger local steps or momentum accelerate gradient norm convergence but worsen model stability, ultimately yielding better excess risk. Experimental results validate theoretical insights.

Conclusion: Libra framework provides insights for hyperparameter tuning and future algorithm design to achieve stronger generalization in FL settings.

Abstract: Federated Learning (FL) is a distributed learning approach that trains machine learning models across multiple devices while keeping their local data private. However, FL often faces challenges due to data heterogeneity, leading to inconsistent local optima among clients. These inconsistencies can cause unfavorable convergence behavior and generalization performance degradation. Existing studies often describe this issue through \textit{convergence analysis} on gradient norms, focusing on how well a model fits training data, or through \textit{algorithmic stability}, which examines the generalization gap. However, neither approach precisely captures the generalization performance of FL algorithms, especially for non-convex neural network training. In response, this paper introduces an innovative generalization dynamics analysis framework, namely \textit{Libra}, for algorithm-dependent excess risk minimization, highlighting the trade-offs between model stability and gradient norms. We present Libra towards a standard federated optimization framework and its variants using server momentum. Through this framework, we show that larger local steps or momentum accelerate convergence of gradient norms, while worsening model stability, yielding better excess risk. Experimental results on standard FL settings prove the insights of our theories. These insights can guide hyperparameter tuning and future algorithm design to achieve stronger generalization.

[888] A Survey and Benchmarking of Spatial-Temporal Traffic Data Imputation Models

Shengnan Guo, Tonglong Wei, Yiheng Huang, Yan Lin, Zekai Shen, Yujuan Dong, Junliang Lin, Youfang Lin, Huaiyu Wan

Main category: cs.LG

TL;DR: This paper addresses key gaps in traffic data imputation by proposing taxonomies for missing patterns and models, creating a unified benchmarking pipeline, and comprehensively evaluating 11 models across multiple dimensions to provide practical guidelines for intelligent transportation systems.

DetailsMotivation: Three main gaps in traffic data imputation: 1) absence of model taxonomy to trace technological development, 2) lack of unified benchmarking pipeline for fair evaluation, and 3) insufficient multi-dimensional analysis of models including effectiveness, efficiency and robustness.

Method: Proposes practice-oriented taxonomies for traffic data missing patterns and imputation models, and introduces a unified benchmarking pipeline to evaluate 11 representative models across various missing patterns and rates.

Result: Comprehensive evaluation of models assessing overall performance, performance under challenging scenarios, computational efficiency, and providing visualizations.

Conclusion: Provides a holistic perspective on traffic data imputation and serves as a practical guideline for model selection and application in intelligent transportation systems.

Abstract: Traffic data imputation is a critical preprocessing step in intelligent transportation systems, underpinning the reliability of downstream transportation services. Despite substantial progress in imputation models, model selection and development for practical applications remains challenging due to three key gaps: 1) the absence of a model taxonomy for traffic data imputation to trace the technological development and highlight their distinct features. 2) the lack of unified benchmarking pipeline for fair and reproducible model evaluation across standardized traffic datasets. 3) insufficient in-depth analysis that jointly compare models across multiple dimensions, including effectiveness, computational efficiency and robustness. To this end, this paper proposes practice-oriented taxonomies for traffic data missing patterns and imputation models, systematically cataloging real-world traffic data loss scenarios and analyzing the characteristics of existing models. We further introduce a unified benchmarking pipeline to comprehensively evaluate 11 representative models across various missing patterns and rates, assessing overall performance, performance under challenging scenarios, computational efficiency, and providing visualizations. This work aims to provide a holistic perspective on traffic data imputation and to serve as a practical guideline for model selection and application in intelligent transportation systems.

[889] CEReBrO: Compact Encoder for Representations of Brain Oscillations Using Efficient Alternating Attention

Alexandru Dimofte, Glenn Anta Bucagu, Thorir Mar Ingolfsson, Xiaying Wang, Andrea Cossettini, Luca Benini, Yawei Li

Main category: cs.LG

TL;DR: CEReBrO is a compact EEG foundation model that uses alternating attention for efficient brain signal modeling, achieving state-of-the-art performance with significantly fewer parameters than existing methods.

DetailsMotivation: Address limitations of current EEG self-supervised learning methods: sub-optimal signal modeling, large model sizes (hundreds of millions of parameters), and reliance on private datasets/inconsistent benchmarks that hinder reproducibility.

Method: Proposes CEReBrO with per-channel patch tokenization and alternating attention mechanism that jointly models intra-channel temporal dynamics and inter-channel spatial correlations, achieving 2x speed improvement with 6x less memory than standard self-attention.

Result: Models ranging from 3.6M to 85M parameters pre-trained on 20,000+ hours of public EEG data set new benchmarks in emotion and seizure detection, with competitive performance in anomaly classification and gait prediction.

Conclusion: CEReBrO validates effectiveness and efficiency of compact EEG foundation models, addressing key limitations in current EEG self-supervised learning approaches.

Abstract: Electroencephalograph (EEG) is a crucial tool for studying brain activity. Recently, self-supervised learning methods leveraging large unlabeled datasets have emerged as a potential solution to the scarcity of widely available annotated EEG data. However, current methods suffer from at least one of the following limitations: i) sub-optimal EEG signal modeling, ii) model sizes in the hundreds of millions of trainable parameters, and iii) reliance on private datasets and/or inconsistent public benchmarks, hindering reproducibility. To address these challenges, we introduce a Compact Encoder for Representations of Brain Oscillations using alternating attention (CEReBrO), a new small EEG foundation model. Our tokenization scheme represents EEG signals at a per-channel patch granularity. We propose an alternating attention mechanism that jointly models intra-channel temporal dynamics and inter-channel spatial correlations, achieving 2x speed improvement with 6x less memory required compared to standard self-attention. We present several model sizes ranging from 3.6 million to 85 million parameters. Pre-trained on over 20,000 hours of publicly available scalp EEG recordings with diverse channel configurations, our models set new benchmarks in emotion detection and seizure detection tasks, with competitive performance in anomaly classification and gait prediction. This validates our models’ effectiveness and efficiency.

[890] Why and How Auxiliary Tasks Improve JEPA Representations

Jiacan Yu, Siyi Chen, Mingrui Liu, Nono Horiuchi, Vladimir Braverman, Zicheng Xu, Dan Haramati, Randall Balestriero

Main category: cs.LG

TL;DR: The paper provides theoretical analysis of JEPA with auxiliary regression, proving it prevents representation collapse and anchors meaningful distinctions in latent space.

DetailsMotivation: JEPA is widely used but poorly understood, particularly regarding representation collapse and how auxiliary tasks affect learned representations.

Method: Theoretical analysis of JEPA with joint auxiliary regression head, proving theorems about representation preservation in deterministic MDPs.

Result: Proved No Unhealthy Representation Collapse theorem: non-equivalent observations map to distinct representations when both losses are minimized. Experiments in counting environment confirm theory.

Conclusion: Joint training with auxiliary functions that encode proper equivalence relations can improve JEPA encoders and prevent representation collapse.

Abstract: Joint-Embedding Predictive Architecture (JEPA) is increasingly used for visual representation learning and as a component in model-based RL, but its behavior remains poorly understood. We provide a theoretical characterization of a simple, practical JEPA variant that has an auxiliary regression head trained jointly with latent dynamics. We prove a No Unhealthy Representation Collapse theorem: in deterministic MDPs, if training drives both the latent-transition consistency loss and the auxiliary regression loss to zero, then any pair of non-equivalent observations, i.e., those that do not have the same transition dynamics or auxiliary value, must map to distinct latent representations. Thus, the auxiliary task anchors which distinctions the representation must preserve. Controlled ablations in a counting environment corroborate the theory and show that training the JEPA model jointly with the auxiliary head generates a richer representation than training them separately. Our work indicates a path to improve JEPA encoders: training them with an auxiliary function that, together with the transition dynamics, encodes the right equivalence relations.

[891] KL-Regularized RLHF with Multiple Reference Models: Exact Solutions and Sample Complexity

Gholamali Aminian, Amir R. Asadi, Idan Shenfeld, Youssef Mroueh

Main category: cs.LG

TL;DR: First exact solution for multiple reference models in RLHF, addressing limitations of single-reference approaches with theoretical guarantees and sample complexity analysis.

DetailsMotivation: Single reference models in LLM alignment limit diversity, cause overfitting, and underutilize available pre-trained models. Multiple reference models can broaden perspectives, reduce bias, and leverage diverse LLM strengths.

Method: Introduces theoretical framework for multiple reference models in reverse KL-regularized RLHF, providing exact solution with rigorous statistical analysis and sample complexity guarantees. Also extends analysis to forward KL-regularized RLHF.

Result: Presents first exact solution to multiple reference model problem in RLHF, with comprehensive theoretical framework and sample complexity guarantees for both reverse and forward KL regularization.

Conclusion: Lays foundation for more advanced LLM alignment techniques using multiple reference models, enabling theoretically sound frameworks better suited to modern AI ecosystem challenges.

Abstract: Recent methods for aligning large language models (LLMs) with human feedback predominantly rely on a single reference model, which limits diversity, model overfitting, and underutilizes the wide range of available pre-trained models. Incorporating multiple reference models has the potential to address these limitations by broadening perspectives, reducing bias, and leveraging the strengths of diverse open-source LLMs. However, integrating multiple reference models into reinforcement learning with human feedback (RLHF) frameworks poses significant theoretical challenges, where achieving exact solutions has remained an open problem. This paper presents the first \emph{exact solution} to the multiple reference model problem in reverse KL-regularized RLHF. We introduce a comprehensive theoretical framework that includes rigorous statistical analysis and provides sample complexity guarantees. Additionally, we extend our analysis to forward KL-regularized RLHF, offering new insights into sample complexity requirements in multiple reference scenarios. Our contributions lay the foundation for more advanced and adaptable LLM alignment techniques, enabling the effective use of multiple reference models. This work paves the way for developing alignment frameworks that are both theoretically sound and better suited to the challenges of modern AI ecosystems.

[892] Communications to Circulations: Real-Time 3D Wind Field Prediction Using 5G GNSS Signals and Deep Learning

Yuchen Ye, Chaoxia Yuan, Mingyu Li, Aoqi Zhou, Hong Liang, Chunqing Shang, Kezuan Wang, Yifeng Zheng, Cong Chen

Main category: cs.LG

TL;DR: G-WindCast is a deep learning framework that uses 5G GNSS signal strength variations to forecast 3D atmospheric wind fields with promising accuracy up to 30 minutes lead time, outperforming traditional NWP models.

DetailsMotivation: Obtaining high spatiotemporal resolution wind data is challenging due to limitations in traditional observations and computational expense of NWP models. There's a need for cost-effective, real-time wind forecasting solutions.

Method: Uses Forward Neural Networks and Transformer networks to capture complex spatiotemporal relationships between GNSS-derived features and wind dynamics from 5G signal strength variations.

Result: Demonstrates promising accuracy in real-time wind forecasts, superior agreement with ground-based radar wind profiler compared to ECMWF ERA5, and maintains excellent performance with reduced GNSS stations (around 100).

Conclusion: This interdisciplinary approach shows transformative potential for exploiting non-traditional data sources and deep learning in environmental monitoring and real-time atmospheric applications.

Abstract: Accurate atmospheric wind field information is crucial for various applications, including weather forecasting, aviation safety, and disaster risk reduction. However, obtaining high spatiotemporal resolution wind data remains challenging due to limitations in traditional in-situ observations and remote sensing techniques, as well as the computational expense and biases of numerical weather prediction (NWP) models. This paper introduces G-WindCast, a novel deep learning framework that leverages signal strength variations from 5G Global Navigation Satellite System (GNSS) signals to forecast three-dimensional (3D) atmospheric wind fields. The framework utilizes Forward Neural Networks (FNN) and Transformer networks to capture complex, nonlinear, and spatiotemporal relationships between GNSS-derived features and wind dynamics. Our preliminary results demonstrate promising accuracy in real-time wind forecasts (up to 30 minutes lead time). The model exhibits robustness across forecast horizons and different pressure levels, and its predictions for wind fields show superior agreement with ground-based radar wind profiler compared to concurrent European Centre for Medium-Range Weather Forecasts (ECMWF) Reanalysis v5 (ERA5). Furthermore, we show that the system can maintain excellent performance for localized forecasting even with a significantly reduced number of GNSS stations (e.g., around 100), highlighting its cost-effectiveness and scalability. This interdisciplinary approach underscores the transformative potential of exploiting non-traditional data sources and deep learning for advanced environmental monitoring and real-time atmospheric applications.

[893] Boosting Graph Robustness Against Backdoor Attacks: An Over-Similarity Perspective

Chang Liu, Hai Huang, Yujie Xing, Xingquan Zuo

Main category: cs.LG

TL;DR: SimGuard is a novel defense method against graph backdoor attacks that uses similarity-based detection and contrastive learning to separate triggers from clean nodes, effectively defending against various attacks while maintaining clean node performance.

DetailsMotivation: GNNs are vulnerable to backdoor attacks in real-world applications, and existing defense methods struggle to clearly distinguish triggers from clean nodes or fully eliminate trigger impact, making it difficult to restore nodes to their pre-attack state.

Method: Uses similarity-based metric to detect triggers and employs contrastive learning to train a backdoor detector that generates embeddings capable of separating triggers from clean nodes, improving detection efficiency.

Result: Extensive experiments on real-world datasets demonstrate that SimGuard effectively defends against various graph backdoor attacks while preserving performance on clean nodes.

Conclusion: SimGuard provides an effective defense against graph backdoor attacks by leveraging the observation that triggers exhibit over-similarity in features and structure, enabling clear separation from clean nodes through similarity-based detection and contrastive learning.

Abstract: Graph Neural Networks (GNNs) have achieved notable success in tasks such as social and transportation networks. However, recent studies have highlighted the vulnerability of GNNs to backdoor attacks, raising significant concerns about their reliability in real-world applications. Despite initial efforts to defend against specific graph backdoor attacks, existing defense methods face two main challenges: either the inability to establish a clear distinction between triggers and clean nodes, resulting in the removal of many clean nodes, or the failure to eliminate the impact of triggers, making it challenging to restore the target nodes to their pre-attack state. Through empirical analysis of various existing graph backdoor attacks, we observe that the triggers generated by these methods exhibit over-similarity in both features and structure. Based on this observation, we propose a novel graph backdoor defense method SimGuard. We first utilizes a similarity-based metric to detect triggers and then employs contrastive learning to train a backdoor detector that generates embeddings capable of separating triggers from clean nodes, thereby improving detection efficiency. Extensive experiments conducted on real-world datasets demonstrate that our proposed method effectively defends against various graph backdoor attacks while preserving performance on clean nodes. The code will be released upon acceptance.

[894] Robust LLM Training Infrastructure at ByteDance

Borui Wan, Gaohong Liu, Zuquan Song, Jun Wang, Yun Zhang, Guangming Sheng, Shuguang Wang, Houmin Wei, Chenyuan Wang, Weiqiang Lou, Xi Yang, Mofan Zhang, Kaihua Jiang, Cheng Ren, Xiaoyun Zhi, Menghan Yu, Zhe Nan, Zhuolin Zheng, Baoquan Zhong, Qinlong Wang, Huan Yu, Jinxin Chi, Wang Zhang, Yuhan Li, Zixian Du, Sida Zhao, Yongqiang Zhang, Jingzhe Tang, Zherui Liu, Chuan Wu, Yanghua Peng, Haibin Lin, Wencong Xiao, Xin Liu, Liang Xiang

Main category: cs.LG

TL;DR: ByteRobust is a GPU infrastructure management system designed to ensure robust and stable large-scale LLM training by enabling efficient fault detection, diagnosis, and recovery with minimal interruptions.

DetailsMotivation: Large-scale LLM training faces frequent failures (CUDA errors, NaN values, job hangs) that challenge training stability and efficiency as training scales to tens of thousands of GPUs.

Method: ByteRobust exploits LLM training uniqueness, uses parallelisms and characteristics of LLM training for fault tolerance, and employs data-driven approaches for prompt fault demarcation and localization.

Result: Deployed on production GPU platform, ByteRobust achieved 97% ETTR (Estimated Time to Recovery) for a three-month training job on 9,600 GPUs.

Conclusion: ByteRobust comprehensively ensures continuous and efficient training of LLM tasks through high-capacity fault tolerance and effective failure management.

Abstract: The training scale of large language models (LLMs) has reached tens of thousands of GPUs and is still continuously expanding, enabling faster learning of larger models. Accompanying the expansion of the resource scale is the prevalence of failures (CUDA error, NaN values, job hang, etc.), which poses significant challenges to training stability. Any large-scale LLM training infrastructure should strive for minimal training interruption, efficient fault diagnosis, and effective failure tolerance to enable highly efficient continuous training. This paper presents ByteRobust, a large-scale GPU infrastructure management system tailored for robust and stable training of LLMs. It exploits the uniqueness of LLM training process and gives top priorities to detecting and recovering failures in a routine manner. Leveraging parallelisms and characteristics of LLM training, ByteRobust enables high-capacity fault tolerance, prompt fault demarcation, and localization with an effective data-driven approach, comprehensively ensuring continuous and efficient training of LLM tasks. ByteRobust is deployed on a production GPU platform and achieves 97% ETTR for a three-month training job on 9,600 GPUs.

[895] Membership Inference Attack Should Move On to Distributional Statistics for Distilled Generative Models

Muxing Li, Zesheng Ye, Sharon Li, Andy Song, Guangquan Zhang, Feng Liu

Main category: cs.LG

TL;DR: Standard membership inference attacks fail against distilled generative models because they lack direct exposure to training data, but distribution-level statistics can detect unauthorized data usage through alignment between student-generated data and teacher’s training distribution.

DetailsMotivation: To address the privacy loophole where distilled models can evade standard membership inference attacks despite potentially being trained on unauthorized data through teacher models, enabling detection of upstream privacy violations.

Method: Propose shifting from instance-level scores to distribution-level statistics for MIAs, with three principles for distribution-based attacks and an exemplar framework to detect unauthorized training data through alignment between student-generated data and teacher’s training distribution.

Result: Distribution-based MIAs can successfully detect unauthorized data usage in distilled generative models even when direct instance-level memorization is absent, by leveraging the memory chain connecting student and teacher’s member data.

Conclusion: Distilled generative models are auditable for upstream privacy violations through distribution-level membership inference attacks, and should not be discarded when privacy is a concern, as the memory chain between student and teacher models enables detection of unauthorized data usage.

Abstract: To detect unauthorized data usage in training large-scale generative models (e.g., ChatGPT or Midjourney), membership inference attacks (MIA) have proven effective in distinguishing a single training instance (a member) from a single non-training instance (a non-member). This success is mainly credited to a memorization effect: models tend to perform better on a member than a non-member. However, we find that standard MIAs fail against distilled generative models (i.e., student models) that are increasingly deployed in practice for efficiency (e.g., ChatGPT 4o-mini). Trained exclusively on data generated from a large-scale model (a teacher model), the student model lacks direct exposure to any members (teacher’s training data), nullifying the memorization effect that standard MIAs rely on. This finding reveals a serious privacy loophole, where generation-service providers could deploy a student model whose teacher was potentially trained on unauthorized data, yet claim the deployed model is clean because it was not directly trained on such data. Hence, are distilled models inherently unauditable for upstream privacy violations, and should we discard them when we care about privacy? We contend no, as we uncover a memory chain connecting the student and teacher’s member data: the distribution of student-generated data aligns more closely with the distribution of the teacher’s members than with non-members, thus we can detect unauthorized data usage even when direct instance-level memorization is absent. This leads us to posit that MIAs on distilled generative models should shift from instance-level scores to distribution-level statistics. We further propose three principles of distribution-based MIAs for detecting unauthorized training data through distilled generative models, and validate our position through an exemplar framework. We lastly discuss the implications of our position.

[896] Graph Coloring for Multi-Task Learning

Santosh Patapati

Main category: cs.LG

TL;DR: SON-GOKU is a scheduler that uses gradient interference analysis and graph coloring to partition tasks into compatible groups, activating only one group per training step to improve multi-task learning by reducing conflicting gradient directions.

DetailsMotivation: Address gradient interference in multi-task learning where conflicting objectives slow convergence and reduce final model performance.

Method: Computes gradient interference, constructs interference graph, applies greedy graph-coloring to partition tasks into compatible groups, and activates only one group per training step with dynamic recomputation of groupings.

Result: Empirical results on six datasets show consistent outperformance over baselines and state-of-the-art multi-task optimizers.

Conclusion: Grouping and sequential updates improve multi-task learning with theoretical guarantees on descent, convergence, and accurate identification of task conflicts/alignments.

Abstract: When different objectives conflict with each other in multi-task learning, gradients begin to interfere and slow convergence, thereby potentially reducing the final model’s performance. To address this, we introduce SON-GOKU, a scheduler that computes gradient interference, constructs an interference graph, and then applies greedy graph-coloring to partition tasks into groups that align well with each other. At each training step, only one group (color class) of tasks are activated, and the grouping partition is constantly recomputed as task relationships evolve throughout training. By ensuring that each mini-batch contains only tasks that pull the model in the same direction, our method improves the effectiveness of any underlying multi-task learning optimizer without additional tuning. Since tasks within these groups will update in compatible directions, multi-task learning will improve model performance rather than impede it. Empirical results on six different datasets show that this interference-aware graph-coloring approach consistently outperforms baselines and state-of-the-art multi-task optimizers. We provide extensive theory showing why grouping and sequential updates improve multi-task learning, with guarantees on descent, convergence, and accurately identifying what tasks conflict or align.

[897] Rao-Blackwell Gradient Estimators for Equivariant Denoising Diffusion

Vinh Tong, Hoang Trung-Dung, Anji Liu, Guy Van den Broeck, Mathias Niepert

Main category: cs.LG

TL;DR: Orbit Diffusion is a framework that enhances both equivariant architectures and data augmentation approaches by providing a lower-variance gradient estimator through Rao-Blackwellization, leading to improved performance in molecular, crystal, and protein generation tasks.

DetailsMotivation: Existing methods for learning invariant distributions in domains like molecular and protein generation face challenges: equivariant architectures are complex and hard to optimize, while data augmentation may not fully capture symmetries. The authors aim to address these limitations by reducing training variance.

Method: The framework interprets data augmentation as a Monte Carlo estimator of training gradients and applies Rao-Blackwellization to create a lower-variance gradient estimator. This requires only a single forward and backward pass per sample. The implementation is called Orbit Diffusion.

Result: Orbit Diffusion achieves state-of-the-art results on GEOM-QM9 for molecular conformation generation, improves crystal structure prediction on Perov-5 and MP-20 benchmarks, and enhances protein designability in protein structure generation.

Conclusion: The proposed framework provides a theoretically guaranteed lower-variance gradient estimator that admits equivariant minimizers, leading to more stable optimization, faster convergence, and improved performance across multiple domains while maintaining computational efficiency.

Abstract: In domains such as molecular and protein generation, physical systems exhibit inherent symmetries that are critical to model. Two main strategies have emerged for learning invariant distributions: designing equivariant network architectures and using data augmentation to approximate equivariance. While equivariant architectures preserve symmetry by design, they often involve greater complexity and pose optimization challenges. Data augmentation, on the other hand, offers flexibility but may fall short in fully capturing symmetries. Our framework enhances both approaches by reducing training variance and providing a provably lower-variance gradient estimator. We achieve this by interpreting data augmentation as a Monte Carlo estimator of the training gradient and applying Rao-Blackwellization. This leads to more stable optimization, faster convergence, and reduced variance, all while requiring only a single forward and backward pass per sample. We also present a practical implementation of this estimator incorporating the loss and sampling procedure through a method we call Orbit Diffusion. Theoretically, we guarantee that our loss admits equivariant minimizers. Empirically, Orbit Diffusion achieves state-of-the-art results on GEOM-QM9 for molecular conformation generation, improves crystal structure prediction, and advances text-guided crystal generation on the Perov-5 and MP-20 benchmarks. Additionally, it enhances protein designability in protein structure generation. Code is available at: https://github.com/vinhsuhi/Orbit-Diffusion.git.

[898] Bayesian Computation in Deep Learning

Wenlong Chen, Bolian Li, Ruqi Zhang, Yingzhen Li

Main category: cs.LG

TL;DR: This chapter introduces approximate Bayesian inference techniques for deep learning models, focusing on Bayesian neural networks and deep generative models, and reviews SG-MCMC and VI methods.

DetailsMotivation: Bayesian methods improve reliability and uncertainty awareness in deep neural networks for predictive tasks and enable inference of latent variables in deep generative models.

Method: The chapter reviews two main approximate Bayesian computational methods: stochastic gradient Markov chain Monte Carlo (SG-MCMC) and variational inference (VI), discussing their challenges and solutions in deep learning contexts.

Result: The paper provides an overview of how SG-MCMC and VI can be effectively applied to Bayesian neural networks and deep generative models for posterior inference.

Conclusion: Approximate Bayesian inference techniques like SG-MCMC and VI are crucial for implementing Bayesian reasoning in deep learning models, enhancing their reliability and enabling effective training of complex generative models.

Abstract: Bayesian methods have shown success in deep learning applications. For example, in predictive tasks, Bayesian neural networks leverage Bayesian reasoning of model uncertainty to improve the reliability and uncertainty awareness of deep neural networks. In generative modeling domain, many widely used deep generative models, such as deep latent variable models, require approximate Bayesian inference to infer their latent variables for the training. In this chapter, we provide an introduction to approximate inference techniques as Bayesian computation methods applied to deep learning models, with a focus on Bayesian neural networks and deep generative models. We review two arguably most popular approximate Bayesian computational methods, stochastic gradient Markov chain Monte Carlo (SG-MCMC) and variational inference (VI), and explain their unique challenges in posterior inference as well as the solutions when applied to deep learning models.

[899] Diffusion Models are Kelly Gamblers

Akhil Premkumar

Main category: cs.LG

TL;DR: This paper connects diffusion models to the Kelly criterion for betting games, showing that conditional diffusion models store mutual information between signals and conditioning information, and that classifier-free guidance boosts this mutual information during sampling.

DetailsMotivation: To establish connections between diffusion models and information theory concepts, specifically exploring how conditional diffusion models handle mutual information between signals and conditioning variables, and to clarify misconceptions about diffusion models as infinitely deep autoencoders.

Method: Theoretical analysis connecting diffusion models to the Kelly criterion, examining how conditional diffusion models store mutual information between X and Y, and analyzing classifier-free guidance as a mechanism to boost mutual information during sampling.

Result: Found that conditional diffusion models store additional information equal to the mutual information between signal X and conditioning Y. Classifier-free guidance effectively increases this mutual information at sampling time, which is particularly beneficial for image models where image-label mutual information is low due to the manifold hypothesis.

Conclusion: The paper provides new theoretical insights connecting diffusion models to information theory and quantum mechanics, showing that classifier-free guidance boosts mutual information and clarifying the relationship between denoising loss and the Fermi Golden Rule from quantum mechanics.

Abstract: We draw a connection between diffusion models and the Kelly criterion for maximizing returns in betting games. We find that conditional diffusion models store additional information to bind the signal $X$ with the conditioning information $Y$, equal to the mutual information between them. Classifier-free guidance effectively boosts the mutual information between $X$ and $Y$ at sampling time. This is especially helpful in image models, since the mutual information between images and their labels is low, a fact which is intimately connected to the manifold hypothesis. Finally, we point out some nuances in the popular perspective that diffusion models are infinitely deep autoencoders. In doing so, we relate the denoising loss to the Fermi Golden Rule from quantum mechanics.

[900] ImpMIA: Leveraging Implicit Bias for Membership Inference Attack under Realistic Scenarios

Yuval Golbari, Navve Wasserman, Gal Vardi, Michal Irani

Main category: cs.LG

TL;DR: ImpMIA is a white-box membership inference attack that exploits neural networks’ implicit bias using KKT conditions to identify training samples without needing reference models or unrealistic assumptions.

DetailsMotivation: Existing black-box MIA methods rely on unrealistic assumptions about hyperparameter knowledge, data distribution, and training data fraction. Removing these assumptions significantly degrades their performance.

Method: Uses KKT optimality conditions from maximum-margin implicit bias theory to find samples whose gradients best reconstruct the trained model’s parameters. This is a white-box approach requiring model weights.

Result: ImpMIA achieves state-of-the-art performance compared to both black-box and white-box attacks in realistic settings where only model weights and a superset of training data are available.

Conclusion: The proposed white-box approach effectively addresses limitations of black-box MIA methods by leveraging implicit bias theory, making membership inference more practical for real-world scenarios.

Abstract: Determining which data samples were used to train a model-known as Membership Inference Attack (MIA)-is a well-studied and important problem with implications for data privacy. Black-box methods presume access only to the model’s outputs and often rely on training auxiliary reference models. While they have shown strong empirical performance, they rely on assumptions that rarely hold in real-world settings: (i) the attacker knows the training hyperparameters; (ii) all available non-training samples come from the same distribution as the training data; and (iii) the fraction of training data in the evaluation set is known. In this paper, we demonstrate that removing these assumptions leads to a significant drop in the performance of black-box attacks. We introduce ImpMIA, a Membership Inference Attack that exploits the Implicit Bias of neural networks, hence removes the need to rely on any reference models and their assumptions. ImpMIA is a white-box attack – a setting which assumes access to model weights and is becoming increasingly realistic given that many models are publicly available (e.g., via Hugging Face). Building on maximum-margin implicit bias theory, ImpMIA uses the Karush-Kuhn-Tucker (KKT) optimality conditions to identify training samples. This is done by finding the samples whose gradients most strongly reconstruct the trained model’s parameters. As a result, ImpMIA achieves state-of-the-art performance compared to both black and white box attacks in realistic settings where only the model weights and a superset of the training data are available.

[901] Weak-to-Strong Generalization Even in Random Feature Networks, Provably

Marko Medvedev, Kaifeng Lyu, Dingli Yu, Sanjeev Arora, Zhiyuan Li, Nathan Srebro

Main category: cs.LG

TL;DR: Weak-to-strong generalization occurs when a strong student model outperforms a weak teacher model, even when trained only on the teacher’s labels. This phenomenon is demonstrated and analyzed using random feature models.

DetailsMotivation: To understand and prove that weak-to-strong generalization doesn't require powerful models like GPT-4, and can occur in simpler settings like random feature models.

Method: Use random feature models (two-layer networks with fixed random bottom layer and trained top layer) where a weak teacher with few units is trained on population data, and a strong student with many units is trained only on teacher-generated labels.

Result: The student significantly outperforms the teacher despite being trained only on teacher-labeled data. Early stopping enables this weak-to-strong generalization.

Conclusion: Weak-to-strong generalization is a general phenomenon that occurs in simpler models, enabled by early stopping, but has quantitative limits in this framework.

Abstract: Weak-to-Strong Generalization (Burns et al., 2024) is the phenomenon whereby a strong student, say GPT-4, learns a task from a weak teacher, say GPT-2, and ends up significantly outperforming the teacher. We show that this phenomenon does not require a strong learner like GPT-4. We consider student and teacher that are random feature models, described by two-layer networks with a random and fixed bottom layer and a trained top layer. A “weak” teacher, with a small number of units (i.e. random features), is trained on the population, and a “strong” student, with a much larger number of units (i.e. random features), is trained only on labels generated by the weak teacher. We demonstrate, prove, and understand how the student can outperform the teacher, even though trained only on data labeled by the teacher. We also explain how such weak-to-strong generalization is enabled by early stopping. Importantly, we also show the quantitative limits of weak-to-strong generalization in this model.

[902] From Equations to Insights: Unraveling Symbolic Structures in PDEs with LLMs

Rohan Bhatnagar, Ling Liang, Krish Patel, Haizhao Yang

Main category: cs.LG

TL;DR: LLMs can predict PDE operators from symbolic information, improving efficiency and accuracy of symbolic machine learning for analytical approximations.

DetailsMotivation: While AI has been applied to solve PDEs, discovering symbolic relationships within these equations remains unexplored. This paper aims to leverage LLMs to learn such symbolic relationships.

Method: Proposed using large language models (LLMs) to predict operators involved in PDE solutions by utilizing symbolic information in the PDEs, both theoretically and numerically.

Result: LLMs can effectively predict PDE operators, and discovering these symbolic relationships substantially improves efficiency and accuracy of symbolic machine learning for finding analytical approximations.

Conclusion: This work opens new avenues for understanding the symbolic structure of scientific problems and advancing their solution processes, delivering a fully interpretable solution pipeline.

Abstract: Motivated by the remarkable success of artificial intelligence (AI) across diverse fields, the application of AI to solve scientific problems, often formulated as partial differential equations (PDEs), has garnered increasing attention. While most existing research concentrates on theoretical properties (such as well-posedness, regularity, and continuity) of the solutions, alongside direct AI-driven methods for solving PDEs, the challenge of uncovering symbolic relationships within these equations remains largely unexplored. In this paper, we propose leveraging large language models (LLMs) to learn such symbolic relationships. Our results demonstrate that LLMs can effectively predict the operators involved in PDE solutions by utilizing the symbolic information in the PDEs both theoretically and numerically. Furthermore, we show that discovering these symbolic relationships can substantially improve both the efficiency and accuracy of symbolic machine learning for finding analytical approximation of PDE solutions, delivering a fully interpretable solution pipeline. This work opens new avenues for understanding the symbolic structure of scientific problems and advancing their solution processes.

[903] TimeEmb: A Lightweight Static-Dynamic Disentanglement Framework for Time Series Forecasting

Mingyuan Xia, Chunxu Zhang, Zijian Zhang, Hao Miao, Qidong Liu, Yuanshao Zhu, Bo Yang

Main category: cs.LG

TL;DR: TimeEmb is a lightweight framework that decomposes time series into time-invariant and time-varying components to handle temporal non-stationarity in forecasting.

DetailsMotivation: Temporal non-stationarity (distribution shifts over time) challenges reliable forecasting. Existing methods conflate time-varying and time-invariant components, leading to suboptimal performance.

Method: Separates time series into: (1) time-invariant component via global embedding module for persistent representations, (2) time-varying component via frequency-domain filtering inspired by full-spectrum analysis.

Result: Outperforms state-of-the-art baselines on real-world datasets with fewer computational resources. Comprehensive analyses verify efficacy of static-dynamic disentanglement.

Conclusion: Lightweight framework effectively handles temporal non-stationarity and can be easily integrated to improve existing forecasting methods.

Abstract: Temporal non-stationarity, the phenomenon that time series distributions change over time, poses fundamental challenges to reliable time series forecasting. Intuitively, the complex time series can be decomposed into two factors, \ie time-invariant and time-varying components, which indicate static and dynamic patterns, respectively. Nonetheless, existing methods often conflate the time-varying and time-invariant components, and jointly learn the combined long-term patterns and short-term fluctuations, leading to suboptimal performance facing distribution shifts. To address this issue, we initiatively propose a lightweight static-dynamic decomposition framework, TimeEmb, for time series forecasting. TimeEmb innovatively separates time series into two complementary components: (1) time-invariant component, captured by a novel global embedding module that learns persistent representations across time series, and (2) time-varying component, processed by an efficient frequency-domain filtering mechanism inspired by full-spectrum analysis in signal processing. Experiments on real-world datasets demonstrate that TimeEmb outperforms state-of-the-art baselines and requires fewer computational resources. We conduct comprehensive quantitative and qualitative analyses to verify the efficacy of static-dynamic disentanglement. This lightweight framework can also improve existing time-series forecasting methods with simple integration. To ease reproducibility, the code is available at https://github.com/showmeon/TimeEmb.

[904] Physics-Informed Deep B-Spline Networks

Zhuoyuan Wang, Raffaele Romagnoli, Saviz Mowlavi, Yorie Nakahira

Main category: cs.LG

TL;DR: Physics-informed deep B-spline networks learn PDE families with varying parameters and ICBCs by approximating solutions through B-spline control points, providing theoretical guarantees and improved efficiency-accuracy tradeoffs.

DetailsMotivation: Existing physics-informed machine learning methods lack theoretical guarantees for learning PDEs with varying parameters and changing initial/boundary conditions, creating a need for more robust approaches.

Method: Propose physics-informed deep B-spline networks that learn B-spline control points through neural networks, reducing the learning task to predicting a compact set of control points while enforcing strict compliance to initial and Dirichlet boundary conditions by construction.

Result: The method achieves improved efficiency-accuracy tradeoffs compared to existing techniques, handles discontinuous ICBCs, nonhomogeneous ICBCs, and non-rectangular domains, with established theoretical guarantees including universal approximation and generalization error bounds.

Conclusion: Physics-informed deep B-spline networks provide a theoretically-grounded framework for learning PDE families with varying parameters and ICBCs, offering practical advantages in handling complex boundary conditions and domains.

Abstract: Physics-informed machine learning offers a promising framework for solving complex partial differential equations (PDEs) by integrating observational data with governing physical laws. However, learning PDEs with varying parameters and changing initial conditions and boundary conditions (ICBCs) with theoretical guarantees remains an open challenge. In this paper, we propose physics-informed deep B-spline networks, a novel technique that approximates a family of PDEs with different parameters and ICBCs by learning B-spline control points through neural networks. The proposed B-spline representation reduces the learning task from predicting solution values over the entire domain to learning a compact set of control points, enforces strict compliance to initial and Dirichlet boundary conditions by construction, and enables analytical computation of derivatives for incorporating PDE residual losses. While existing approximation and generalization theories are not applicable in this setting - where solutions of parametrized PDE families are represented via B-spline bases - we fill this gap by showing that B-spline networks are universal approximators for such families under mild conditions. We also derive generalization error bounds for physics-informed learning in both elliptic and parabolic PDE settings, establishing new theoretical guarantees. Finally, we demonstrate in experiments that the proposed technique has improved efficiency-accuracy tradeoffs compared to existing techniques in a dynamical system problem with discontinuous ICBCs and can handle nonhomogeneous ICBCs and non-rectangular domains.

[905] Market-Driven Subset Selection for Budgeted Training

Ashish Jha, Valentin Leplat, AH Phan

Main category: cs.LG

TL;DR: A market-based framework for data subset selection that aggregates multiple utility signals using logarithmic market scoring rules, achieving efficient training under fixed computational budgets with theoretical guarantees.

DetailsMotivation: Current data subset selection methods combine heterogeneous utility signals through ad hoc weighted sums without theoretical grounding, leading to inefficiencies in large language model training.

Method: Treats training examples as tradeable contracts, uses Logarithmic Market Scoring Rule to aggregate utility signals, employs topic-wise normalization, and handles token budgets explicitly with price-per-token decision rules.

Result: Achieves parity with strong baselines on GSM8K mathematical reasoning under 60k-token budgets with low variance and minimal overhead (<0.1 GPU-hour), and delivers competitive accuracy with improved stability on AGNews classification at 5-25% retention rates.

Conclusion: The market-based framework successfully unifies multi-signal data curation under fixed computational budgets, providing a theoretically grounded approach for efficient training data selection.

Abstract: Training large language models on massive datasets is computationally expensive, yet empirical evidence suggests that substantial portions of training examples contribute minimally to final performance. Data subset selection addresses this inefficiency by identifying small, high-utility subsets under resource constraints. However, example utility is inherently multi-faceted, encompassing uncertainty, distributional rarity, and diversity signals that are heterogeneous and typically combined through ad hoc weighted sums lacking theoretical grounding. We propose a market-based framework that treats each training example as a tradeable contract and employs the Logarithmic Market Scoring Rule to aggregate multiple utility signals into coherent prices. Heterogeneous signals act as traders, a single liquidity parameter controls concentration versus smoothing, and topic-wise normalization ensures calibrated aggregation. Token budgets are handled explicitly through a price-per-token decision rule with an interpretable length-bias parameter. We establish theoretical connections to maximum-entropy aggregation and provide utility recovery guarantees under noisy but monotone signals. On GSM8K mathematical reasoning under strict 60k-token budgets, our selector achieves parity with strong single-signal baselines while exhibiting lower variance and incurring less than 0.1 GPU-hour overhead. On AGNews classification at 5-25% retention rates, the market formulation delivers competitive accuracy with improved stability. Our framework unifies multi-signal data curation under fixed computational budgets for prompt-level reasoning and classification tasks.

[906] LANGTRAJ: Diffusion Model and Dataset for Language-Conditioned Trajectory Simulation

Wei-Jer Chang, Wei Zhan, Masayoshi Tomizuka, Manmohan Chandraker, Francesco Pittaluga

Main category: cs.LG

TL;DR: LangTraj is a language-conditioned scene-diffusion model that simulates realistic traffic scenarios with natural language control, enabling flexible and scalable autonomous vehicle testing.

DetailsMotivation: To enable scalable testing of autonomous vehicles in counterfactual settings by providing intuitive language-based control over traffic scenarios, overcoming limitations of domain-specific guidance functions.

Method: Developed LangTraj with language conditioning during training, proposed novel closed-loop training strategy for diffusion models, and created Inter-Drive dataset with diverse interactive labels for training.

Result: LangTraj demonstrates strong performance in realism, language controllability, and safety-critical simulation on Waymo Open Motion Dataset, establishing a new paradigm for AV testing.

Conclusion: Language-conditioned simulation provides flexible and scalable approach for autonomous vehicle evaluation, with LangTraj showing promising results in realistic traffic scenario generation.

Abstract: Evaluating autonomous vehicles with controllability enables scalable testing in counterfactual or structured settings, enhancing both efficiency and safety. We introduce LangTraj, a language-conditioned scene-diffusion model that simulates the joint behavior of all agents in traffic scenarios. By conditioning on natural language inputs, LangTraj provides flexible and intuitive control over interactive behaviors, generating nuanced and realistic scenarios. Unlike prior approaches that depend on domain-specific guidance functions, LangTraj incorporates language conditioning during training, facilitating more intuitive traffic simulation control. We propose a novel closed-loop training strategy for diffusion models, explicitly tailored to enhance stability and realism during closed-loop simulation. To support language-conditioned simulation, we develop Inter-Drive, a large-scale dataset with diverse and interactive labels for training language-conditioned diffusion models. Our dataset is built upon a scalable pipeline for annotating agent-agent interactions and single-agent behaviors, ensuring rich and varied supervision. Validated on the Waymo Open Motion Dataset, LangTraj demonstrates strong performance in realism, language controllability, and language-conditioned safety-critical simulation, establishing a new paradigm for flexible and scalable autonomous vehicle testing. Project Website: https://langtraj.github.io/

[907] Score-based deterministic density sampling

Vasily Ilin, Peter Sushko, Jingwei Hu

Main category: cs.LG

TL;DR: A deterministic sampling framework using Score-Based Transport Modeling that approximates Wasserstein gradient flow for sampling unnormalized densities using score matching, producing smooth deterministic trajectories with monotone convergence.

DetailsMotivation: To develop a deterministic alternative to stochastic Langevin dynamics that maintains the same marginal distribution but produces smooth trajectories with noise-free convergence and potentially better sample efficiency.

Method: Approximates Wasserstein gradient flow on KL divergence by learning time-varying scores through score matching, creating deterministic transport maps that evolve samples along smooth trajectories.

Result: The method converges at optimal rate with smooth trajectories, is often more sample efficient than stochastic counterparts, produces high-quality image generations in as few as 15 steps, and scales linearly with sample size in memory and runtime.

Conclusion: The proposed deterministic sampling framework successfully provides smooth, noise-free convergence with optimal convergence rates and improved sample efficiency compared to stochastic methods, while maintaining theoretical guarantees and practical scalability.

Abstract: We propose a deterministic sampling framework using Score-Based Transport Modeling for sampling an unnormalized target density $\pi$ given only its score $\nabla \log \pi$. Our method approximates the Wasserstein gradient flow on $\mathrm{KL}(f_t|\pi)$ by learning the time-varying score $\nabla \log f_t$ on the fly using score matching. While having the same marginal distribution as Langevin dynamics, our method produces smooth deterministic trajectories, resulting in monotone noise-free convergence. We prove that our method dissipates relative entropy at the same rate as the exact gradient flow, provided sufficient training. Numerical experiments validate our theoretical findings: our method converges at the optimal rate, has smooth trajectories, and is often more sample efficient than its stochastic counterpart. Experiments on high-dimensional image data show that our method produces high-quality generations in as few as 15 steps and exhibits natural exploratory behavior. The memory and runtime scale linearly in the sample size.

[908] SDAR: A Synergistic Diffusion-AutoRegression Paradigm for Scalable Sequence Generation

Shuang Cheng, Yihan Bian, Dawei Liu, Linfeng Zhang, Qian Yao, Zhongbo Tian, Wenhai Wang, Qipeng Guo, Kai Chen, Biqing Qi, Bowen Zhou

Main category: cs.LG

TL;DR: SDAR is a novel paradigm that converts autoregressive models into blockwise diffusion models, enabling parallel inference while maintaining training efficiency and performance.

DetailsMotivation: To combine the training efficiency of autoregressive models with the parallel inference capability of diffusion models, avoiding costly end-to-end diffusion training.

Method: Lightweight paradigm conversion that transforms well-trained autoregressive models into blockwise diffusion models through brief, data-efficient adaptation. Uses autoregressive generation across blocks for global coherence and parallel diffusion decoding within blocks.

Result: Achieves efficient AR-to-diffusion conversion with minimal cost, preserving AR-level performance while enabling parallel generation. 30B MoE model surpasses AR counterpart on scientific reasoning benchmarks (GPQA, ChemBench) and improves further with test-time scaling methods.

Conclusion: SDAR establishes a practical paradigm that combines the strengths of autoregression and diffusion for scalable, high-throughput reasoning, with larger models showing stronger robustness and greater speedups without accuracy loss.

Abstract: We propose SDAR, a Synergistic Diffusion-Autoregression paradigm that unifies the training efficiency of autoregressive models with the parallel inference capability of diffusion. Instead of costly end-to-end diffusion training, SDAR performs a lightweight paradigm conversion that transforms a well-trained autoregressive (AR) model into a blockwise diffusion model through brief, data-efficient adaptation. During inference, SDAR generates sequences autoregressively across blocks for global coherence while decoding all tokens within each block in parallel via a discrete diffusion process. Extensive experiments show that AR models remain substantially more compute-efficient than masked diffusion models, providing a strong foundation for adaptation. Building on this insight, SDAR achieves efficient AR-to-diffusion conversion with minimal cost, preserving AR-level performance while enabling parallel generation. Scaling studies across dense and Mixture-of-Experts architectures confirm that SDAR scales without compromise: larger models exhibit stronger robustness to block size and decoding thresholds, yielding greater speedups without accuracy loss. Beyond efficiency, SDAR demonstrates enhanced reasoning and domain adaptability. Our 30B MoE model surpasses its AR counterpart on challenging scientific reasoning benchmarks such as GPQA and ChemBench, and gains further improvements under test-time scaling methods like majority voting and pass@k. Together, these results establish SDAR as a practical paradigm that combines the strengths of autoregression and diffusion for scalable, high-throughput reasoning.

[909] Challenges and proposed solutions in modeling multimodal data: A systematic review

Maryam Farhadizadeh, Maria Weymann, Michael Blaß, Johann Kraus, Christopher Gundler, Sebastian Walter, Noah Hempen, Harald Binder, Nadine Binder

Main category: cs.LG

TL;DR: Systematic review of 69 studies on multimodal data modeling in clinical research, identifying key challenges and recent methodological advances for integrating diverse medical data types.

DetailsMotivation: To address the technical challenges in modeling heterogeneous clinical data (imaging, genomics, wearables, EHRs) despite its potential to improve diagnostic accuracy and personalized care.

Method: Conducted a systematic review synthesizing findings from 69 studies to identify common obstacles and highlight recent methodological advances in multimodal data fusion.

Result: Identified key challenges including missing modalities, limited sample sizes, dimensionality imbalance, interpretability issues, and optimal fusion techniques. Found promising solutions in transfer learning, generative models, attention mechanisms, and neural architecture search.

Conclusion: The review provides a comprehensive overview of current trends and practical insights to guide future research in multimodal modeling for medical applications.

Abstract: Multimodal data modeling has emerged as a powerful approach in clinical research, enabling the integration of diverse data types such as imaging, genomics, wearable sensors, and electronic health records. Despite its potential to improve diagnostic accuracy and support personalized care, modeling such heterogeneous data presents significant technical challenges. This systematic review synthesizes findings from 69 studies to identify common obstacles, including missing modalities, limited sample sizes, dimensionality imbalance, interpretability issues, and finding the optimal fusion techniques. We highlight recent methodological advances, such as transfer learning, generative models, attention mechanisms, and neural architecture search that offer promising solutions. By mapping current trends and innovations, this review provides a comprehensive overview of the field and offers practical insights to guide future research and development in multimodal modeling for medical applications.

[910] Synthetic Series-Symbol Data Generation for Time Series Foundation Models

Wenxuan Wang, Kai Wu, Yujian Betterest Li, Dan Wang, Xiaoyu Zhang

Main category: cs.LG

TL;DR: SymTime is a foundation model for time series analysis that uses series-symbol data generation to overcome data scarcity issues, achieving competitive performance across five major TSA tasks.

DetailsMotivation: To address challenges in foundation models for time series analysis, particularly training data scarcity and imbalance, by leveraging complex dynamic system theories.

Method: Developed a series-symbol data generation mechanism to create unlimited high-quality time series data with corresponding symbolic expressions, then pre-trained SymTime model using these correlated series-symbol pairs.

Result: SymTime demonstrates competitive performance across five major time series analysis tasks when fine-tuned, rivaling foundation models pre-trained on real-world datasets.

Conclusion: The approach shows the potential of series-symbol data generation and pretraining mechanisms to overcome data scarcity and enhance task performance in time series analysis.

Abstract: Foundation models for time series analysis (TSA) have attracted significant attention. However, challenges such as training data scarcity and imbalance continue to hinder their development. Inspired by complex dynamic system theories, we design a series-symbol data generation mechanism, enabling the unrestricted creation of high-quality time series data paired with corresponding symbolic expressions. To leverage series-symbol data pairs with strong correlations, we develop SymTime, a pre-trained foundation model for enhancing time series representation using symbolic information. SymTime demonstrates competitive performance across five major TSA tasks when fine-tunes with downstream tasks, rivaling foundation models pre-trained on real-world datasets. This approach underscores the potential of series-symbol data generation and pretraining mechanisms in overcoming data scarcity and enhancing task performance. The code is available at https://github.com/wwhenxuan/SymTime.

[911] MergeBench: A Benchmark for Merging Domain-Specialized LLMs

Yifei He, Siqi Zeng, Yuzheng Hu, Rui Yang, Tong Zhang, Han Zhao

Main category: cs.LG

TL;DR: MergeBench is a comprehensive evaluation suite for model merging methods, assessing 8 techniques across 5 domains using Llama and Gemma models at 2B-9B scales, providing practical guidelines and identifying remaining challenges.

DetailsMotivation: Existing model merging evaluations are limited in model scale and task diversity, leaving questions about applicability to large, domain-specialized LLMs.

Method: Built MergeBench evaluation suite using state-of-the-art open-source LLMs (Llama, Gemma families at 2B-9B scales) covering 5 domains: instruction following, mathematics, multilingual understanding, coding, and safety. Standardized finetuning and evaluation protocols to assess 8 merging methods.

Result: Model merging performs better on stronger base models, with techniques like merging coefficient tuning and sparsification improving knowledge retention. However, challenges remain including computational costs, performance gaps compared to multi-task models, and unexplored roles in standard LLM training pipelines.

Conclusion: MergeBench provides a foundation for future research to advance understanding and practical application of model merging, with identified challenges guiding further development in the field.

Abstract: Model merging provides a scalable alternative to multi-task training by combining specialized finetuned models through parameter arithmetic, enabling efficient deployment without the need for joint training or access to all task data. While recent methods have shown promise, existing evaluations are limited in both model scale and task diversity, leaving open questions about their applicability to large, domain-specialized LLMs. To tackle the challenges, we introduce MergeBench, a comprehensive evaluation suite designed to assess model merging at scale. MergeBench builds on state-of-the-art open-source language models, including Llama and Gemma families at 2B to 9B scales, and covers five key domains: instruction following, mathematics, multilingual understanding, coding and safety. We standardize finetuning and evaluation protocols, and assess eight representative merging methods across multi-task performance, forgetting and runtime efficiency. Based on extensive experiments, we provide practical guidelines for algorithm selection and share insights showing that model merging tends to perform better on stronger base models, with techniques such as merging coefficient tuning and sparsification improving knowledge retention. However, several challenges remain, including the computational cost on large models, the gap for in-domain performance compared to multi-task models, and the underexplored role of model merging in standard LLM training pipelines. We hope MergeBench provides a foundation for future research to advance the understanding and practical application of model merging. Our project page is at \href{https://yifei-he.github.io/mergebench/}{https://yifei-he.github.io/mergebench/}.

[912] A Generic Framework for Conformal Fairness

Aditya T. Vadlamani, Anutam Srinivasan, Pranav Maneriker, Ali Payani, Srinivasan Parthasarathy

Main category: cs.LG

TL;DR: This paper introduces Conformal Fairness, a framework that extends conformal prediction to ensure fairness by controlling coverage gaps between different sensitive groups, with applications to non-IID data like graph data.

DetailsMotivation: Conformal prediction provides uncertainty quantification but is agnostic to sensitive attributes, potentially leading to unfair coverage across different demographic groups. The authors aim to address this limitation by formalizing fairness within the conformal prediction framework.

Method: The authors develop a theoretically well-founded algorithm that leverages the exchangeability assumption (inherent in conformal prediction) rather than the typical IID assumption. This allows their framework to handle non-IID data types like graph data while controlling for coverage gaps between sensitive groups.

Result: Experiments on both graph and tabular datasets demonstrate that the proposed algorithm can effectively control fairness-related gaps in coverage while maintaining coverage aligned with theoretical expectations.

Conclusion: The paper successfully formalizes Conformal Fairness and provides a practical framework that extends conformal prediction to ensure fairness across sensitive groups, with particular applicability to non-IID data scenarios.

Abstract: Conformal Prediction (CP) is a popular method for uncertainty quantification with machine learning models. While conformal prediction provides probabilistic guarantees regarding the coverage of the true label, these guarantees are agnostic to the presence of sensitive attributes within the dataset. In this work, we formalize \textit{Conformal Fairness}, a notion of fairness using conformal predictors, and provide a theoretically well-founded algorithm and associated framework to control for the gaps in coverage between different sensitive groups. Our framework leverages the exchangeability assumption (implicit to CP) rather than the typical IID assumption, allowing us to apply the notion of Conformal Fairness to data types and tasks that are not IID, such as graph data. Experiments were conducted on graph and tabular datasets to demonstrate that the algorithm can control fairness-related gaps in addition to coverage aligned with theoretical expectations.

[913] PICT – A Differentiable, GPU-Accelerated Multi-Block PISO Solver for Simulation-Coupled Learning Tasks in Fluid Dynamics

Aleksandra Franz, Hao Wei, Luca Guastoni, Nils Thuerey

Main category: cs.LG

TL;DR: PICT is a differentiable pressure-implicit fluid simulator in PyTorch with GPU support that enables learning of turbulence models through gradient-based optimization.

DetailsMotivation: Fluid simulation is challenging in scientific computing, and differentiable simulators are needed for optimization and learning in physics simulations using gradient information.

Method: Developed a differentiable pressure-implicit solver in PyTorch with GPU support, verified accuracy in benchmarks, and applied supervised/unsupervised training with physical priors to learn turbulence models.

Result: Successfully learned stable sub-grid scale models for 3D turbulent channel flows that run faster than high-resolution references while maintaining or improving accuracy. Provided insights into solver gradients and developed physically informed regularization.

Conclusion: PICT demonstrates that differentiable fluid simulation enables effective learning of complex turbulence models, offering faster computation with comparable or better accuracy than traditional high-resolution methods.

Abstract: Despite decades of advancements, the simulation of fluids remains one of the most challenging areas of in scientific computing. Supported by the necessity of gradient information in deep learning, differentiable simulators have emerged as an effective tool for optimization and learning in physics simulations. In this work, we present our fluid simulator PICT, a differentiable pressure-implicit solver coded in PyTorch with Graphics-processing-unit (GPU) support. We first verify the accuracy of both the forward simulation and our derived gradients in various established benchmarks like lid-driven cavities and turbulent channel flows before we show that the gradients provided by our solver can be used to learn complicated turbulence models in 2D and 3D. We apply both supervised and unsupervised training regimes using physical priors to match flow statistics. In particular, we learn a stable sub-grid scale (SGS) model for a 3D turbulent channel flow purely based on reference statistics. The low-resolution corrector trained with our solver runs substantially faster than the highly resolved references, while keeping or even surpassing their accuracy. Finally, we give additional insights into the physical interpretation of different solver gradients, and motivate a physically informed regularization technique. To ensure that the full potential of PICT can be leveraged, it is published as open source: https://github.com/tum-pbs/PICT.

[914] LeapFactual: Reliable Visual Counterfactual Explanation Using Conditional Flow Matching

Zhuo Cao, Xuan Zhao, Lena Krieger, Hanno Scharr, Ira Assent

Main category: cs.LG

TL;DR: LeapFactual is a novel counterfactual explanation algorithm using conditional flow matching to generate reliable counterfactuals even when true and learned decision boundaries diverge, overcoming limitations of existing methods like gradient vanishing and discontinuous latent spaces.

DetailsMotivation: The integration of ML/AI models into high-stakes domains requires interpretable models. Current counterfactual methods have limitations including gradient vanishing, discontinuous latent spaces, and overreliance on decision boundary alignment.

Method: LeapFactual uses conditional flow matching to generate counterfactual explanations. It is model-agnostic, works without differentiable loss functions, and can handle human-in-the-loop systems.

Result: Extensive experiments show LeapFactual generates accurate, in-distribution counterfactual explanations with actionable insights. Reliable counterfactual samples can be used as new training data to enhance models.

Conclusion: The method is broadly applicable and enhances both scientific knowledge discovery and non-expert interpretability, expanding counterfactual explanations to domains requiring human participation like citizen science.

Abstract: The growing integration of machine learning (ML) and artificial intelligence (AI) models into high-stakes domains such as healthcare and scientific research calls for models that are not only accurate but also interpretable. Among the existing explainable methods, counterfactual explanations offer interpretability by identifying minimal changes to inputs that would alter a model’s prediction, thus providing deeper insights. However, current counterfactual generation methods suffer from critical limitations, including gradient vanishing, discontinuous latent spaces, and an overreliance on the alignment between learned and true decision boundaries. To overcome these limitations, we propose LeapFactual, a novel counterfactual explanation algorithm based on conditional flow matching. LeapFactual generates reliable and informative counterfactuals, even when true and learned decision boundaries diverge. Following a model-agnostic approach, LeapFactual is not limited to models with differentiable loss functions. It can even handle human-in-the-loop systems, expanding the scope of counterfactual explanations to domains that require the participation of human annotators, such as citizen science. We provide extensive experiments on benchmark and real-world datasets showing that LeapFactual generates accurate and in-distribution counterfactual explanations that offer actionable insights. We observe, for instance, that our reliable counterfactual samples with labels aligning to ground truth can be beneficially used as new training data to enhance the model. The proposed method is broadly applicable and enhances both scientific knowledge discovery and non-expert interpretability.

[915] HERO: Heterogeneous Continual Graph Learning via Meta-Knowledge Distillation

Guiquan Sun, Xikun Zhang, Jingchao Ni, Dongjin Song

Main category: cs.LG

TL;DR: HERO is a continual learning framework for heterogeneous graphs that uses meta-adaptation and knowledge distillation to handle evolving web data while preventing catastrophic forgetting.

DetailsMotivation: Real-world heterogeneous graphs (social networks, knowledge graphs, recommendation systems) are dynamic and continuously evolving, but existing methods assume static graphs. This requires models to adapt to new data while preserving existing knowledge.

Method: HERO employs: 1) Meta-adaptation using gradient-based meta-learning for rapid adaptation to new tasks; 2) DiSCo sampling method for heterogeneity-aware sampling that maximizes node diversity and expands subgraphs along metapaths; 3) Heterogeneity-aware knowledge distillation that aligns knowledge at node and semantic levels.

Result: Extensive experiments on four web-related heterogeneous graph benchmarks show HERO substantially mitigates catastrophic forgetting while achieving efficient and consistent knowledge reuse in dynamic web environments.

Conclusion: HERO provides a unified framework for continual learning on heterogeneous graphs that effectively balances adaptation to new data with preservation of existing knowledge in dynamic web environments.

Abstract: Heterogeneous graph neural networks have seen rapid progress in web applications such as social networks, knowledge graphs, and recommendation systems, driven by the inherent heterogeneity of web data. However, existing methods typically assume static graphs, while real-world graphs are continuously evolving. This dynamic nature requires models to adapt to new data while preserving existing knowledge. To this end, this work introduces HERO (HEterogeneous continual gRaph learning via meta-knOwledge distillation), a unified framework for continual learning on heterogeneous graphs. HERO employs meta-adaptation, a gradient-based meta-learning strategy that provides directional guidance for rapid adaptation to new tasks with limited samples. To enable efficient and effective knowledge reuse, we propose DiSCo (Diversity Sampling with semantic Consistency), a heterogeneity-aware sampling method that maximizes target node diversity and expands subgraphs along metapaths, retaining critical semantic and structural information with minimal overhead. Furthermore, HERO incorporates heterogeneity-aware knowledge distillation, which aligns knowledge at both the node and semantic levels to balance adaptation and retention across tasks. Extensive experiments on four web-related heterogeneous graph benchmarks demonstrate that HERO substantially mitigates catastrophic forgetting while achieving efficient and consistent knowledge reuse in dynamic web environments.

[916] Temperature is All You Need for Generalization in Langevin Dynamics and other Markov Processes

Itamar Harel, Yonathan Wolanowsky, Gal Vardi, Nathan Srebro, Daniel Soudry

Main category: cs.LG

TL;DR: The paper provides a generalization bound for over-parametrized models trained using Markovian stochastic algorithms, specifically Langevin dynamics, with no dependence on training time, dimensionality, or gradient norms.

DetailsMotivation: To understand the generalization gap in over-parametrized models trained with stochastic algorithms, avoiding reliance on mixing, training time, or model-specific properties.

Method: Analyze Langevin dynamics (gradient descent with infinitesimal step size and Gaussian noise) and use a generalized second law of thermodynamics to bound the marginal distribution divergence from initialization.

Result: The generalization gap is bounded by √((β𝔼L(θ₀) + log(1/δ))/N) with probability 1-δ, where β is temperature, 𝔼L(θ₀)=O(1), and N is sample size.

Conclusion: A simple proof shows that Markov process-based training with Gibbs-style stationary distributions provides strong generalization guarantees without typical dependencies.

Abstract: We analyze the generalization gap (gap between the training and test errors) when training a potentially over-parametrized model using a Markovian stochastic training algorithm, initialized from some distribution $\theta_0 \sim p_0$. We focus on Langevin dynamics with a positive temperature $\beta^{-1}$, i.e. gradient descent on a training loss $L$ with infinitesimal step size, perturbed with $\beta^{-1}$-variances Gaussian noise, and lightly regularized or bounded. There, we bound the generalization gap, at any time during training, by $\sqrt{(\beta\mathbb{E} L (\theta_0) + \log(1/\delta))/N}$ with probability $1-\delta$ over the dataset, where $N$ is the sample size, and $\mathbb{E} L (\theta_0) =O(1)$ with standard initialization scaling. In contrast to previous guarantees, we have no dependence on either training time or reliance on mixing, nor a dependence on dimensionality, gradient norms, or any other properties of the loss or model. This guarantee follows from a general analysis of any Markov process-based training that has a Gibbs-style stationary distribution. The proof is surprisingly simple, once we observe that the marginal distribution divergence from initialization remains bounded, as implied by a generalized second law of thermodynamics.

[917] Navigating the Latent Space Dynamics of Neural Models

Marco Fumero, Luca Moschella, Emanuele Rodolà, Francesco Locatello

Main category: cs.LG

TL;DR: Neural networks can be interpreted as dynamical systems on latent manifolds, where autoencoders implicitly define vector fields that reveal attractor points, enabling analysis of generalization, memorization, and prior knowledge without additional training.

DetailsMotivation: To provide an alternative interpretation of neural models as dynamical systems and leverage the implicit vector fields in autoencoders to analyze model properties and data characteristics without requiring extra training.

Method: Interpret autoencoders as dynamical systems by iteratively applying the encoding-decoding map to define latent vector fields, then analyze the emergent attractor points and trajectories in these fields.

Result: The approach enables analysis of generalization/memorization regimes, extraction of prior knowledge from attractors without input data, and identification of out-of-distribution samples through trajectory analysis.

Conclusion: The vector field representation provides a powerful tool for analyzing neural network properties and data characteristics, with demonstrated effectiveness on vision foundation models in real-world scenarios.

Abstract: Neural networks transform high-dimensional data into compact, structured representations, often modeled as elements of a lower dimensional latent space. In this paper, we present an alternative interpretation of neural models as dynamical systems acting on the latent manifold. Specifically, we show that autoencoder models implicitly define a latent vector field on the manifold, derived by iteratively applying the encoding-decoding map, without any additional training. We observe that standard training procedures introduce inductive biases that lead to the emergence of attractor points within this vector field. Drawing on this insight, we propose to leverage the vector field as a representation for the network, providing a novel tool to analyze the properties of the model and the data. This representation enables to: (i) analyze the generalization and memorization regimes of neural models, even throughout training; (ii) extract prior knowledge encoded in the network’s parameters from the attractors, without requiring any input data; (iii) identify out-of-distribution samples from their trajectories in the vector field. We further validate our approach on vision foundation models, showcasing the applicability and effectiveness of our method in real-world scenarios.

[918] Language Models are Injective and Hence Invertible

Giorgos Nikolaou, Tommaso Mencattini, Donato Crisostomi, Andrea Santilli, Yannis Panagakis, Emanuele Rodolà

Main category: cs.LG

TL;DR: Transformer language models are injective (lossless) - different inputs map to different representations, enabling exact input recovery from hidden activations.

DetailsMotivation: Challenge the view that transformer components like non-linear activations prevent exact input recovery, despite their non-injective nature.

Method: Mathematical proof of injectivity at initialization and during training, empirical collision tests on six state-of-the-art models, and development of SipIt algorithm for exact input reconstruction.

Result: No collisions found in billions of tests, SipIt algorithm achieves linear-time exact input reconstruction from hidden activations.

Conclusion: Injectivity is a fundamental property of language models with implications for transparency, interpretability, and safe deployment.

Abstract: Transformer components such as non-linear activations and normalization are inherently non-injective, suggesting that different inputs could map to the same output and prevent exact recovery of the input from a model’s representations. In this paper, we challenge this view. First, we prove mathematically that transformer language models mapping discrete input sequences to their corresponding sequence of continuous representations are injective and therefore lossless, a property established at initialization and preserved during training. Second, we confirm this result empirically through billions of collision tests on six state-of-the-art language models, and observe no collisions. Third, we operationalize injectivity: we introduce SipIt, the first algorithm that provably and efficiently reconstructs the exact input text from hidden activations, establishing linear-time guarantees and demonstrating exact invertibility in practice. Overall, our work establishes injectivity as a fundamental and exploitable property of language models, with direct implications for transparency, interpretability, and safe deployment.

[919] Improved Best-of-Both-Worlds Regret for Bandits with Delayed Feedback

Ofir Schlisselberg, Tal Lancewicki, Peter Auer, Yishay Mansour

Main category: cs.LG

TL;DR: A new Best-of-Both-Worlds algorithm for multi-armed bandits with adversarial delays that achieves near-optimal regret bounds in both stochastic and adversarial environments, matching known lower bounds up to logarithmic factors.

DetailsMotivation: To address the gap in existing algorithms that suffer from significant performance gaps to known lower bounds in delayed bandit environments, particularly in stochastic settings.

Method: Developed a new algorithm for multi-armed bandits with adversarially chosen delays that operates in the Best-of-Both-Worlds framework.

Result: Achieves optimal regret of Õ(√KT + √D) in adversarial case and improved stochastic regret bound of Σ(log T/Δ_i) + (1/K)ΣΔ_iσ_max, where σ_max is maximum missing observations.

Conclusion: This is the first BoBW algorithm to simultaneously match lower bounds in both stochastic and adversarial regimes under delays, with the stochastic bound improving the best known result by a factor of K.

Abstract: We study the multi-armed bandit problem with adversarially chosen delays in the Best-of-Both-Worlds (BoBW) framework, which aims to achieve near-optimal performance in both stochastic and adversarial environments. While prior work has made progress toward this goal, existing algorithms suffer from significant gaps to the known lower bounds, especially in the stochastic settings. Our main contribution is a new algorithm that, up to logarithmic factors, matches the known lower bounds in each setting individually. In the adversarial case, our algorithm achieves regret of $\widetilde{O}(\sqrt{KT} + \sqrt{D})$, which is optimal up to logarithmic terms, where $T$ is the number of rounds, $K$ is the number of arms, and $D$ is the cumulative delay. In the stochastic case, we provide a regret bound which scale as $\sum_{i:\Delta_i>0}\left(\log T/\Delta_i\right) + \frac{1}{K}\sum \Delta_i \sigma_{max}$, where $\Delta_i$ is the sub-optimality gap of arm $i$ and $\sigma_{\max}$ is the maximum number of missing observations. To the best of our knowledge, this is the first BoBW algorithm to simultaneously match the lower bounds in both stochastic and adversarial regimes in delayed environment. Moreover, even beyond the BoBW setting, our stochastic regret bound is the first to match the known lower bound under adversarial delays, improving the second term over the best known result by a factor of $K$.

[920] Neural Network Reprogrammability: A Unified Theme on Model Reprogramming, Prompt Tuning, and Prompt Instruction

Zesheng Ye, Chengyi Cai, Ruijiang Dong, Jianzhong Qi, Lei Feng, Pin-Yu Chen, Feng Liu

Main category: cs.LG

TL;DR: This survey introduces neural network reprogrammability as a unifying framework that bridges model reprogramming, prompt tuning, and prompt instruction - all methods that repurpose pre-trained models by manipulating information at interfaces while keeping parameters frozen.

DetailsMotivation: As large-scale pre-trained models grow in size and capability, efficiently adapting them to downstream tasks becomes critical. Existing adaptation approaches have evolved in isolation without clear understanding of their interrelationships.

Method: The paper presents a taxonomy categorizing information manipulation-based adaptation approaches across four dimensions: manipulation format (fixed/learnable), location (interfaces), operator (application method), and output alignment requirement.

Result: The framework applies consistently across data modalities and model architectures, revealing theoretical connections between established techniques like in-context learning and chain-of-thought prompting.

Conclusion: Neural network reprogrammability is positioned as a fundamental paradigm for efficient model adaptation, with identified research directions emerging from this integrative viewpoint.

Abstract: As large-scale pre-trained foundation models continue to expand in size and capability, efficiently adapting them to specific downstream tasks has become increasingly critical. Despite substantial progress, existing adaptation approaches have evolved largely in isolation, without a clear understanding of their interrelationships. This survey introduces neural network reprogrammability as a unifying framework that bridges mainstream model adaptation techniques–model reprogramming, prompt tuning, and prompt instruction–previously fragmented research areas yet converges on a shared principle: repurposing a pre-trained model by manipulating information at the interfaces while keeping the model parameters frozen. These methods exploit neural networks’ sensitivity to manipulation on different interfaces, be it through perturbing inputs, inserting tokens into intermediate layers, or providing task-specific examples in context, to redirect model behaviors towards desired outcomes. We then present a taxonomy that categorizes such information manipulation-based adaptation approaches across four key dimensions: manipulation format (fixed or learnable), location (interfaces where manipulations occur), operator (how they are applied), and output alignment requirement (post-processing needed to align outputs with downstream tasks). Notably, this framework applies consistently across data modalities, independent of specific model architectures. Moreover, viewing established techniques like in-context learning and chain-of-thought prompting through this lens reveals both their theoretical connections and practical distinctions. We further analyze remaining technical challenges and ethical considerations, positioning neural network reprogrammability as a fundamental paradigm for efficient model adaptation. We lastly identify promising research directions emerging from this integrative viewpoint.

[921] Progressive Tempering Sampler with Diffusion

Severi Rissanen, RuiKang OuYang, Jiajun He, Wenlin Chen, Markus Heinonen, Arno Solin, José Miguel Hernández-Lobato

Main category: cs.LG

TL;DR: PTSD combines Parallel Tempering with diffusion models to create a neural sampler that generates uncorrelated samples efficiently while improving target evaluation performance over pure diffusion methods.

DetailsMotivation: Current neural samplers are less efficient than Parallel Tempering in target evaluations, while PT produces dependent samples and requires expensive reruns for new samples.

Method: Trains diffusion models sequentially across temperatures using PT advantages, combines high-temperature diffusion models to generate approximate lower-temperature samples, then minimally refines them with MCMC for training next diffusion model.

Result: PTSD significantly improves target evaluation efficiency and outperforms diffusion-based neural samplers while generating well-mixed, uncorrelated samples.

Conclusion: The proposed PTSD method effectively combines the strengths of Parallel Tempering and neural samplers, enabling efficient reuse of sample information across temperatures and addressing key weaknesses of both approaches.

Abstract: Recent research has focused on designing neural samplers that amortize the process of sampling from unnormalized densities. However, despite significant advancements, they still fall short of the state-of-the-art MCMC approach, Parallel Tempering (PT), when it comes to the efficiency of target evaluations. On the other hand, unlike a well-trained neural sampler, PT yields only dependent samples and needs to be rerun – at considerable computational cost – whenever new samples are required. To address these weaknesses, we propose the Progressive Tempering Sampler with Diffusion (PTSD), which trains diffusion models sequentially across temperatures, leveraging the advantages of PT to improve the training of neural samplers. We also introduce a novel method to combine high-temperature diffusion models to generate approximate lower-temperature samples, which are minimally refined using MCMC and used to train the next diffusion model. PTSD enables efficient reuse of sample information across temperature levels while generating well-mixed, uncorrelated samples. Our method significantly improves target evaluation efficiency, outperforming diffusion-based neural samplers.

[922] BLUR: A Bi-Level Optimization Approach for LLM Unlearning

Hadi Reisizadeh, Jinghan Jia, Zhiqi Bu, Bhanukiran Vinzamuri, Anil Ramakrishna, Kai-Wei Chang, Volkan Cevher, Sijia Liu, Mingyi Hong

Main category: cs.LG

TL;DR: Proposes BLUR, a bi-level optimization approach for LLM unlearning that prioritizes forgetting over retention, outperforming existing methods.

DetailsMotivation: Current unlearning formulations using weighted sum of forget/retain losses lead to performance degradation due to inherent trade-offs. Need better formulation that respects the hierarchical priority of forgetting over retention.

Method: Bi-level optimization formulation: lower-level minimizes forget loss, upper-level maintains model utility. BLUR algorithm implements this with theoretical guarantees.

Result: BLUR consistently outperforms state-of-the-art unlearning algorithms across various tasks, models, and metrics in extensive experiments.

Conclusion: The hierarchical bi-level optimization formulation effectively addresses LLM unlearning, with BLUR delivering superior performance while maintaining theoretical soundness.

Abstract: Enabling large language models (LLMs) to unlearn knowledge and capabilities acquired during training has proven vital for ensuring compliance with data regulations and promoting ethical practices in generative AI. Although there are growing interests in developing various unlearning algorithms, it remains unclear how to best formulate the unlearning problem. The most popular formulation uses a weighted sum of forget and retain loss, but it often leads to performance degradation due to the inherent trade-off between forget and retain losses. In this work, we argue that it is important to model the hierarchical structure of the unlearning problem, where the forget problem (which \textit{unlearns} certain knowledge and/or capabilities) takes priority over the retain problem (which preserves model utility). This hierarchical structure naturally leads to a bi-level optimization formulation where the lower-level objective focuses on minimizing the forget loss, while the upper-level objective aims to maintain the model’s utility. Based on this new formulation, we propose a novel algorithm, termed Bi-Level UnleaRning (\texttt{BLUR}), which not only possesses strong theoretical guarantees but more importantly, delivers superior performance. In particular, our extensive experiments demonstrate that \texttt{BLUR} consistently outperforms all the state-of-the-art algorithms across various unlearning tasks, models, and metrics. Codes are available at https://github.com/OptimAI-Lab/BLURLLMUnlearning.

[923] FlexQuant: A Flexible and Efficient Dynamic Precision Switching Framework for LLM Quantization

Fangxin Liu, Zongwu Wang, JinHong Xia, Junping Zhao, Shouren Zhao, Jinjin Li, Jian Liu, Li Jiang, Haibing Guan

Main category: cs.LG

TL;DR: FlexQuant is a dynamic precision-switching framework that enables fine-grained, layer-wise mixed-precision quantization for LLMs, achieving 1.3x speedup with minimal accuracy loss by dynamically adjusting bit-widths during token generation.

DetailsMotivation: Address the memory bottleneck in LLMs caused by the gap between model parameter scaling and hardware capabilities, overcoming limitations of static quantization methods that struggle with dynamic workloads.

Method: Uses model perplexity entropy and Kullback-Leibler divergence to enable fine-grained layer-wise mixed-precision quantization, dynamically adjusts bit-widths during each token generation, and implements a precision requirement model for optimal switching.

Result: Achieves 1.3x end-to-end speedup across diverse language tasks with negligible accuracy loss, providing flexible and adaptive quantization for efficient LLM deployment.

Conclusion: FlexQuant offers an effective dynamic precision-switching solution that optimizes the trade-off between inference speed and accuracy in LLM deployment.

Abstract: The rapid advancement of large language models (LLMs) has exacerbated the memory bottleneck due to the widening gap between model parameter scaling and hardware capabilities. While post-training quantization techniques effectively reduce memory overhead, existing methods predominantly rely on static quantization strategies, which struggle to adapt to dynamic workloads. To address this, we propose FlexQuant, a dynamic precision-switching framework that optimizes the trade-off between inference speed and accuracy. Leveraging model perplexity entropy and Kullback-Leibler divergence, FlexQuant enables fine-grained, layer-wise mixed-precision quantization and dynamically adjusts bit-widths during each token generation. FlexQuant provides a comprehensive analysis of quantization strategies, introduces a precision requirement model for optimal switching, and implements efficient fine-grained precision management. Evaluations demonstrate that FlexQuant achieves a 1.3x end-to-end speedup across diverse language tasks with negligible accuracy loss introduced. This framework offers a flexible and adaptive solution for efficient LLM deployment. Code is released at https://github.com/ZongwuWang/FlexQuant.git.

[924] GeoRecon: Graph-Level Representation Learning for 3D Molecules via Reconstruction-Based Pretraining

Shaoheng Yan, Zian Li, Muhan Zhang

Main category: cs.LG

TL;DR: GeoRecon is a graph-level pretraining framework that shifts from node-level denoising to holistic molecular structure reconstruction, improving performance on molecular property prediction tasks.

DetailsMotivation: Current molecular pretraining focuses on node-level denoising which captures local atomic environments but fails to encode global molecular structure needed for graph-level property prediction tasks like energy estimation.

Method: GeoRecon formulates a graph-level reconstruction task where the model learns to produce informative graph representations that guide geometry reconstruction, inducing smoother and more transferable latent spaces.

Result: GeoRecon generally improves over backbone baselines on multiple molecular benchmarks including QM9, MD17, MD22, and 3BPA without external supervision.

Conclusion: Graph-level reconstruction is effective for learning holistic and geometry-aware molecular embeddings that capture coherent global structural features beyond isolated atomic details.

Abstract: The pretraining-finetuning paradigm has powered major advances in domains such as natural language processing and computer vision, with representative examples including masked language modeling and next-token prediction. In molecular representation learning, however, pretraining tasks remain largely restricted to node-level denoising, which effectively captures local atomic environments but is often insufficient for encoding the global molecular structure critical to graph-level property prediction tasks such as energy estimation and molecular regression. To address this gap, we introduce GeoRecon, a graph-level pretraining framework that shifts the focus from individual atoms to the molecule as an integrated whole. GeoRecon formulates a graph-level reconstruction task: during pretraining, the model is trained to produce an informative graph representation that guides geometry reconstruction while inducing smoother and more transferable latent spaces. This encourages the learning of coherent, global structural features beyond isolated atomic details. Without relying on external supervision, GeoRecon generally improves over backbone baselines on multiple molecular benchmarks including QM9, MD17, MD22, and 3BPA, demonstrating the effectiveness of graph-level reconstruction for holistic and geometry-aware molecular embeddings.

[925] Improving Rectified Flow with Boundary Conditions

Xixi Hu, Runlong Liao, Keyang Xu, Bo Liu, Yeqing Li, Eugene Ie, Hongliang Fei, Qiang Liu

Main category: cs.LG

TL;DR: Boundary-enforced Rectified Flow Model improves generative modeling by enforcing boundary conditions in velocity field learning, achieving significant FID score improvements on ImageNet.

DetailsMotivation: Rectified Flow's unconstrained neural network velocity modeling fails to satisfy boundary conditions, causing inaccurate velocity field estimations that deviate from the desired ODE, particularly problematic during stochastic sampling.

Method: Proposed Boundary-enforced Rectified Flow Model that enforces boundary conditions with minimal code modification to the vanilla RF model.

Result: Demonstrated 8.01% improvement in FID score on ImageNet using ODE sampling and 8.98% improvement using SDE sampling compared to vanilla RF model.

Conclusion: Enforcing boundary conditions in Rectified Flow models significantly improves performance and addresses critical issues in velocity field estimation near boundaries.

Abstract: Rectified Flow offers a simple and effective approach to high-quality generative modeling by learning a velocity field. However, we identify a limitation in directly modeling the velocity with an unconstrained neural network: the learned velocity often fails to satisfy certain boundary conditions, leading to inaccurate velocity field estimations that deviate from the desired ODE. This issue is particularly critical during stochastic sampling at inference, as the score function’s errors are amplified near the boundary. To mitigate this, we propose a Boundary-enforced Rectified Flow Model (Boundary RF Model), in which we enforce boundary conditions with a minimal code modification. Boundary RF Model improves performance over vanilla RF model, demonstrating 8.01% improvement in FID score on ImageNet using ODE sampling and 8.98% improvement using SDE sampling.

[926] Online Learning of Whittle Indices for Restless Bandits with Non-Stationary Transition Kernels

Md Kamran Chowdhury Shisher, Vishrant Tripathi, Mung Chiang, Christopher G. Brinton

Main category: cs.LG

TL;DR: The paper proposes SW-Whittle, a sliding-window online Whittle index policy for restless multi-armed bandits with unknown and non-stationary dynamics, achieving dynamic regret guarantees without requiring prior knowledge of variation budgets.

DetailsMotivation: Existing Whittle index policies require stationary transition kernels, which is unrealistic in many applications. Solving RMABs optimally is PSPACE-hard even with full model knowledge, creating a need for efficient adaptive policies that can handle unknown and time-varying dynamics.

Method: Proposed SW-Whittle policy combines sliding-window estimation with online Whittle index computation. Uses Bandit-over-Bandit framework to tune window lengths based on estimated variation, computes Whittle indices via upper-confidence-bound of estimated transition kernels and bilinear optimization.

Result: Achieves dynamic regret of Õ(T^{2/3}Ṽ^{1/3}+T^{4/5}) for large RMABs, where T is episodes and Ṽ is total variation distance. Numerical experiments show consistent outperformance over baselines with lowest cumulative regret across non-stationary environments.

Conclusion: SW-Whittle provides a computationally efficient solution for RMABs with unknown non-stationary dynamics, adapting to time-varying kernels without requiring prior knowledge of variation budgets, making it practical for real-world applications.

Abstract: We study optimal resource allocation in restless multi-armed bandits (RMABs) under unknown and non-stationary dynamics. Solving RMABs optimally is PSPACE-hard even with full knowledge of model parameters, and while the Whittle index policy offers asymptotic optimality with low computational cost, it requires access to stationary transition kernels - an unrealistic assumption in many applications. To address this challenge, we propose a Sliding-Window Online Whittle (SW-Whittle) policy that remains computationally efficient while adapting to time-varying kernels. Our algorithm achieves a dynamic regret of $\tilde O(T^{2/3}\tilde V^{1/3}+T^{4/5})$ for large RMABs, where $T$ is the number of episodes and $\tilde V$ is the total variation distance between consecutive transition kernels. Importantly, we handle the challenging case where the variation budget is unknown in advance by combining a Bandit-over-Bandit framework with our sliding-window design. Window lengths are tuned online as a function of the estimated variation, while Whittle indices are computed via an upper-confidence-bound of the estimated transition kernels and a bilinear optimization routine. Numerical experiments demonstrate that our algorithm consistently outperforms baselines, achieving the lowest cumulative regret across a range of non-stationary environments.

[927] Importance-Aware Activation Space Reconstruction

Md Mokarram Chowdhury, Daniel Agyei Asante, Ernie Chang, Yang Li

Main category: cs.LG

TL;DR: IMPACT is a framework for importance-aware activation reconstruction in LLM compression that considers both activation structure and gradient sensitivity, achieving better size reduction while maintaining accuracy.

DetailsMotivation: Traditional low-rank weight compression assumes weights are low-rank, but this doesn't hold for LLMs. Instead, activations exhibit stronger low-rank structure, but uniform activation reconstruction can harm performance since activation dimensions contribute unequally to model performance.

Method: IMPACT formulates an optimization problem considering both activation structure and gradient sensitivity, deriving a closed-form solution where optimal reconstruction bases are eigenvectors of an importance-weighted activation covariance matrix.

Result: Experiments show IMPACT achieves up to 48.6% greater model size reduction with accuracy comparable to state-of-the-art baselines across diverse models and tasks.

Conclusion: IMPACT provides a principled framework for LLM compression that explicitly optimizes for accuracy preservation through importance-aware activation reconstruction.

Abstract: Large language models (LLMs) achieve strong performance across many domains but are difficult to deploy in resource-constrained settings due to their size. Low-rank weight matrix compression is a popular strategy for reducing model size, typically by minimizing weight reconstruction error under the assumption that weights are low-rank. However, this assumption often does not hold in LLMs. Instead, LLM activations exhibit stronger low-rank structure-prompting a shift toward minimizing activation reconstruction error. We show that this shift alone is insufficient: activation dimensions contribute unequally to model performance, and uniform reconstruction can harm performance. We propose IMPACT, a principled framework for importance-aware activation reconstruction that links model compression decisions to their impact on model behavior. IMPACT formulates an optimization problem that considers both activation structure and gradient sensitivity, and derives a closed-form solution where the optimal reconstruction bases are the eigenvectors of an importance-weighted activation covariance matrix. This enables low-rank approximations explicitly optimized to preserve accuracy. Experiments across diverse models and tasks show that IMPACT achieves up to 48.6% greater model size reduction with accuracy comparable to state-of-the-art baselines.

[928] ESSA: Evolutionary Strategies for Scalable Alignment

Daria Korotyshova, Boris Shaposhnikov, Alexey Malakhov, Alexey Khokhulin, Nikita Surnachev, Kirill Ovcharenko, George Bredis, Alexey Gorbatovski, Viacheslav Sinii, Daniil Gavrilov

Main category: cs.LG

TL;DR: ESSA is a gradient-free framework that uses evolutionary strategies to align LLMs through forward inference and black-box optimization, focusing on Low-Rank Adapters with SVD compression for efficient scaling.

DetailsMotivation: Current LLM alignment methods like RLHF with PPO/GRPO require complex distributed training, large memory budgets, and careful hyperparameter tuning, which become increasingly difficult at billion-parameter scale.

Method: ESSA uses evolutionary strategies for scalable alignment, focusing optimization on Low-Rank Adapters and compressing parameter space by optimizing only singular values from SVD decomposition of adapter matrices, enabling efficient INT4/INT8 quantized inference.

Result: ESSA improved Qwen2.5-Math-7B by 12.6% on GSM8K and 14.8% on PRM800K, raised LLaMA3.1-8B accuracy on IFEval by 22.5% vs GRPO. On Qwen2.5-32B for PRM800K, it reached near-optimal accuracy 2x faster on 16 GPUs and 6x faster on 128 GPUs compared to GRPO.

Conclusion: Evolutionary strategies present a compelling, hardware-friendly alternative to gradient-based LLM alignment, offering competitive quality with substantially reduced wall-clock time and engineering overhead.

Abstract: Alignment of Large Language Models (LLMs) typically relies on Reinforcement Learning from Human Feedback (RLHF) with gradient-based optimizers such as Proximal Policy Optimization (PPO) or Group Relative Policy Optimization (GRPO). While effective, these methods require complex distributed training, large memory budgets, and careful hyperparameter tuning, all of which become increasingly difficult at billion-parameter scale. We present ESSA, Evolutionary Strategies for Scalable Alignment, a gradient-free framework that aligns LLMs using only forward inference and black-box optimization. ESSA focuses optimization on Low-Rank Adapters (LoRA) and further compresses their parameter space by optimizing only the singular values from an SVD decomposition of each adapter matrix. This dimensionality reduction makes evolutionary search practical even for very large models and allows efficient operation in quantized INT4 and INT8 inference mode. Across these benchmarks ESSA improves the test accuracy of Qwen2.5-Math-7B by 12.6% on GSM8K and 14.8% on PRM800K, and raises the accuracy of LLaMA3.1-8B on IFEval by 22.5%, all compared with GRPO. In large-scale settings ESSA shows stronger scaling than gradient-based methods: on Qwen2.5-32B for PRM800K it reaches near-optimal accuracy twice as fast on 16 GPUs and six times as fast on 128 GPUs compared with GRPO. These results position evolutionary strategies as a compelling, hardware-friendly alternative to gradient-based LLM alignment, combining competitive quality with substantially reduced wall-clock time and engineering overhead.

[929] Greedy Low-Rank Gradient Compression for Distributed Learning with Convergence Guarantees

Chuyan Chen, Yutong He, Pengrui Li, Weichen Jia, Kun Yuan

Main category: cs.LG

TL;DR: GreedyLore is the first greedy low-rank gradient compression algorithm with convergence guarantees, achieving linear speedup convergence rate through error feedback and semi-lazy subspace updates.

DetailsMotivation: Distributed optimization faces communication bottlenecks, and existing low-rank compression methods either have high variance (randomized) or lack convergence guarantees (greedy).

Method: GreedyLore uses greedy low-rank compression with error feedback to correct bias and semi-lazy subspace updates to maintain contractive compression operators.

Result: GreedyLore achieves convergence rate of O(σ/√NT + 1/T) under standard optimizers like MSGD and Adam, providing the first linear speedup for low-rank gradient compression.

Conclusion: The proposed GreedyLore algorithm successfully bridges the gap between empirical performance and theoretical guarantees in low-rank gradient compression for distributed learning.

Abstract: Distributed optimization is pivotal for large-scale signal processing and machine learning, yet communication overhead remains a major bottleneck. Low-rank gradient compression, in which the transmitted gradients are approximated by low-rank matrices to reduce communication, offers a promising remedy. Existing methods typically adopt either randomized or greedy compression strategies: randomized approaches project gradients onto randomly chosen subspaces, introducing high variance and degrading empirical performance; greedy methods select the most informative subspaces, achieving strong empirical results but lacking convergence guarantees. To address this gap, we propose GreedyLore–the first Greedy Low-Rank gradient compression algorithm for distributed learning with rigorous convergence guarantees. GreedyLore incorporates error feedback to correct the bias introduced by greedy compression and introduces a semi-lazy subspace update that ensures the compression operator remains contractive throughout all iterations. With these techniques, we prove that GreedyLore achieves a convergence rate of $\mathcal{O}(\sigma/\sqrt{NT} + 1/T)$ under standard optimizers such as MSGD and Adam–marking the first linear speedup convergence rate for low-rank gradient compression. Extensive experiments are conducted to validate our theoretical findings.

[930] Multiscale Neural PDE Surrogates for Prediction and Downscaling: Application to Ocean Currents

Abdessamad El-Kabid, Loubna Benabbou, Redouane Lguensat, Alex Hernández-García

Main category: cs.LG

TL;DR: A deep learning framework using neural operators for solving PDEs and downscaling ocean current data to arbitrary resolution, applied to Copernicus satellite data and Navier-Stokes simulations.

DetailsMotivation: High-resolution ocean current data is crucial for coastal management and safety, but existing satellite products and global models lack sufficient spatial granularity for detailed local analyses.

Method: Supervised deep learning framework based on neural operators that can solve PDEs and provide arbitrary resolution solutions, with specific application to downscaling Copernicus ocean current data.

Result: The model was evaluated on real-world Copernicus ocean current data and synthetic Navier-Stokes simulation datasets, demonstrating capability to predict solutions at arbitrary resolution regardless of input resolution.

Conclusion: The proposed neural operator framework successfully addresses the resolution limitations of existing ocean current data products by enabling arbitrary resolution downscaling and PDE solution prediction.

Abstract: Accurate modeling of physical systems governed by partial differential equations is a central challenge in scientific computing. In oceanography, high-resolution current data are critical for coastal management, environmental monitoring, and maritime safety. However, available satellite products, such as Copernicus data for sea water velocity at ~0.08 degrees spatial resolution and global ocean models, often lack the spatial granularity required for detailed local analyses. In this work, we (a) introduce a supervised deep learning framework based on neural operators for solving PDEs and providing arbitrary resolution solutions, and (b) propose downscaling models with an application to Copernicus ocean current data. Additionally, our method can model surrogate PDEs and predict solutions at arbitrary resolution, regardless of the input resolution. We evaluated our model on real-world Copernicus ocean current data and synthetic Navier-Stokes simulation datasets.

[931] Reliable Wireless Indoor Localization via Cross-Validated Prediction-Powered Calibration

Seonghoon Yoo, Houssem Sifaou, Sangwoo Park, Joonhyuk Kang, Osvaldo Simeone

Main category: cs.LG

TL;DR: Proposes a method to efficiently use limited calibration data for wireless indoor localization by simultaneously fine-tuning predictors and estimating synthetic label bias, ensuring rigorous coverage guarantees.

DetailsMotivation: Wireless indoor localization using RSSI requires calibration, but synthetic labels from different models need fine-tuning and bias estimation, which demands additional data and worsens calibration data scarcity.

Method: An approach that uses limited calibration data to simultaneously fine-tune a predictor and estimate the bias of synthetic labels, producing prediction sets with rigorous coverage guarantees.

Result: Experiments on a fingerprinting dataset validate the effectiveness of the proposed method.

Conclusion: The proposed method efficiently addresses calibration data scarcity in wireless indoor localization while maintaining reliable position estimates with coverage guarantees.

Abstract: Wireless indoor localization using predictive models with received signal strength information (RSSI) requires proper calibration for reliable position estimates. One remedy is to employ synthetic labels produced by a (generally different) predictive model. But fine-tuning an additional predictor, as well as estimating residual bias of the synthetic labels, demands additional data, aggravating calibration data scarcity in wireless environments. This letter proposes an approach that efficiently uses limited calibration data to simultaneously fine-tune a predictor and estimate the bias of synthetic labels, yielding prediction sets with rigorous coverage guarantees. Experiments on a fingerprinting dataset validate the effectiveness of the proposed method.

[932] Towards Explainable Deep Clustering for Time Series Data

Udo Schlegel, Gabriel Marques Tavares, Thomas Seidl

Main category: cs.LG

TL;DR: This survey provides a comprehensive overview of explainable deep clustering methods for time series data, analyzing current approaches and identifying research gaps across healthcare, finance, IoT, and climate science domains.

DetailsMotivation: Deep clustering reveals hidden patterns in time series data but lacks transparency, limiting its use in safety-critical applications where explainability is crucial.

Method: The authors conducted a structured survey of peer-reviewed and preprint papers, analyzing methods based on autoencoder and attention architectures, and comparing applications across multiple domains.

Result: Analysis shows most current work relies on autoencoder and attention architectures with limited support for streaming, irregularly sampled, or privacy-preserved data, and interpretability is typically treated as an add-on rather than a core design feature.

Conclusion: The paper outlines six key research opportunities to advance the field: combining complex networks with built-in interpretability, developing faithfulness-focused evaluation metrics, adaptive stream explainers, domain-specific explanations, human-in-the-loop methods, and better understanding of model internals - proposing interpretability as a primary design goal for trustworthy deep clustering analytics.

Abstract: Deep clustering uncovers hidden patterns and groups in complex time series data, yet its opaque decision-making limits use in safety-critical settings. This survey offers a structured overview of explainable deep clustering for time series, collecting current methods and their real-world applications. We thoroughly discuss and compare peer-reviewed and preprint papers through application domains across healthcare, finance, IoT, and climate science. Our analysis reveals that most work relies on autoencoder and attention architectures, with limited support for streaming, irregularly sampled, or privacy-preserved series, and interpretability is still primarily treated as an add-on. To push the field forward, we outline six research opportunities: (1) combining complex networks with built-in interpretability; (2) setting up clear, faithfulness-focused evaluation metrics for unsupervised explanations; (3) building explainers that adapt to live data streams; (4) crafting explanations tailored to specific domains; (5) adding human-in-the-loop methods that refine clusters and explanations together; and (6) improving our understanding of how time series clustering models work internally. By making interpretability a primary design goal rather than an afterthought, we propose the groundwork for the next generation of trustworthy deep clustering time series analytics.

[933] Wavy Transformer

Satoshi Noguchi, Yoshinobu Kawahara

Main category: cs.LG

TL;DR: Wavy Transformer addresses over-smoothing in deep transformers by modeling attention layers as graph neural diffusion and introducing second-order wavy dynamics to prevent token representation convergence.

DetailsMotivation: Deep transformer models suffer from over-smoothing where token representations become similar across layers, limiting model performance. The paper establishes an equivalence between attention layers and graph neural diffusion to understand this phenomenon.

Method: Proposes Wavy Transformer with second-order wavy dynamics attention layers, plus modified feed-forward networks and normalization layers that preserve physical state-velocity relationships under the chain rule.

Result: Wavy Transformer consistently improves performance across various NLP and CV tasks with minimal additional parameters and no extra hyperparameter tuning required.

Conclusion: Modeling transformer dynamics as physical diffusion processes provides effective solutions to over-smoothing, enabling better performance in deep transformer architectures.

Abstract: Transformers have achieved remarkable success across natural language processing (NLP) and computer vision (CV). However, deep transformer models often suffer from an over-smoothing issue, in which token representations converge to similar values as they pass through successive transformer blocks. In this paper, we establish an equivalence between the hidden-state dynamics induced by stacked attention layers and graph neural diffusion on a complete graph. From this perspective, over-smoothing can be interpreted as a consequence of the dissipative nature of the underlying diffusion dynamics. Motivated by this physical interpretation, we propose Wavy Transformer, which consists of a novel attention layer based on second-order wavy dynamics. We also introduce a feed-forward network and a normalization layer designed to preserve the physical state-velocity relationship under the chain rule, thereby extending the transformer architecture. We further validate our proposed techniques on various transformer models for NLP and CV tasks. The results consistently demonstrate that Wavy Transformer improves performance with minimal additional parameters and no extra hyperparameter tuning.

[934] From AI for Science to Agentic Science: A Survey on Autonomous Scientific Discovery

Jiaqi Wei, Yuejin Yang, Xiang Zhang, Yuhan Chen, Xiang Zhuang, Zhangyang Gao, Dongzhan Zhou, Guangshuai Wang, Zhiqiang Gao, Juntai Cao, Zijie Qiu, Ming Hu, Chenglong Ma, Shixiang Tang, Junjun He, Chunfeng Song, Xuming He, Qiang Zhang, Chenyu You, Shuangjia Zheng, Ning Ding, Wanli Ouyang, Nanqing Dong, Yu Cheng, Siqi Sun, Lei Bai, Bowen Zhou

Main category: cs.LG

TL;DR: This survey establishes Agentic Science as a new paradigm where AI systems evolve from partial assistance to full scientific agency, enabled by LLMs and multimodal systems. It provides a comprehensive framework unifying three perspectives and reviews applications across life sciences, chemistry, materials science, and physics.

DetailsMotivation: To position Agentic Science as a pivotal stage in AI for Science evolution, where AI systems progress from computational tools to autonomous research partners capable of hypothesis generation, experimental design, execution, analysis, and iterative refinement.

Method: The survey provides a domain-oriented review across multiple scientific fields, unifying process-oriented, autonomy-oriented, and mechanism-oriented perspectives through a comprehensive framework that connects foundational capabilities, core processes, and domain-specific realizations.

Result: The work establishes a structured paradigm for Agentic Science, tracing AI evolution in science, identifying five core capabilities for scientific agency, modeling discovery as a four-stage workflow, and reviewing applications across life sciences, chemistry, materials science, and physics.

Conclusion: This survey establishes Agentic Science as a structured paradigm for advancing AI-driven research, providing a domain-oriented synthesis of autonomous scientific discovery and identifying key challenges and future opportunities in the field.

Abstract: Artificial intelligence (AI) is reshaping scientific discovery, evolving from specialized computational tools into autonomous research partners. We position Agentic Science as a pivotal stage within the broader AI for Science paradigm, where AI systems progress from partial assistance to full scientific agency. Enabled by large language models (LLMs), multimodal systems, and integrated research platforms, agentic AI shows capabilities in hypothesis generation, experimental design, execution, analysis, and iterative refinement – behaviors once regarded as uniquely human. This survey provides a domain-oriented review of autonomous scientific discovery across life sciences, chemistry, materials science, and physics. We unify three previously fragmented perspectives – process-oriented, autonomy-oriented, and mechanism-oriented – through a comprehensive framework that connects foundational capabilities, core processes, and domain-specific realizations. Building on this framework, we (i) trace the evolution of AI for Science, (ii) identify five core capabilities underpinning scientific agency, (iii) model discovery as a dynamic four-stage workflow, (iv) review applications across the above domains, and (v) synthesize key challenges and future opportunities. This work establishes a domain-oriented synthesis of autonomous scientific discovery and positions Agentic Science as a structured paradigm for advancing AI-driven research.

[935] Frozen in Time: Parameter-Efficient Time Series Transformers via Reservoir-Induced Feature Expansion and Fixed Random Dynamics

Pradeep Singh, Mehak Sharma, Anupriya Dey, Balasubramanian Raman

Main category: cs.LG

TL;DR: FreezeTST is a hybrid model combining frozen random-feature blocks with trainable Transformer layers for efficient long-term time-series forecasting, reducing training costs while maintaining performance.

DetailsMotivation: Transformers have quadratic self-attention costs and weak temporal bias, making long-range forecasting expensive and brittle. The goal is to create a more efficient alternative that maintains performance.

Method: Interleaves frozen random-feature (reservoir) blocks with standard trainable Transformer layers. Frozen blocks provide nonlinear memory at no optimization cost, while trainable layers learn to query this memory through self-attention.

Result: On seven standard long-term forecasting benchmarks, FreezeTST consistently matches or surpasses specialized variants like Informer, Autoformer, and PatchTST with substantially lower compute. Reduces trainable parameters and wall-clock training time while maintaining inference complexity.

Conclusion: Embedding reservoir principles within Transformers offers a simple, principled route to efficient long-term time-series prediction, demonstrating that hybrid approaches can achieve strong performance with reduced computational costs.

Abstract: Transformers are the de-facto choice for sequence modelling, yet their quadratic self-attention and weak temporal bias can make long-range forecasting both expensive and brittle. We introduce FreezeTST, a lightweight hybrid that interleaves frozen random-feature (reservoir) blocks with standard trainable Transformer layers. The frozen blocks endow the network with rich nonlinear memory at no optimisation cost; the trainable layers learn to query this memory through self-attention. The design cuts trainable parameters and also lowers wall-clock training time, while leaving inference complexity unchanged. On seven standard long-term forecasting benchmarks, FreezeTST consistently matches or surpasses specialised variants such as Informer, Autoformer, and PatchTST; with substantially lower compute. Our results show that embedding reservoir principles within Transformers offers a simple, principled route to efficient long-term time-series prediction.

[936] HypER: Hyperbolic Echo State Networks for Capturing Stretch-and-Fold Dynamics in Chaotic Flows

Pradeep Singh, Sutirtha Ghosh, Ashutosh Kumar, Hrishit B P, Balasubramanian Raman

Main category: cs.LG

TL;DR: HypER is a novel Echo State Network that uses hyperbolic geometry in the Poincare ball to better match the structure of chaotic systems, enabling longer prediction horizons than traditional ESNs.

DetailsMotivation: Traditional ESNs have Euclidean geometry that mismatches the stretch-and-fold structure of chaos, limiting their ability to forecast chaotic dynamics beyond a few Lyapunov times.

Method: HypER samples neurons in the Poincare ball with connections decaying exponentially with hyperbolic distance, embedding negative curvature directly into the latent space while preserving standard ESN features like sparsity and spectral-radius control.

Result: On chaotic systems (Lorenz-63, Roessler, Chen-Ueta) and real-world benchmarks (heart-rate variability, sunspot numbers), HypER consistently outperforms Euclidean and graph-structured ESN baselines, with statistically significant gains in prediction horizon.

Conclusion: The hyperbolic embedding approach successfully aligns reservoir dynamics with chaotic system structure, enabling improved long-term forecasting of chaotic phenomena.

Abstract: Forecasting chaotic dynamics beyond a few Lyapunov times is difficult because infinitesimal errors grow exponentially. Existing Echo State Networks (ESNs) mitigate this growth but employ reservoirs whose Euclidean geometry is mismatched to the stretch-and-fold structure of chaos. We introduce the Hyperbolic Embedding Reservoir (HypER), an ESN whose neurons are sampled in the Poincare ball and whose connections decay exponentially with hyperbolic distance. This negative-curvature construction embeds an exponential metric directly into the latent space, aligning the reservoir’s local expansion-contraction spectrum with the system’s Lyapunov directions while preserving standard ESN features such as sparsity, leaky integration, and spectral-radius control. Training is limited to a Tikhonov-regularized readout. On the chaotic Lorenz-63 and Roessler systems, and the hyperchaotic Chen-Ueta attractor, HypER consistently lengthens the mean valid-prediction horizon beyond Euclidean and graph-structured ESN baselines, with statistically significant gains confirmed over 30 independent runs; parallel results on real-world benchmarks, including heart-rate variability from the Santa Fe and MIT-BIH datasets and international sunspot numbers, corroborate its advantage. We further establish a lower bound on the rate of state divergence for HypER, mirroring Lyapunov growth.

[937] MatPROV: A Provenance Graph Dataset of Material Synthesis Extracted from Scientific Literature

Hirofumi Tsuruta, Masaya Kumagai

Main category: cs.LG

TL;DR: MatPROV introduces a dataset using PROV-DM standard to extract structured synthesis procedures from materials literature via LLMs, capturing complex causal relationships as graphs.

DetailsMotivation: Existing approaches use rigid schemas or linear sequences that fail to capture the structural complexity of real-world materials synthesis procedures.

Method: Adopt PROV-DM international standard for provenance information to model synthesis procedures as flexible graphs, using large language models to extract procedures from scientific literature.

Result: Created MatPROV dataset with visually intuitive directed graphs that capture structural complexities and causal relationships among materials, operations, and conditions.

Conclusion: PROV-DM-based representation enables machine-interpretable synthesis knowledge for applications like automated synthesis planning and optimization.

Abstract: Synthesis procedures play a critical role in materials research, as they directly affect material properties. With data-driven approaches increasingly accelerating materials discovery, there is growing interest in extracting synthesis procedures from scientific literature as structured data. However, existing studies often rely on rigid, domain-specific schemas with predefined fields for structuring synthesis procedures or assume that synthesis procedures are linear sequences of operations, which limits their ability to capture the structural complexity of real-world procedures. To address these limitations, we adopt PROV-DM, an international standard for provenance information, which supports flexible, graph-based modeling of procedures. We present MatPROV, a dataset of PROV-DM-compliant synthesis procedures extracted from scientific literature using large language models. MatPROV captures structural complexities and causal relationships among materials, operations, and conditions through visually intuitive directed graphs. This representation enables machine-interpretable synthesis knowledge, opening opportunities for future research such as automated synthesis planning and optimization.

[938] Robust Anomaly Detection through Multi-Modal Autoencoder Fusion for Small Vehicle Damage Detection

Sara Khan, Mehmed Yüksel, Frank Kirchner

Main category: cs.LG

TL;DR: A multi-modal anomaly detection system using IMUs and microphones for real-time wear and tear detection in fleet vehicles, achieving 92% ROC-AUC.

DetailsMotivation: Manual inspection is labor-intensive and error-prone, while image-based methods fail during motion and cannot detect underbody damage effectively.

Method: Multi-modal autoencoder-based architectures using IMUs and microphones mounted on windshield, with ensemble models and pooling techniques.

Result: Multi-modal ensemble model with pooling achieved 92% ROC-AUC, outperforming unimodal and state-of-the-art methods.

Conclusion: The approach effectively detects vehicle damage in real-time, can integrate with safety systems, and has potential applications in autonomous vehicle collision detection.

Abstract: Wear and tear detection in fleet and shared vehicle systems is a critical challenge, particularly in rental and car-sharing services, where minor damage, such as dents, scratches, and underbody impacts, often goes unnoticed or is detected too late. Currently, manual inspection methods are the default approach, but are labour-intensive and prone to human error. In contrast, state-of-the-art image-based methods are less reliable when the vehicle is moving, and they cannot effectively capture underbody damage due to limited visual access and spatial coverage. This work introduces a novel multi-modal architecture based on anomaly detection to address these issues. Sensors such as Inertial Measurement Units (IMUs) and microphones are integrated into a compact device mounted on the vehicle’s windshield. This approach supports real-time damage detection while avoiding the need for highly resource-intensive sensors. We developed multiple variants of multi-modal autoencoder-based architectures and evaluated them against unimodal and state-of-the-art methods. Our multi-modal ensemble model with pooling achieved the highest performance, with a Receiver Operating Characteristic-Area Under Curve (ROC-AUC) of 92%, demonstrating its effectiveness in real-world applications. This approach can also be extended to other applications, such as improving automotive safety. It can integrate with airbag systems for efficient deployment and help autonomous vehicles by complementing other sensors in collision detection.

[939] Unlocking the Power of Mixture-of-Experts for Task-Aware Time Series Analytics

Xingjian Wu, Zhengyu Li, Hanyin Cheng, Xiangfei Qiu, Jilin Hu, Chenjuan Guo, Bin Yang

Main category: cs.LG

TL;DR: PatchMoE is a novel Mixture-of-Experts framework for time series analysis that addresses limitations of traditional MoE by introducing task-aware routing and modeling channel correlations through recurrent noisy gating and temporal/channel load balancing.

DetailsMotivation: Traditional MoE architectures are task-agnostic and lack capability in modeling channel correlations, making them ineffective for versatile time series analytics tasks despite their success in NLP.

Method: Proposes PatchMoE with Recurrent Noisy Gating to utilize hierarchical information for task-specific routing, operates routing on time series tokens in temporal and channel dimensions, and uses Temporal & Channel Load Balancing Loss to model correlations.

Result: Comprehensive experiments on five downstream tasks demonstrate state-of-the-art performance.

Conclusion: PatchMoE effectively supports intricate knowledge utilization for distinct time series tasks through task-aware routing and correlation modeling, achieving superior performance across multiple applications.

Abstract: Time Series Analysis is widely used in various real-world applications such as weather forecasting, financial fraud detection, imputation for missing data in IoT systems, and classification for action recognization. Mixture-of-Experts (MoE), as a powerful architecture, though demonstrating effectiveness in NLP, still falls short in adapting to versatile tasks in time series analytics due to its task-agnostic router and the lack of capability in modeling channel correlations. In this study, we propose a novel, general MoE-based time series framework called PatchMoE to support the intricate ``knowledge’’ utilization for distinct tasks, thus task-aware. Based on the observation that hierarchical representations often vary across tasks, e.g., forecasting vs. classification, we propose a Recurrent Noisy Gating to utilize the hierarchical information in routing, thus obtaining task-sepcific capability. And the routing strategy is operated on time series tokens in both temporal and channel dimensions, and encouraged by a meticulously designed Temporal & Channel Load Balancing Loss to model the intricate temporal and channel correlations. Comprehensive experiments on five downstream tasks demonstrate the state-of-the-art performance of PatchMoE.

[940] Aurora: Towards Universal Generative Multimodal Time Series Forecasting

Xingjian Wu, Jianxin Jin, Wanghui Qiu, Peng Chen, Yang Shu, Bin Yang, Chenjuan Guo

Main category: cs.LG

TL;DR: Aurora is a multimodal time series foundation model that supports multimodal inputs and zero-shot inference for cross-domain generalization in time series forecasting.

DetailsMotivation: Existing unimodal time series foundation models lack explicit utilization of domain-specific knowledge from text/image modalities, while supervised multimodal models don't support zero-shot inference for cross-domain scenarios.

Method: Uses tokenization, encoding, and distillation to extract multimodal domain knowledge, employs Modality-Guided Multi-head Self-Attention to inject knowledge into temporal modeling, and uses Prototype-Guided Flow Matching for generative probabilistic forecasting.

Result: Achieves state-of-the-art performance on TimeMMD, TSFM-Bench and ProbTS benchmarks in both unimodal and multimodal scenarios.

Conclusion: Aurora demonstrates strong cross-domain generalization capability through adaptive extraction of domain knowledge from multimodal inputs and supports zero-shot inference.

Abstract: Cross-domain generalization is very important in Time Series Forecasting because similar historical information may lead to distinct future trends due to the domain-specific characteristics. Recent works focus on building unimodal time series foundation models and end-to-end multimodal supervised models. Since domain-specific knowledge is often contained in modalities like texts, the former lacks the explicit utilization of them, thus hindering the performance. The latter is tailored for end-to-end scenarios and does not support zero-shot inference for cross-domain scenarios. In this work, we introduce Aurora, a Multimodal Time Series Foundation Model, which supports multimodal inputs and zero-shot inference. Pretrained on Corss-domain Multimodal Time Series Corpus, Aurora can adaptively extract and focus on key domain knowledge contained in corrsponding text or image modalities, thus possessing strong Cross-domain generalization capability. Through tokenization, encoding, and distillation, Aurora can extract multimodal domain knowledge as guidance and then utilizes a Modality-Guided Multi-head Self-Attention to inject them into the modeling of temporal representations. In the decoding phase, the multimodal representations are used to generate the conditions and prototypes of future tokens, contributing to a novel Prototype-Guided Flow Matching for generative probabilistic forecasting. Comprehensive experiments on well-recognized benchmarks, including TimeMMD, TSFM-Bench and ProbTS, demonstrate the consistent state-of-the-art performance of Aurora on both unimodal and multimodal scenarios.

[941] Conformal Prediction for Signal Temporal Logic Inference

Danyang Li, Yixuan Wang, Matthew Cleaveland, Mingyu Cai, Roberto Tron

Main category: cs.LG

TL;DR: End-to-end differentiable conformal prediction framework for STL inference that provides statistical guarantees while improving both reliability and interpretability of learned formulas.

DetailsMotivation: Existing STL inference methods lack formal confidence guarantees, and standard conformal prediction is typically applied as a post-training wrapper without improving model learning.

Method: Introduces robustness-based nonconformity score, embeds smooth CP layer directly into training, and uses new loss function that simultaneously optimizes inference accuracy and CP prediction sets.

Result: Reduces prediction uncertainty (high coverage with smaller prediction sets) and improves accuracy over state-of-the-art baselines on benchmark time-series tasks.

Conclusion: The framework successfully integrates conformal prediction into STL inference training, providing statistical guarantees while enhancing both reliability and interpretability of learned temporal logic formulas.

Abstract: Signal Temporal Logic (STL) inference seeks to extract human-interpretable rules from time-series data, but existing methods lack formal confidence guarantees for the inferred rules. Conformal prediction (CP) is a technique that can provide statistical correctness guarantees, but is typically applied as a post-training wrapper without improving model learning. Instead, we introduce an end-to-end differentiable CP framework for STL inference that enhances both reliability and interpretability of the resulting formulas. We introduce a robustness-based nonconformity score, embed a smooth CP layer directly into training, and employ a new loss function that simultaneously optimizes inference accuracy and CP prediction sets with a single term. Following training, an exact CP procedure delivers statistical guarantees for the learned STL formulas. Experiments on benchmark time-series tasks show that our approach reduces uncertainty in predictions (i.e., it achieves high coverage while reducing prediction set size), and improves accuracy (i.e., the number of misclassifications when using a fixed threshold) over state-of-the-art baselines.

[942] Accelerated Evolving Set Processes for Local PageRank Computation

Binbin Huang, Luo Luo, Yanghua Xiao, Deqing Yang, Baojian Zhou

Main category: cs.LG

TL;DR: A novel framework using nested evolving set processes to accelerate Personalized PageRank computation, achieving time complexity independent of graph size under certain conditions.

DetailsMotivation: To develop faster algorithms for Personalized PageRank computation that overcome limitations of existing methods, particularly their dependency on graph size.

Method: Uses nested evolving set processes with localized inexact proximal point iterations to solve simplified linear systems, requiring only ~O(1/√α) such systems to be solved.

Result: Achieves time complexity of min{~O(R²/ε²), ~O(m)} for ε-approximation, and ~O(R²/(√αε²)) when 1/ε² ≪ m, independent of graph size. Validated on real-world graphs with early convergence.

Conclusion: The framework successfully accelerates PPR computation and resolves an open conjecture, providing graph-size-independent complexity under practical conditions.

Abstract: This work proposes a novel framework based on nested evolving set processes to accelerate Personalized PageRank (PPR) computation. At each stage of the process, we employ a localized inexact proximal point iteration to solve a simplified linear system. We show that the time complexity of such localized methods is upper bounded by $\min{\tilde{\mathcal{O}}(R^2/\epsilon^2), \tilde{\mathcal{O}}(m)}$ to obtain an $\epsilon$-approximation of the PPR vector, where $m$ denotes the number of edges in the graph and $R$ is a constant defined via nested evolving set processes. Furthermore, the algorithms induced by our framework require solving only $\tilde{\mathcal{O}}(1/\sqrt{\alpha})$ such linear systems, where $\alpha$ is the damping factor. When $1/\epsilon^2\ll m$, this implies the existence of an algorithm that computes an $\ epsilon $-approximation of the PPR vector with an overall time complexity of $\tilde{\mathcal{O}}\left(R^2 / (\sqrt{\alpha}\epsilon^2)\right)$, independent of the underlying graph size. Our result resolves an open conjecture from existing literature. Experimental results on real-world graphs validate the efficiency of our methods, demonstrating significant convergence in the early stages.

[943] When LLM Agents Meet Graph Optimization: An Automated Data Quality Improvement Approach

Zhihan Zhang, Xunkai Li, Yilong Zuo, Zhaoxin Fan, Zhenjun Li, Bing Zhou, Rong-Hua Li, Guoren Wang

Main category: cs.LG

TL;DR: LAGA is a unified multi-agent framework that uses Large Language and Graph Agents to comprehensively optimize Text-attributed Graph (TAG) quality through automated detection, planning, action, and evaluation cycles.

DetailsMotivation: Graph neural networks (GNNs) are highly sensitive to data quality issues in TAGs, with both conventional and LLM-enhanced GNNs degrading significantly under textual, structural, and label imperfections. Existing approaches lack systematic quality improvement methods.

Method: LAGA formulates graph quality control as a data-centric process with four coordinated agents: detection, planning, action, and evaluation. It performs multi-modal optimization across textual, structural, and label aspects in an automated loop.

Result: Extensive experiments on 5 datasets and 16 baselines across 9 scenarios demonstrate LAGA’s effectiveness, robustness, and scalability in improving TAG quality for reliable analytics.

Conclusion: Data-centric quality optimization is crucial for reliable TAG analytics, and LAGA provides a comprehensive framework that addresses the limitations of existing approaches by systematically improving multiple quality aspects.

Abstract: Text-attributed graphs (TAGs) have become a key form of graph-structured data in modern data management and analytics, combining structural relationships with rich textual semantics for diverse applications. However, the effectiveness of analytical models, particularly graph neural networks (GNNs), is highly sensitive to data quality. Our empirical analysis shows that both conventional and LLM-enhanced GNNs degrade notably under textual, structural, and label imperfections, underscoring TAG quality as a key bottleneck for reliable analytics. Existing studies have explored data-level optimization for TAGs, but most focus on specific degradation types and target a single aspect like structure or label, lacking a systematic and comprehensive perspective on data quality improvement. To address this gap, we propose LAGA (Large Language and Graph Agent), a unified multi-agent framework for comprehensive TAG quality optimization. LAGA formulates graph quality control as a data-centric process, integrating detection, planning, action, and evaluation agents into an automated loop. It holistically enhances textual, structural, and label aspects through coordinated multi-modal optimization. Extensive experiments on 5 datasets and 16 baselines across 9 scenarios demonstrate the effectiveness, robustness and scalability of LAGA, confirming the importance of data-centric quality optimization for reliable TAG analytics.

[944] Enhancing the Cross-Size Generalization for Solving Vehicle Routing Problems via Continual Learning

Jingwen Li, Zhiguang Cao, Yaoxin Wu, Tang Liu

Main category: cs.LG

TL;DR: A continual learning framework for vehicle routing problems that trains deep models sequentially with instances of ascending sizes, using inter-task and intra-task regularization plus experience replay to improve generalization across different problem sizes.

DetailsMotivation: Existing deep models for vehicle routing problems are typically trained and evaluated on instances of a single size, which limits their ability to generalize across different problem sizes and hampers practical applicability.

Method: Proposes a continual learning framework with: 1) inter-task regularization to retain knowledge from smaller problem sizes, 2) intra-task regularization to consolidate model by imitating latest desirable behaviors, and 3) experience replay to revisit instances from previously trained sizes to mitigate catastrophic forgetting.

Result: Achieves superior performance across various problem sizes (both seen and unseen in training) compared to state-of-the-art deep models, including those specialized for generalizability enhancement. Ablation studies show synergistic effect of the key designs.

Conclusion: The proposed continual learning framework effectively addresses the generalization limitation of existing deep models for vehicle routing problems across different problem sizes, demonstrating practical applicability through superior performance.

Abstract: Exploring machine learning techniques for addressing vehicle routing problems has attracted considerable research attention. To achieve decent and efficient solutions, existing deep models for vehicle routing problems are typically trained and evaluated using instances of a single size. This substantially limits their ability to generalize across different problem sizes and thus hampers their practical applicability. To address the issue, we propose a continual learning based framework that sequentially trains a deep model with instances of ascending problem sizes. Specifically, on the one hand, we design an inter-task regularization scheme to retain the knowledge acquired from smaller problem sizes in the model training on a larger size. On the other hand, we introduce an intra-task regularization scheme to consolidate the model by imitating the latest desirable behaviors during training on each size. Additionally, we exploit the experience replay to revisit instances of formerly trained sizes for mitigating the catastrophic forgetting. Experimental results show that our approach achieves predominantly superior performance across various problem sizes (either seen or unseen in the training), as compared to state-of-the-art deep models including the ones specialized for generalizability enhancement. Meanwhile, the ablation studies on the key designs manifest their synergistic effect in the proposed framework.

[945] Budget Allocation for Unknown Value Functions in a Lipschitz Space

MohammadHossein Bateni, Hossein Esfandiari, Samira HosseinGhorban, Alireza Mirrokni, Radin Shahdaei

Main category: cs.LG

TL;DR: This paper addresses the challenge of optimally allocating a limited budget to evaluate intermediate models during machine learning development, formalizing it as a budget allocation problem over unknown-value functions in a Lipschitz space.

DetailsMotivation: Building learning models requires evaluating many intermediate models during feature selection, model structure search, and parameter tuning. These evaluations influence subsequent exploration decisions, but true performance is only revealed after evaluation, creating a need for optimal budget allocation.

Method: The authors formalize the problem as a general budget allocation problem over unknown-value functions within a Lipschitz space, providing a theoretical framework for optimal exploration of intermediate models.

Result: The paper presents a formalization of the budget allocation problem for intermediate model evaluation, though specific experimental results are not detailed in the abstract.

Conclusion: The work provides a theoretical foundation for optimally allocating evaluation budgets when exploring intermediate machine learning models, addressing a key challenge in model development workflows.

Abstract: Building learning models frequently requires evaluating numerous intermediate models. Examples include models considered during feature selection, model structure search, and parameter tunings. The evaluation of an intermediate model influences subsequent model exploration decisions. Although prior knowledge can provide initial quality estimates, true performance is only revealed after evaluation. In this work, we address the challenge of optimally allocating a bounded budget to explore the space of intermediate models. We formalize this as a general budget allocation problem over unknown-value functions within a Lipschitz space.

[946] Federated Conditional Conformal Prediction via Generative Models

Rui Xu, Xingyuan Chen, Wenxing Huang, Minxuan Huang, Yun Xie, Weiyan Chen, Sihong Xie

Main category: cs.LG

TL;DR: Fed-CCP is a federated conditional conformal prediction method that uses generative models to achieve conditional coverage in non-i.i.d. federated learning settings, addressing the limitations of marginal coverage guarantees.

DetailsMotivation: Standard conformal prediction assumes i.i.d. data, which is violated in federated learning where client distributions differ. Existing federated CP methods only provide marginal coverage per client, failing to capture input-conditional uncertainty.

Method: Fed-CCP uses generative models (normalizing flows or diffusion models) to approximate conditional data distributions without sharing raw data. Each client locally calibrates conformal scores reflecting its unique uncertainty, with global consistency maintained through federated aggregation.

Result: Experiments on real datasets show that Fed-CCP achieves more adaptive prediction sets compared to existing methods.

Conclusion: Fed-CCP successfully addresses the challenge of conditional coverage in federated settings by leveraging generative models, providing more accurate uncertainty quantification for heterogeneous client data while preserving privacy.

Abstract: Conformal Prediction (CP) provides distribution-free uncertainty quantification by constructing prediction sets that guarantee coverage of the true labels. This reliability makes CP valuable for high-stakes federated learning scenarios such as multi-center healthcare. However, standard CP assumes i.i.d. data, which is violated in federated settings where client distributions differ substantially. Existing federated CP methods address this by maintaining marginal coverage on each client, but such guarantees often fail to reflect input-conditional uncertainty. In this work, we propose Federated Conditional Conformal Prediction (Fed-CCP) via generative models, which aims for conditional coverage that adapts to local data heterogeneity. Fed-CCP leverages generative models, such as normalizing flows or diffusion models, to approximate conditional data distributions without requiring the sharing of raw data. This enables each client to locally calibrate conformal scores that reflect its unique uncertainty, while preserving global consistency through federated aggregation. Experiments on real datasets demonstrate that Fed-CCP achieves more adaptive prediction sets.

[947] Going with the Flow: Approximating Banzhaf Values via Graph Neural Networks

Benjamin Kempinski, Tal Kachman

Main category: cs.LG

TL;DR: A novel Graph Neural Network approach for approximating Banzhaf values in network flow games, achieving high-fidelity approximation with significant speedups and strong zero-shot generalization across different network configurations.

DetailsMotivation: Exact computation of Banzhaf values is intractable for large systems (>20 agents) due to exponential complexity, and Monte Carlo methods suffer from high sample complexity and lack knowledge transfer across network configurations.

Method: Framed as a graph-level prediction task using Graph Neural Networks (GNNs) - specifically GAT, GINE, and EdgeConv architectures - trained on 200,000 synthetic graphs varying in size, agent count, and edge probability.

Result: GNN models achieve high-fidelity Banzhaf value approximation with order-of-magnitude speedups compared to exact and sampling methods, and demonstrate strong zero-shot generalization to networks with different structural properties.

Conclusion: GNNs establish as a practical tool for scalable cooperative game-theoretic analysis of complex networked systems, enabling efficient Banzhaf value computation without retraining for new network configurations.

Abstract: Computing the Banzhaf value in network flow games is fundamental for quantifying agent influence in multi-agent systems, with applications ranging from cybersecurity to infrastructure planning. However, exact computation is intractable for systems with more than $\sim20$ agents due to exponential complexity $\mathcal{O}(2^m)$. While Monte Carlo sampling methods provide statistical estimates, they suffer from high sample complexity and cannot transfer knowledge across different network configurations, making them impractical for large-scale or dynamic systems. We present a novel learning-based approach using Graph Neural Networks (GNNs) to approximate Banzhaf values in cardinal network flow games. By framing the problem as a graph-level prediction task, our method learns generalisable patterns of agent influence directly from network topology and control structure. We conduct a comprehensive empirical study comparing three state-of-the-art GNN architectures-Graph Attention Networks (GAT), Graph Isomorphism Networks with Edge features (GINE), and EdgeConv-on a large-scale synthetic dataset of 200,000 graphs per configuration, varying in size (20-100 nodes), agent count (5-20), and edge probability (0.5-1.0). Our results demonstrate that trained GNN models achieve high-fidelity Banzhaf value approximation with order-of-magnitude speedups compared to exact and sampling-based methods. Most significantly, we show strong zero-shot generalisation: models trained on graphs of a specific size and topology accurately predict Banzhaf values for entirely new networks with different structural properties, without requiring retraining. This work establishes GNNs as a practical tool for scalable cooperative game-theoretic analysis of complex networked systems.

[948] SWIR-LightFusion: Multi-spectral Semantic Fusion of Synthetic SWIR with Thermal IR (LWIR/MWIR) and RGB

Muhammad Ishfaq Hussain, Ma Van Linh, Zubia Naz, Unse Fatima, Yeongmin Ko, Moongu Jeon

Main category: cs.LG

TL;DR: The paper introduces a method to synthetically generate SWIR-like images from LWIR data and proposes a multimodal fusion framework combining synthetic SWIR, LWIR, and RGB modalities using an encoder-decoder network with modality-specific encoders and softmax-gated fusion.

DetailsMotivation: Address limitations of conventional RGB and thermal infrared fusion in adverse visibility conditions by leveraging SWIR imaging's advantages, while overcoming the scarcity of public SWIR datasets.

Method: Synthetic generation of SWIR-like images from LWIR data using contrast enhancement techniques, followed by multimodal fusion with RGB and LWIR using optimized encoder-decoder architecture with modality-specific encoders and softmax-gated fusion head.

Result: Improved fused-image quality (contrast, edge definition, structural fidelity) while maintaining real-time performance across multiple benchmarks (M3FD, TNO, CAMEL, MSRS, RoadScene) and private RGB-MWIR-SWIR dataset.

Conclusion: The synthetic-SWIR-enhanced fusion framework shows substantial potential for real-world surveillance and autonomous systems applications.

Abstract: Enhancing scene understanding in adverse visibility conditions remains a critical challenge for surveillance and autonomous navigation systems. Conventional imaging modalities, such as RGB and thermal infrared (MWIR / LWIR), when fused, often struggle to deliver comprehensive scene information, particularly under conditions of atmospheric interference or inadequate illumination. To address these limitations, Short-Wave Infrared (SWIR) imaging has emerged as a promising modality due to its ability to penetrate atmospheric disturbances and differentiate materials with improved clarity. However, the advancement and widespread implementation of SWIR-based systems face significant hurdles, primarily due to the scarcity of publicly accessible SWIR datasets. In response to this challenge, our research introduces an approach to synthetically generate SWIR-like structural/contrast cues (without claiming spectral reproduction) images from existing LWIR data using advanced contrast enhancement techniques. We then propose a multimodal fusion framework integrating synthetic SWIR, LWIR, and RGB modalities, employing an optimized encoder-decoder neural network architecture with modality-specific encoders and a softmax-gated fusion head. Comprehensive experiments on public RGB-LWIR benchmarks (M3FD, TNO, CAMEL, MSRS, RoadScene) and an additional private real RGB-MWIR-SWIR dataset demonstrate that our synthetic-SWIR-enhanced fusion framework improves fused-image quality (contrast, edge definition, structural fidelity) while maintaining real-time performance. We also add fair trimodal baselines (LP, LatLRR, GFF) and cascaded trimodal variants of U2Fusion/SwinFusion under a unified protocol. The outcomes highlight substantial potential for real-world applications in surveillance and autonomous systems.

[949] Enhancing Time Series Forecasting through Selective Representation Spaces: A Patch Perspective

Xingjian Wu, Xiangfei Qiu, Hanyin Cheng, Zhengyu Li, Jilin Hu, Chenjuan Guo, Bin Yang

Main category: cs.LG

TL;DR: The paper proposes a Selective Representation Space (SRS) module that uses learnable selective patching and dynamic reassembly to adaptively select and shuffle patches from time series, improving forecasting performance of patch-based models.

DetailsMotivation: Conventional patching techniques partition time series into adjacent patches, creating fixed representation spaces that result in insufficiently expressive representations for time series forecasting.

Method: Proposed SRS module with learnable Selective Patching and Dynamic Reassembly techniques to adaptively select and shuffle patches from contextual time series. Implemented SRSNet combining SRS with MLP head.

Result: Achieves state-of-the-art performance on real-world datasets from multiple domains. Also enhances existing patch-based models as a plug-and-play module.

Conclusion: SRS module effectively constructs selective representation space to flexibly include informative patches, improving forecasting performance while being compatible with existing patch-based models.

Abstract: Time Series Forecasting has made significant progress with the help of Patching technique, which partitions time series into multiple patches to effectively retain contextual semantic information into a representation space beneficial for modeling long-term dependencies. However, conventional patching partitions a time series into adjacent patches, which causes a fixed representation space, thus resulting in insufficiently expressful representations. In this paper, we pioneer the exploration of constructing a selective representation space to flexibly include the most informative patches for forecasting. Specifically, we propose the Selective Representation Space (SRS) module, which utilizes the learnable Selective Patching and Dynamic Reassembly techniques to adaptively select and shuffle the patches from the contextual time series, aiming at fully exploiting the information of contextual time series to enhance the forecasting performance of patch-based models. To demonstrate the effectiveness of SRS module, we propose a simple yet effective SRSNet consisting of SRS and an MLP head, which achieves state-of-the-art performance on real-world datasets from multiple domains. Furthermore, as a novel plugin-and-play module, SRS can also enhance the performance of existing patch-based models. The resources are available at https://github.com/decisionintelligence/SRSNet.

[950] Reflections from Research Roundtables at the Conference on Health, Inference, and Learning (CHIL) 2025

Emily Alsentzer, Marie-Laure Charpignon, Bill Chen, Niharika D’Souza, Jason Fries, Yixing Jiang, Aparajita Kashyap, Chanwoo Kim, Simon Lee, Aishwarya Mandyam, Ashery Christopher Mbilinyi, Nikita Mehandru, Nitish Nagesh, Brighton Nuwagira, Emma Pierson, Arvind Pillai, Akane Sano, Tanveer Syeda-Mahmood, Shashank Yadav, Elias Adhanom, Muhammad Umar Afza, Amelia Archer, Suhana Bedi, Vasiliki Bikia, Trenton Chang, George H. Chen, Winston Chen, Erica Chiang, Edward Choi, Octavia Ciora, Paz Dozie-Nnamah, Shaza Elsharief, Matthew Engelhard, Ali Eshragh, Jean Feng, Josh Fessel, Scott Fleming, Kei Sen Fong, Thomas Frost, Soham Gadgil, Judy Gichoya, Leeor Hershkovich, Sujeong Im, Bhavya Jain, Vincent Jeanselme, Furong Jia, Qixuan Jin, Yuxuan Jin, Daniel Kapash, Geetika Kapoor, Behdokht Kiafar, Matthias Kleiner, Stefan Kraft, Annika Kumar, Daeun Kyung, Zhongyuan Liang, Joanna Lin, Qianchu Liu, Chang Liu, Hongzhou Luan, Chris Lunt, Leopoldo Julían Lechuga López, Matthew B. A. McDermott, Shahriar Noroozizadeh, Connor O’Brien, YongKyung Oh, Mixail Ota, Stephen Pfohl, Meagan Pi, Tanmoy Sarkar Pias, Emma Rocheteau, Avishaan Sethi, Toru Shirakawa, Anita Silver, Neha Simha, Kamile Stankeviciute, Max Sunog, Peter Szolovits, Shengpu Tang, Jialu Tang, Aaron Tierney, John Valdovinos, Byron Wallace, Will Ke Wang, Peter Washington, Jeremy Weiss, Daniel Wolfe, Emily Wong, Hye Sun Yun, Xiaoman Zhang, Xiao Yu Cindy Zhang, Hayoung Jeong, Kaveri A. Thakoor

Main category: cs.LG

TL;DR: The CHIL 2025 conference featured Research Roundtables with 19 chairs discussing 8 key topics at the intersection of machine learning and healthcare.

DetailsMotivation: To catalyze collaborative dialogue and address critical challenges in machine learning applications for healthcare through small-group discussions.

Method: Hosted 8 Research Roundtables moderated by senior and junior chairs, focusing on rigorous discussion, exploration of opportunities, and collective ideation.

Result: Successful roundtable sessions covering topics including explainability, uncertainty/bias/fairness, causality, domain adaptation, foundation models, small medical data learning, multimodal methods, and scalable healthcare solutions.

Conclusion: The Research Roundtables effectively fostered inclusive engagement and collaborative discussion to advance actionable directions in health machine learning research.

Abstract: The 6th Annual Conference on Health, Inference, and Learning (CHIL 2025), hosted by the Association for Health Learning and Inference (AHLI), was held in person on June 25-27, 2025, at the University of California, Berkeley, in Berkeley, California, USA. As part of this year’s program, we hosted Research Roundtables to catalyze collaborative, small-group dialogue around critical, timely topics at the intersection of machine learning and healthcare. Each roundtable was moderated by a team of senior and junior chairs who fostered open exchange, intellectual curiosity, and inclusive engagement. The sessions emphasized rigorous discussion of key challenges, exploration of emerging opportunities, and collective ideation toward actionable directions in the field. In total, eight roundtables were held by 19 roundtable chairs on topics of “Explainability, Interpretability, and Transparency,” “Uncertainty, Bias, and Fairness,” “Causality,” “Domain Adaptation,” “Foundation Models,” “Learning from Small Medical Data,” “Multimodal Methods,” and “Scalable, Translational Healthcare Solutions.”

[951] Machine Learning for Early Detection of Meningitis: Stacked Ensemble Learning with EHR Data

Han Ouyang, Jesse Hamilton, Saeed Amal

Main category: cs.LG

TL;DR: Ensemble learning approach using Random Forest, LightGBM, and DNN models achieves high AUC scores (0.9637 and 0.9472) for meningitis diagnosis from MIMIC-III data, simulating real-world ER scenarios.

DetailsMotivation: To develop an AI-driven diagnostic tool for meningitis that can be clinically useful in real-world emergency room settings, addressing the challenge of accurate and timely diagnosis.

Method: Used MIMIC-III database with 214 meningitis and 46,303 non-meningitis patients. Applied data preprocessing, feature selection including gender and high-risk ICD codes, and ensemble learning with three base models (Random Forest, LightGBM, DNN) stacked with logistic regression meta-model.

Result: Excellent performance with AUC of 0.9637 on Testing Set 1 and 0.9472 on Testing Set 2, demonstrating high diagnostic accuracy for meningitis.

Conclusion: The ensemble learning approach shows promising results for meningitis diagnosis and paves the way for future AI-driven diagnostic tools in clinical settings, though direct deployment remains challenging.

Abstract: We utilized a cohort of 214 meningitis patients and 46,303 non-meningitis patients from the MIMIC-III database. After extensive data preprocessing, which included ICD-based cohort selection, one-hot encoding of coding, and a two-stage feature selection process (for both the training set and the testing sets), clinically relevant features such as gender and high-risk ICD codes (including subarachnoid hemorrhage, secondary malignant neoplasm of the brain, and generalized epilepsy) are selected. Overall, these clinically reasonable and temporally adherent features provided excellent modeling performance. Three models (Random Forest, LightGBM, and Deep Neural Networks (DNN) are trained as base models for Ensemble Learning. Base model outputs are aggregated and stacked into a meta model (Logistic Regression) that uses the base model outputs as input values in training. Ultimately, soldier outputs (AUC of Testing Set 1: 0.9637, AUC of Testing Set 2: 0.9472) are obtained through ensemble learning. We created a challenging condition for diagnosing meningitis, simulating a real-world ER (Emergency Room) scenario to enhance clinical use in real-world applications. While directly deploying a diagnostic tool that clinicians can use is challenging, this paper paves the way for a potential future AI-driven diagnostic approach for meningitis using Ensemble Learning.

cs.MA

[952] Disaster Management in the Era of Agentic AI Systems: A Vision for Collective Human-Machine Intelligence for Augmented Resilience

Bo Li, Junwei Ma, Kai Yin, Yiming Xiao, Chia-Wei Hsu, Ali Mostafavi

Main category: cs.MA

TL;DR: Disaster Copilot is a multi-agent AI system that coordinates specialized agents for disaster management, integrating multi-modal data to provide real-time operational awareness and serve as an AI backbone for Disaster Digital Twins.

DetailsMotivation: Current disaster management faces challenges from fragmented data, siloed technologies, resource constraints, and institutional memory loss, which hinder timely and effective decision-making during disasters.

Method: A multi-agent AI architecture with a central orchestrator coordinating specialized sub-agents for predictive risk analytics, situational awareness, and impact assessment, using on-device orchestration for resource-limited environments.

Result: The system delivers holistic real-time operational pictures, advances Disaster Digital Twins from passive to active intelligent environments, and incorporates mechanisms to capture institutional knowledge.

Conclusion: Disaster Copilot offers a transformative vision for building more adaptive, data-driven, and resilient communities through collective human-machine intelligence, with a three-phased roadmap for implementation.

Abstract: The escalating frequency and severity of disasters routinely overwhelm traditional response capabilities, exposing critical vulnerability in disaster management. Current practices are hindered by fragmented data streams, siloed technologies, resource constraints, and the erosion of institutional memory, which collectively impede timely and effective decision making. This study introduces Disaster Copilot, a vision for a multi-agent artificial intelligence system designed to overcome these systemic challenges by unifying specialized AI tools within a collaborative framework. The proposed architecture utilizes a central orchestrator to coordinate diverse sub-agents, each specializing in critical domains such as predictive risk analytics, situational awareness, and impact assessment. By integrating multi-modal data, the system delivers a holistic, real-time operational picture and serve as the essential AI backbone required to advance Disaster Digital Twins from passive models to active, intelligent environments. Furthermore, it ensures functionality in resource-limited environments through on-device orchestration and incorporates mechanisms to capture institutional knowledge, mitigating the impact of staff turnover. We detail the system architecture and propose a three-phased roadmap emphasizing the parallel growth of technology, organizational capacity, and human-AI teaming. Disaster Copilot offers a transformative vision, fostering collective human-machine intelligence to build more adaptive, data-driven and resilient communities.

[953] Zero-Shot Coordination in Ad Hoc Teams with Generalized Policy Improvement and Difference Rewards

Rupal Nigam, Niket Parikh, Hamid Osooli, Mikihisa Yuasa, Jacob Heglund, Huy T. Tran

Main category: cs.MA

TL;DR: GPAT enables zero-shot ad hoc teaming by leveraging all pretrained policies through generalized policy improvement and difference rewards, achieving successful coordination with unseen teammates in simulated and real-world environments.

DetailsMotivation: Real-world multi-agent systems require ad hoc teaming where agents must coordinate with previously unseen teammates in zero-shot scenarios, but existing approaches either select policies based on inferred teammate models or use single robust policies.

Method: Proposes GPAT algorithm that uses generalized policy improvement and difference rewards for knowledge transfer between different teams in ad hoc multi-agent Markov decision processes.

Result: Successfully demonstrated zero-shot transfer to new teams in three simulated environments (cooperative foraging, predator-prey, Overcooked) and in real-world multi-robot settings.

Conclusion: GPAT provides an effective solution for zero-shot ad hoc teaming by leveraging all pretrained policies rather than selecting single policies or relying on teammate modeling.

Abstract: Real-world multi-agent systems may require ad hoc teaming, where an agent must coordinate with other previously unseen teammates to solve a task in a zero-shot manner. Prior work often either selects a pretrained policy based on an inferred model of the new teammates or pretrains a single policy that is robust to potential teammates. Instead, we propose to leverage all pretrained policies in a zero-shot transfer setting. We formalize this problem as an ad hoc multi-agent Markov decision process and present a solution that uses two key ideas, generalized policy improvement and difference rewards, for efficient and effective knowledge transfer between different teams. We empirically demonstrate that our algorithm, Generalized Policy improvement for Ad hoc Teaming (GPAT), successfully enables zero-shot transfer to new teams in three simulated environments: cooperative foraging, predator-prey, and Overcooked. We also demonstrate our algorithm in a real-world multi-robot setting.

[954] Heterogeneous Multi-Agent Task-Assignment with Uncertain Execution Times and Preferences

Qinshuang Wei, Vaibhav Srivastava, Vijay Gupta

Main category: cs.MA

TL;DR: The paper proposes a bandit algorithm for multi-agent task assignment with heterogeneous capabilities and preferences, analyzing regret for both exact and approximate solutions.

DetailsMotivation: Multi-agent task assignment problems with heterogeneous task preferences and capabilities are less studied compared to single-agent sequential task assignment.

Method: A bandit algorithm that repeatedly solves optimal task assignment problems, considering stochastic rewards, execution times, and resource consumption with unknown distributions.

Result: The algorithm is analyzed for achievable regret in two scenarios: when optimal task assignment can be solved exactly and when only approximate solutions are available.

Conclusion: The proposed approach addresses the challenge of multi-agent task assignment under uncertainty with heterogeneous capabilities and provides regret analysis for both exact and approximate solution cases.

Abstract: While sequential task assignment for a single agent has been widely studied, such problems in a multi-agent setting, where the agents have heterogeneous task preferences or capabilities, remain less well-characterized. We study a multi-agent task assignment problem where a central planner assigns recurring tasks to multiple members of a team over a finite time horizon. For any given task, the members have heterogeneous capabilities in terms of task completion times, task resource consumption (which can model variables such as energy or attention), and preferences in terms of the rewards they collect upon task completion. We assume that the reward, execution time, and resource consumption for each member to complete any task are stochastic with unknown distributions. The goal of the planner is to maximize the total expected reward that the team receives over the problem horizon while ensuring that the resource consumption required for any assigned task is within the capability of the agent. We propose and analyze a bandit algorithm for this problem. Since the bandit algorithm relies on solving an optimal task assignment problem repeatedly, we analyze the achievable regret in two cases: when we can solve the optimal task assignment exactly and when we can solve it only approximately.

[955] Prompt Optimization via Retrieved Reasoning Assets and Multi-Agent Analysis

Wonduk Seo, Juhyeon Lee, Junseo Koh, Hyunjin An, Jian Park, Seunghyun Lee, Haihua Chen, Yi Bu

Main category: cs.MA

TL;DR: MA-SAPO is a multi-agent framework that improves prompt optimization by coupling evaluation scores with structured reasoning, enabling more transparent and controllable prompt refinements.

DetailsMotivation: Existing prompt optimization methods treat evaluation as a black box, relying on trial-and-error without providing insights into why prompts succeed or fail, making them difficult to interpret and control.

Method: A two-stage framework: Reasoning Phase where agents collaboratively explain metric scores, diagnose weaknesses, and synthesize refinements as reusable reasoning assets; Test Phase where agents retrieve these assets to analyze optimized prompts and apply evidence-grounded edits.

Result: Experiments on HelpSteer1/2 benchmarks show consistent improvements over single-pass prompting, retrieval-augmented baselines, and prior multi-agent strategies.

Conclusion: MA-SAPO effectively turns evaluation signals into interpretable reasoning chains, producing more transparent, auditable, and controllable prompt refinements.

Abstract: Prompt optimization has emerged as an effective alternative to retraining for improving the performance of Large Language Models (LLMs). However, most existing approaches treat evaluation as a black box, relying solely on numerical scores while offering limited insight into why a prompt succeeds or fails. They also depend heavily on trial-and-error refinements, which are difficult to interpret and control. In this paper, we introduce MA-SAPO, a Multi-Agent framework for Score-Aware Prompt Optimization. Compared to prior methods, MA-SAPO explicitly couples evaluation outcomes with structured reasoning to guide systematic edits. The framework specifically consists of two stages: during the Reasoning Phase, agents collaboratively explain metric scores, diagnose weaknesses, and synthesize targeted refinements that are stored as reusable reasoning assets; during the Test Phase, agents retrieve these assets to analyze optimized prompts and apply only evidence-grounded edits. By turning evaluation signals into interpretable reasoning chains, MA-SAPO produces prompt refinements that are more transparent, auditable, and controllable. Experiments on the HelpSteer1/2 benchmarks demonstrate consistent improvements over single-pass prompting, retrieval-augmented baselines, and prior multi-agent strategies, validating the effectiveness of our approach.

[956] DiRAC - Distributed Robot Awareness and Consensus

Uday Gopan, Manjari Kulkarni, Lakshasri S, Kashish Mittal, Sriram Radhakrishna, Aditya Naskar, Rameshwar DL

Main category: cs.MA

TL;DR: DiRAC is a scalable distributed framework for efficient task assignment and path planning in large robotic swarms using zone-partitioned architecture with dynamic leaders and force-based decentralized planning.

DetailsMotivation: To enable efficient coordination and collision-free operation in very large robotic swarms for industrial and logistics applications.

Method: Zone-partitioned architecture with dynamically elected leaders, tick-synchronized consensus protocol, and force-based decentralized planner for real-time collision resolution.

Result: Demonstrated architectural scalability and modular efficiency in ROS 2 simulations within warehouse environments.

Conclusion: DiRAC provides a foundation for real-world deployment in large-scale industrial and logistics domains with strong consistency and deterministic outcomes.

Abstract: DiRAC is a scalable, distributed framework designed to enable efficient task assignment and path planning in very large robotic swarms. It introduces a novel zone-partitioned architecture with dynamically elected leaders and a tick-synchronized consensus protocol that yields strong consistency and deterministic outcomes. For path planning, DiRAC uses a novel algorithm, a force-based decentralized planner for real-time collision resolution. Validated within ROS 2 middleware through preliminary simulation, DiRAC demonstrates architectural scalability and modular efficiency in simulated warehouse environments, laying the groundwork for real-world deployment in large-scale industrial and logistics domains.

[957] Lark: Biologically Inspired Neuroevolution for Multi-Stakeholder LLM Agents

Dheeraj Chintapalli, Rikhil Tanugula, Sunkalp Chandra

Main category: cs.MA

TL;DR: Lark is a biologically inspired decision-making framework combining LLM reasoning with evolutionary multi-agent systems, featuring plasticity, duplication/maturation, stakeholder aggregation, and compute awareness mechanisms to generate diverse strategies efficiently.

DetailsMotivation: To address verbosity and stakeholder trade-offs in decision-making systems while maintaining cost efficiency and transparency in strategy generation.

Method: Integrates four mechanisms: plasticity for concise adjustments, duplication/maturation for specialization, ranked-choice stakeholder aggregation with Borda scoring, and compute awareness via token-based penalties. Uses iterative evolutionary process with LLM-driven reasoning.

Result: In 30-round evaluation, Lark Full achieved mean rank of 2.55 and composite score of 29.4/50, finishing Top-3 in 80% of rounds at $0.016 per task. All four mechanisms contributed significantly, with ablation studies showing duplication/maturation had largest impact.

Conclusion: Lark provides a practical, compute-aware neuroevolutionary approach for stakeholder-aligned strategy generation that makes trade-offs transparent, serving as proof-of-concept for real-world validation studies.

Abstract: We present Lark, a biologically inspired decision-making framework that couples LLM-driven reasoning with an evolutionary, stakeholder-aware Multi-Agent System (MAS). To address verbosity and stakeholder trade-offs, we integrate four mechanisms: (i) plasticity, which applies concise adjustments to candidate solutions; (ii) duplication and maturation, which copy high-performing candidates and specialize them into new modules; (iii) ranked-choice stakeholder aggregation using influence-weighted Borda scoring; and (iv) compute awareness via token-based penalties that reward brevity. The system iteratively proposes diverse strategies, applies plasticity tweaks, simulates stakeholder evaluations, aggregates preferences, selects top candidates, and performs duplication/maturation while factoring compute cost into final scores. In a controlled evaluation over 30 rounds comparing 14 systems, Lark Full achieves a mean rank of 2.55 (95% CI [2.17, 2.93]) and a mean composite score of 29.4/50 (95% CI [26.34, 32.46]), finishing Top-3 in 80% of rounds while remaining cost competitive with leading commercial models ($0.016 per task). Paired Wilcoxon tests confirm that all four mechanisms contribute significantly as ablating duplication/maturation yields the largest deficit ({\Delta}Score = 3.5, Cohen’s d_z = 2.53, p < 0.001), followed by plasticity ({\Delta}Score = 3.4, d_z = 1.86), ranked-choice voting ({\Delta}Score = 2.4, d_z = 1.20), and token penalties ({\Delta}Score = 2.2, d_z = 1.63). Rather than a formal Markov Decision Process with constrained optimization, Lark is a practical, compute-aware neuroevolutionary loop that scales stakeholder-aligned strategy generation and makes trade-offs transparent through per-step metrics. Our work presents proof-of-concept findings and invites community feedback as we expand toward real-world validation studies.

[958] ReclAIm: A multi-agent framework for degradation-aware performance tuning of medical imaging AI

Eleftherios Tzanis, Michail E. Klontzas

Main category: cs.MA

TL;DR: ReclAIm is a multi-agent framework that autonomously monitors, evaluates, and fine-tunes medical image classification models using natural language interaction, eliminating the need for programming expertise.

DetailsMotivation: To ensure long-term reliability of AI models in clinical practice through continuous performance monitoring and corrective actions when degradation occurs.

Method: Built on a large language model core, the system operates entirely through natural language interaction and autonomously executes state-of-the-art fine-tuning procedures when performance degradation is detected.

Result: ReclAIm successfully maintained consistent performance across MRI, CT, and X-ray datasets, reducing performance gaps to within 1.5% of initial model results even after drops of up to -41.1%.

Conclusion: ReclAIm enables automated, continuous maintenance of medical imaging AI models in a user-friendly and adaptable manner that facilitates broader adoption in both research and clinical environments.

Abstract: Ensuring the long-term reliability of AI models in clinical practice requires continuous performance monitoring and corrective actions when degradation occurs. Addressing this need, this manuscript presents ReclAIm, a multi-agent framework capable of autonomously monitoring, evaluating, and fine-tuning medical image classification models. The system, built on a large language model core, operates entirely through natural language interaction, eliminating the need for programming expertise. ReclAIm successfully trains, evaluates, and maintains consistent performance of models across MRI, CT, and X-ray datasets. Once ReclAIm detects significant performance degradation, it autonomously executes state-of-the-art fine-tuning procedures that substantially reduce the performance gap. In cases with performance drops of up to -41.1% (MRI InceptionV3), ReclAIm managed to readjust performance metrics within 1.5% of the initial model results. ReclAIm enables automated, continuous maintenance of medical imaging AI models in a user-friendly and adaptable manner that facilitates broader adoption in both research and clinical environments.

[959] MiCRO for Multilateral Negotiations

David Aguilera-Luzon, Dave de Jonge, Javier Larrosa

Main category: cs.MA

TL;DR: The paper introduces a multilateral version of the MiCRO negotiation strategy, showing it outperforms ANAC competition winners and forms an empirical Nash equilibrium.

DetailsMotivation: Previous research showed MiCRO performed well in bilateral negotiations without complex modeling, but its generalization to multilateral settings was an open question that needed to be addressed.

Method: Developed a multilateral variant of MiCRO strategy and compared it against winners of Automated Negotiating Agents Competitions (2015, 2017, 2018). Conducted empirical game-theoretical analysis.

Result: The multilateral MiCRO variant outperforms the ANAC competition winners and forms an empirical Nash equilibrium.

Conclusion: The simple MiCRO strategy can be successfully extended to multilateral negotiations, achieving strong performance without complex opponent modeling or parameter tuning.

Abstract: Recently, a very simple new bilateral negotiation strategy called MiCRO was introduced that does not make use of any kind of opponent modeling or machine learning techniques and that does not require fine-tuning of any parameters. Despite its simplicity, it was shown that MiCRO performs similar to – or even better than – most state-of-the-art negotiation strategies. This lead its authors to argue that the benchmark domains on which negotiation algorithms are typically tested may be too simplistic. However, one question that was left open, was how MiCRO could be generalized to multilateral negotiations. In this paper we fill this gap by introducing a multilateral variant of MiCRO. We compare it with the winners of the Automated Negotiating Agents Competitions (ANAC) of 2015, 2017 and 2018 and show that it outperforms them. Furthermore, we perform an empirical game-theoretical analysis to show that our new version of MiCRO forms an empirical Nash equilibrium.

[960] Strategyproof Facility Location for Five Agents on a Circle using PCD

Ido Farjoun, Reshef Meir

Main category: cs.MA

TL;DR: The paper analyzes strategyproof facility location on a circle for 5 agents, finding a tight bound for the PCD mechanism and hypothesizing approximation ratios for general odd n.

DetailsMotivation: To understand the performance of proportional circle division (PCD) strategyproof mechanisms for facility location problems on circular domains, particularly focusing on the case with 5 agents.

Method: Systematically reduces the instance space size and applies standard optimization techniques to find and prove tight bounds for the PCD mechanism.

Result: Found a tight bound for the PCD strategyproof mechanism with 5 agents on a circle.

Conclusion: The PCD mechanism achieves optimal performance for 5 agents, and the authors hypothesize approximation ratios for general odd numbers of agents.

Abstract: We consider the strategyproof facility location problem on a circle. We focus on the case of 5 agents, and find a tight bound for the PCD strategyproof mechanism, which selects the reported location of an agent in proportion to the length of the arc in front of it. We methodically “reduce” the size of the instance space and then use standard optimization techniques to find and prove the bound is tight. Moreover we hypothesize the approximation ratio of PCD for general odd $n$.

[961] Asynchronous Agents with Perfect Recall: Model Reductions, Knowledge-Based Construction, and Model Checking for Coalitional Strategies

Dilian Gurov, Filip Jamroga, Wojciech Jamroga, Mateusz Kamiński, Damian Kurpiewski, Wojciech Penczek, Teofil Sidoruk

Main category: cs.MA

TL;DR: This paper presents two advances for model checking strategic abilities of agents with memory: extending partial-order reduction to memoryful agents and adapting Knowledge-Based Subset Construction for asynchronous multi-agent systems.

DetailsMotivation: Model checking of strategic abilities for agents with memory is notoriously difficult, with few existing solutions. The paper aims to address this challenging problem.

Method: 1) Extend partial-order reduction scheme to work for agents with memory (previously only for memoryless agents). 2) Adapt Knowledge-Based Subset Construction from synchronous to asynchronous multi-agent systems to preserve memoryful agent abilities. 3) Propose new execution semantics combining Concurrent Game Structures and Interleaved Interpreted Systems.

Result: The paper successfully shows that partial-order reduction preserves individual and coalitional abilities for memoryful agents, and adapts Knowledge-Based Subset Construction for asynchronous settings.

Conclusion: These contributions represent important steps towards solving the hard problem of model checking strategic abilities for agents with memory in multi-agent systems.

Abstract: Model checking of strategic abilities for agents with memory is a notoriously hard problem, and very few attempts have been made to tackle it. In this paper, we present two important steps towards this goal. First, we take the partial-order reduction scheme that was recently proved to preserve individual and coalitional abilities of memoryless agents, and show that it also works for agents with memory. Secondly, we take the Knowledge-Based Subset Construction, that was recently studied for synchronous concurrent games, and adapt it to preserve abilities of memoryful agents in asynchronous MAS. On the way, we also propose a new execution semantics for strategies in asynchronous MAS, that combines elements of Concurrent Game Structures and Interleaved Interpreted Systems in a natural and intuitive way.

[962] First Field-Trial Demonstration of L4 Autonomous Optical Network for Distributed AI Training Communication: An LLM-Powered Multi-AI-Agent Solution

Yihao Zhang, Qizhi Qiu, Xiaomin Liu, Dianxuan Fu, Xingyu Liu, Leyan Fei, Yuming Cheng, Lilin Yi, Weisheng Hu, Qunbi Zhuge

Main category: cs.MA

TL;DR: First cross-domain cross-layer level-4 autonomous optical network achieved using multi-AI-agent system with 98% task completion rate in field trials.

DetailsMotivation: To create a fully autonomous optical network that can operate across different domains and layers without human intervention.

Method: Implemented a multi-AI-agent system for distributed AI training lifecycle management in optical networks.

Result: Field trials showed ~98% task completion rate, which is 3.2x higher than single agents using state-of-the-art LLMs.

Conclusion: Multi-AI-agent systems significantly outperform single-agent approaches in autonomous optical network operations.

Abstract: We demonstrate the first cross-domain cross-layer level-4 autonomous optical network via a multi-AI-agent system. Field trials show ~98% task completion rate across the distributed AI training lifecycle-3.2x higher than single agents using state-of-the-art LLMs.

[963] Sequence Modeling for N-Agent Ad Hoc Teamwork

Caroline Wang, Di Yang Shi, Elad Liebman, Ishan Durugkar, Arrasy Rahman, Peter Stone

Main category: cs.MA

TL;DR: A transformer-based method for N-agent ad hoc teamwork that outperforms existing approaches in StarCraft II tasks with better sample efficiency and generalization.

DetailsMotivation: Existing independent learning methods like POAM fail to capture inter-agent dynamics essential for effective collaboration in N-agent ad hoc teamwork scenarios.

Method: Centralized transformer-based approach that incorporates historical observations and actions of all controlled agents to handle varying team sizes and unknown teammates.

Result: MAT-NAHT outperforms POAM in StarCraft II tasks, achieving superior sample efficiency and generalization without needing auxiliary agent-modeling objectives.

Conclusion: Transformer-based centralized methods are effective for N-agent ad hoc teamwork, handling variable team sizes and unknown teammates better than independent learning approaches.

Abstract: N-agent ad hoc teamwork (NAHT) is a newly introduced challenge in multi-agent reinforcement learning, where controlled subteams of varying sizes must dynamically collaborate with varying numbers and types of unknown teammates without pre-coordination. The existing learning algorithm (POAM) considers only independent learning for its flexibility in dealing with a changing number of agents. However, independent learning fails to fully capture the inter-agent dynamics essential for effective collaboration. Based on our observation that transformers deal effectively with sequences with varying lengths and have been shown to be highly effective for a variety of machine learning problems, this work introduces a centralized, transformer-based method for N-agent ad hoc teamwork. Our proposed approach incorporates historical observations and actions of all controlled agents, enabling optimal responses to diverse and unseen teammates in partially observable environments. Empirical evaluation on a StarCraft II task demonstrates that MAT-NAHT outperforms POAM, achieving superior sample efficiency and generalization, without auxiliary agent-modeling objectives.

[964] COMPASS: Cooperative Multi-Agent Persistent Monitoring using Spatio-Temporal Attention Network

Xingjian Zhang, Yizhuo Wang, Guillaume Sartoretti

Main category: cs.MA

TL;DR: COMPASS is a multi-agent reinforcement learning framework for persistent monitoring of moving targets using decentralized agents with spatio-temporal attention and Gaussian Process modeling.

DetailsMotivation: Persistent monitoring of dynamic targets is essential in applications like disaster response, environmental sensing, and wildlife conservation, where mobile agents must continuously gather information under uncertainty.

Method: Models environment as a graph, uses decentralized agents with shared spatio-temporal attention network, employs Gaussian Processes for target dynamics modeling, and trains with centralized value estimation and decentralized policy execution under adaptive rewards.

Result: Extensive experiments show COMPASS consistently outperforms strong baselines in uncertainty reduction, target coverage, and coordination efficiency across dynamic multi-target scenarios.

Conclusion: COMPASS provides an effective MARL framework for persistent monitoring that enables decentralized agents to efficiently monitor multiple moving targets through structured reasoning and uncertainty-aware planning.

Abstract: Persistent monitoring of dynamic targets is essential in real-world applications such as disaster response, environmental sensing, and wildlife conservation, where mobile agents must continuously gather information under uncertainty. We propose COMPASS, a multi-agent reinforcement learning (MARL) framework that enables decentralized agents to persistently monitor multiple moving targets efficiently. We model the environment as a graph, where nodes represent spatial locations and edges capture topological proximity, allowing agents to reason over structured layouts and revisit informative regions as needed. Each agent independently selects actions based on a shared spatio-temporal attention network that we design to integrate historical observations and spatial context. We model target dynamics using Gaussian Processes (GPs), which support principled belief updates and enable uncertainty-aware planning. We train COMPASS using centralized value estimation and decentralized policy execution under an adaptive reward setting. Our extensive experiments demonstrate that COMPASS consistently outperforms strong baselines in uncertainty reduction, target coverage, and coordination efficiency across dynamic multi-target scenarios.

[965] Audit the Whisper: Detecting Steganographic Collusion in Multi-Agent LLMs

Om Tailor

Main category: cs.MA

TL;DR: Audit the Whisper is a comprehensive framework for detecting covert coordination among LLM agents with theoretical guarantees, benchmark design, and reproducibility infrastructure.

DetailsMotivation: Multi-agent LLM deployments are increasingly used in critical workflows, but covert coordination can silently undermine trust and social welfare. Existing audits lack theoretical guarantees and reproducibility.

Method: Developed channel-capacity analysis with mutual-information thresholds, created ColludeBench-v0 with configurable covert schemes, and built calibrated auditing pipeline with multiple detection techniques.

Result: Achieved state-of-the-art detection power at fixed false-positive rates across multiple benchmarks, revealing fairness-driven colluders invisible to mutual information alone.

Conclusion: The framework provides reproducible, theoretically-grounded detection of covert coordination with practical deployment capabilities for external auditors.

Abstract: Multi-agent deployments of large language models (LLMs) are increasingly embedded in market, allocation, and governance workflows, yet covert coordination among agents can silently erode trust and social welfare. Existing audits are dominated by heuristics that lack theoretical guarantees, struggle to transfer across tasks, and seldom ship with the infrastructure needed for independent replication. We introduce Audit the Whisper, a conference-grade research artifact that spans theory, benchmark design, detection, and reproducibility. Our contributions are: (i) a channel-capacity analysis showing how interventions such as paraphrase, rate limiting, and role permutation impose quantifiable capacity penalties-operationalised via paired-run Kullback–Leibler diagnostics-that tighten mutual-information thresholds with finite-sample guarantees and full proofs; (ii) ColludeBench-v0, covering pricing, first-price auctions, peer review, and hosted Gemini/Groq APIs with configurable covert schemes, deterministic manifests, and reward instrumentation; and (iii) a calibrated auditing pipeline that fuses cross-run mutual information, permutation invariance, watermark variance, and fairness-aware acceptance bias, each tuned to a $10^{-3}$ false-positive budget and validated by 10k honest runs plus an e-value martingale. Across ColludeBench and external suites including Secret Collusion, CASE, Perfect Collusion Benchmark, and SentinelAgent, the union meta-test attains state-of-the-art power at fixed FPR while ablations surface price-of-auditing trade-offs and fairness-driven colluders invisible to MI alone. We release regeneration scripts, anonymized manifests, and documentation so that external auditors can reproduce every figure, satisfy double-blind requirements, and extend the framework with minimal effort.

[966] A Vision for Access Control in LLM-based Agent Systems

Xinfeng Li, Dong Huang, Jie Li, Hongyi Cai, Zhenhong Zhou, Wei Dong, XiaoFeng Wang, Yang Liu

Main category: cs.MA

TL;DR: Proposes Agent Access Control (AAC) as a new framework replacing traditional binary access control with dynamic, context-aware information flow governance for LLM-based agents.

DetailsMotivation: Traditional access control mechanisms are insufficient for LLM-based agents due to their autonomy and contextual complexity, requiring a paradigm shift from permission-based to information flow governance.

Method: AAC framework with two core modules: multi-dimensional contextual evaluation (assessing identity, relationships, scenarios, norms) and adaptive response formulation (using redaction, summarization, paraphrasing beyond simple allow/deny).

Result: A conceptual framework that reframes access control as dynamic information flow governance, powered by a dedicated AC reasoning engine.

Conclusion: AAC bridges human-like nuanced judgment with scalable AI safety, providing a new conceptual lens for trustworthy agent design research.

Abstract: The autonomy and contextual complexity of LLM-based agents render traditional access control (AC) mechanisms insufficient. Static, rule-based systems designed for predictable environments are fundamentally ill-equipped to manage the dynamic information flows inherent in agentic interactions. This position paper argues for a paradigm shift from binary access control to a more sophisticated model of information governance, positing that the core challenge is not merely about permission, but about governing the flow of information. We introduce Agent Access Control (AAC), a novel framework that reframes AC as a dynamic, context-aware process of information flow governance. AAC operates on two core modules: (1) multi-dimensional contextual evaluation, which assesses not just identity but also relationships, scenarios, and norms; and (2) adaptive response formulation, which moves beyond simple allow/deny decisions to shape information through redaction, summarization, and paraphrasing. This vision, powered by a dedicated AC reasoning engine, aims to bridge the gap between human-like nuanced judgment and scalable Al safety, proposing a new conceptual lens for future research in trustworthy agent design.

[967] Static Sandboxes Are Inadequate: Modeling Societal Complexity Requires Open-Ended Co-Evolution in LLM-Based Multi-Agent Simulations

Jinkun Chen, Sher Badshah, Xuemin Yu, Sijia Han, Jiechao Gao

Main category: cs.MA

TL;DR: The paper argues that current multi-agent simulations using LLMs are too static and limited, and proposes moving towards open-ended, co-evolving systems that can better model real-world complexity.

DetailsMotivation: Current LLM-powered multi-agent simulations are constrained by predefined tasks and rigid evaluation, preventing them from capturing the complexity of real-world societies. The authors want to move beyond these static paradigms.

Method: The paper critically reviews emerging LLM-multi-agent architectures, identifies key challenges (balancing stability/diversity, evaluating unexpected behaviors, scaling complexity), and introduces a new taxonomy for the field.

Result: The authors develop a research roadmap focused on open-endedness, continuous co-evolution, and building resilient, socially aligned AI ecosystems.

Conclusion: The community should move beyond static simulation paradigms and help develop the next generation of adaptive, socially-aware multi-agent systems that can evolve and reshape their environments in unpredictable ways.

Abstract: What if artificial agents could not just communicate, but also evolve, adapt, and reshape their worlds in ways we cannot fully predict? With llm now powering multi-agent systems and social simulations, we are witnessing new possibilities for modeling open-ended, ever-changing environments. Yet, most current simulations remain constrained within static sandboxes, characterized by predefined tasks, limited dynamics, and rigid evaluation criteria. These limitations prevent them from capturing the complexity of real-world societies. In this paper, we argue that static, task-specific benchmarks are fundamentally inadequate and must be rethought. We critically review emerging architectures that blend llm with multi-agent dynamics, highlight key hurdles such as balancing stability and diversity, evaluating unexpected behaviors, and scaling to greater complexity, and introduce a fresh taxonomy for this rapidly evolving field. Finally, we present a research roadmap centered on open-endedness, continuous co-evolution, and the development of resilient, socially aligned AI ecosystems. We call on the community to move beyond static paradigms and help shape the next generation of adaptive, socially-aware multi-agent simulations.

cs.MM

[968] Taming Modality Entanglement in Continual Audio-Visual Segmentation

Yuyang Hong, Qi Yang, Tao Zhang, Zili Wang, Zhaojin Fu, Kun Ding, Bin Fan, Shiming Xiang

Main category: cs.MM

TL;DR: The paper introduces Continual Audio-Visual Segmentation (CAVS) to address modality entanglement in fine-grained continual learning, proposing a Collision-based Multi-modal Rehearsal (CMR) framework with Multi-modal Sample Selection and Collision-based Sample Rehearsal mechanisms.

DetailsMotivation: Existing multi-modal continual learning methods focus on coarse-grained tasks and struggle with modality entanglement in fine-grained settings, particularly for audio-visual segmentation where sounding objects may be mislabeled as background.

Method: Proposes CMR framework with: 1) Multi-modal Sample Selection (MSS) to select samples with high modal consistency for rehearsal, and 2) Collision-based Sample Rehearsal (CSR) to increase rehearsal frequency for confusable co-occurring classes.

Result: The method significantly outperforms single-modal continual learning methods across three constructed audio-visual incremental scenarios, demonstrating effectiveness in addressing multi-modal semantic drift and co-occurrence confusion.

Conclusion: The proposed CAVS task and CMR framework effectively address fine-grained multi-modal continual learning challenges, particularly multi-modal semantic drift and co-occurrence confusion in audio-visual segmentation.

Abstract: Recently, significant progress has been made in multi-modal continual learning, aiming to learn new tasks sequentially in multi-modal settings while preserving performance on previously learned ones. However, existing methods mainly focus on coarse-grained tasks, with limitations in addressing modality entanglement in fine-grained continual learning settings. To bridge this gap, we introduce a novel Continual Audio-Visual Segmentation (CAVS) task, aiming to continuously segment new classes guided by audio. Through comprehensive analysis, two critical challenges are identified: 1) multi-modal semantic drift, where a sounding objects is labeled as background in sequential tasks; 2) co-occurrence confusion, where frequent co-occurring classes tend to be confused. In this work, a Collision-based Multi-modal Rehearsal (CMR) framework is designed to address these challenges. Specifically, for multi-modal semantic drift, a Multi-modal Sample Selection (MSS) strategy is proposed to select samples with high modal consistency for rehearsal. Meanwhile, for co-occurence confusion, a Collision-based Sample Rehearsal (CSR) mechanism is designed, allowing for the increase of rehearsal sample frequency of those confusable classes during training process. Moreover, we construct three audio-visual incremental scenarios to verify effectiveness of our method. Comprehensive experiments demonstrate that our method significantly outperforms single-modal continual learning methods.

[969] Nexus: An Omni-Perceptive And -Interactive Model for Language, Audio, And Vision

Che Liu, Yingji Zhang, Dong Zhang, Weijie Zhang, Chenggong Gong, Yu Lu, Shilin Zhou, Ziliang Gan, Ziao Wang, Haipang Wu, Ji Liu, André Freitas, Qifan Wang, Zenglin Xu, Rongjuncheng Zhang, Yong Dai

Main category: cs.MM

TL;DR: Proposes Nexus, an industry-level omni-modal LLM pipeline integrating auditory, visual, and linguistic modalities with modular architecture, lightweight training strategy, and audio synthesis pipeline to overcome tri-modal dataset limitations and computational costs.

DetailsMotivation: To address challenges in multi-modal AI including limited tri-modal datasets, high computational costs, and complex feature alignments across auditory, visual, and linguistic modalities.

Method: Three-component pipeline: 1) Modular framework for flexible encoder-LLM-decoder architectures, 2) Lightweight training strategy pre-training audio-language alignment on Qwen2.5-VL to avoid costly vision-specific pre-training, 3) Audio synthesis pipeline generating high-quality audio-text data from real-world scenarios.

Result: Nexus demonstrates superior performance in visual understanding over Qwen2.5-VL-7B, better accuracy in English Spoken QA than MiniCPM-o2.6-7B, outstanding ASR performance in real-world tests, outperforms Qwen2-Audio-Instruct-7B in speech-to-text translation, comparable to backbone vocoders in text-to-speech, and enhanced tri-modal alignment.

Conclusion: The proposed omni-modal pipeline effectively integrates three modalities with efficient training strategy, achieving state-of-the-art performance across multiple tasks while demonstrating that audio modality enhances vision-language alignment.

Abstract: This work proposes an industry-level omni-modal large language model (LLM) pipeline that integrates auditory, visual, and linguistic modalities to overcome challenges such as limited tri-modal datasets, high computational costs, and complex feature alignments. Our pipeline consists of three main components: First, a modular framework enabling flexible configuration of various encoder-LLM-decoder architectures. Second, a lightweight training strategy that pre-trains audio-language alignment on the state-of-the-art vision-language model Qwen2.5-VL, thus avoiding the costly pre-training of vision-specific modalities. Third, an audio synthesis pipeline that generates high-quality audio-text data from diverse real-world scenarios, supporting applications such as Automatic Speech Recognition and Speech-to-Speech chat. To this end, we introduce an industry-level omni-modal LLM, Nexus. Extensive experiments validate the efficacy of our pipeline, yielding the following key findings:(1) In the visual understanding task, Nexus exhibits superior performance compared with its backbone model - Qwen2.5-VL-7B, validating the efficiency of our training strategy. (2) Within the English Spoken Question-Answering task, the model achieves better accuracy than the same-period competitor (i.e, MiniCPM-o2.6-7B) in the LLaMA Q. benchmark. (3) In our real-world ASR testset, Nexus achieves outstanding performance, indicating its robustness in real scenarios. (4) In the Speech-to-Text Translation task, our model outperforms Qwen2-Audio-Instruct-7B. (5) In the Text-to-Speech task, based on pretrained vocoder (e.g., Fishspeech1.4 or CosyVoice2.0), Nexus is comparable to its backbone vocoder on Seed-TTS benchmark. (6) An in-depth analysis of tri-modal alignment reveals that incorporating the audio modality enhances representational alignment between vision and language.

[970] VGGSounder: Audio-Visual Evaluations for Foundation Models

Daniil Zverev, Thaddäus Wiedemer, Ameya Prabhu, Matthias Bethge, Wieland Brendel, A. Sophia Koepke

Main category: cs.MM

TL;DR: VGGSounder is a re-annotated, multi-label test set that addresses limitations in VGGSound for evaluating audio-visual foundation models, providing detailed modality annotations and a new modality confusion metric.

DetailsMotivation: VGGSound dataset has limitations including incomplete labeling, overlapping classes, and misaligned modalities that distort evaluations of audio-visual models.

Method: Created VGGSounder by comprehensively re-annotating VGGSound with multi-label annotations, detailed modality-specific labels, and introduced a modality confusion metric to analyze performance degradation.

Result: VGGSounder enables precise analysis of modality-specific performance and reveals model limitations through the modality confusion metric.

Conclusion: VGGSounder provides a more reliable benchmark for evaluating audio-visual foundation models by addressing VGGSound’s limitations and enabling detailed modality analysis.

Abstract: The emergence of audio-visual foundation models underscores the importance of reliably assessing their multi-modal understanding. The VGGSound dataset is commonly used as a benchmark for evaluation audio-visual classification. However, our analysis identifies several limitations of VGGSound, including incomplete labelling, partially overlapping classes, and misaligned modalities. These lead to distorted evaluations of auditory and visual capabilities. To address these limitations, we introduce VGGSounder, a comprehensively re-annotated, multi-label test set that extends VGGSound and is specifically designed to evaluate audio-visual foundation models. VGGSounder features detailed modality annotations, enabling precise analyses of modality-specific performance. Furthermore, we reveal model limitations by analysing performance degradation when adding another input modality with our new modality confusion metric.

eess.AS

[971] AsyncVoice Agent: Real-Time Explanation for LLM Planning and Reasoning

Yueqian Lin, Zhengmian Hu, Jayakumar Subramanian, Qinsi Wang, Nikos Vlassis, Hai “Helen” Li, Yiran Chen

Main category: eess.AS

TL;DR: AsyncVoice Agent enables real-time human-AI collaboration by decoupling streaming LLM backend from voice frontend, allowing users to interrupt and steer model reasoning with 600x latency reduction.

DetailsMotivation: Current interfaces lack real-time verbalization and user barge-in capabilities, preventing effective understanding and interaction with AI reasoning processes.

Method: Asynchronous architecture that separates streaming LLM backend from conversational voice frontend, enabling parallel narration and inference with interruptible user interaction.

Result: Reduces interaction latency by more than 600x compared to monolithic baselines while maintaining high fidelity and competitive task accuracy.

Conclusion: Enables two-way dialogue with model’s thought process, creating more effective, steerable, and trustworthy human-AI systems for high-stakes tasks.

Abstract: Effective human-AI collaboration on complex reasoning tasks requires that users understand and interact with the model’s process, not just receive an output. However, the monolithic text from methods like Chain-of-Thought (CoT) prevents this, as current interfaces lack real-time verbalization and robust user barge-in. We present AsyncVoice Agent, a system whose asynchronous architecture decouples a streaming LLM backend from a conversational voice frontend. This design allows narration and inference to run in parallel, empowering users to interrupt, query, and steer the model’s reasoning process at any time. Objective benchmarks show this approach reduces interaction latency by more than 600x compared to monolithic baselines while ensuring high fidelity and competitive task accuracy. By enabling a two-way dialogue with a model’s thought process, AsyncVoice Agent offers a new paradigm for building more effective, steerable, and trustworthy human-AI systems for high-stakes tasks.

[972] Audio-Visual Speech Enhancement for Spatial Audio - Spatial-VisualVoice and the MAVE Database

Danielle Yaffe, Ferdinand Campe, Prachi Sharma, Dorothea Kolossa, Boaz Rafaely

Main category: eess.AS

TL;DR: A multi-channel audio-visual speech enhancement framework that combines spatial cues from microphone arrays with visual information to enhance target speakers in noisy environments, particularly effective at low SNR levels.

DetailsMotivation: Address the gap in AVSE methods tailored for spatial audio enhancement under low-SNR conditions, which is important for augmented reality applications where visual features remain immune to acoustic noise.

Method: Multi-channel AVSE framework based on VisualVoice that leverages spatial cues from microphone arrays and visual information. Also introduced MAVe database containing multi-channel audio-visual signals in controlled room conditions across various SNR levels.

Result: Consistently achieves significant gains in SI-SDR, STOI, and PESQ metrics, especially at low SNRs. Binaural signal analysis confirms preservation of spatial cues and intelligibility.

Conclusion: The proposed multi-channel audio-visual approach effectively enhances spatial audio under challenging low-SNR conditions while maintaining spatial information important for augmented reality applications.

Abstract: Audio-visual speech enhancement (AVSE) has been found to be particularly useful at low signal-to-noise (SNR) ratios due to the immunity of the visual features to acoustic noise. However, a significant gap exists in AVSE methods tailored to enhance spatial audio under low-SNR conditions. The latter is of growing interest with augmented reality applications. To address this gap, we present a multi-channel AVSE framework based on VisualVoice that leverages spatial cues from microphone arrays and visual information for enhancing the target speaker in noisy environments. We also introduce MAVe, a novel database containing multi-channel audio-visual signals in controlled, reproducible room conditions across a wide range of SNR levels. Experiments demonstrate that the proposed method consistently achieves significant gains in SI-SDR, STOI, and PESQ, particularly in low SNRs. Binaural signal analysis further confirms the preservation of spatial cues and intelligibility.

[973] Audio dequantization using instantaneous frequency

Vojtěch Kovanda, Pavel Rajmic

Main category: eess.AS

TL;DR: PHADQ is a phase-aware audio dequantization method that maintains temporal continuity of sinusoidal components and avoids energy loss artifacts common with l1-based approaches.

DetailsMotivation: To address energy loss artifacts and temporal discontinuity issues in audio dequantization that are commonly encountered with traditional l1-based regularization methods.

Method: Uses a phase-aware regularizer adapted from audio inpainting to maintain temporal continuity of sinusoidal components in time-frequency representation.

Result: Evaluated using objective metrics SDR and PEMO-Q ODG, showing improved performance over l1-based approaches.

Conclusion: PHADQ effectively handles audio dequantization while preserving temporal continuity and avoiding energy loss artifacts.

Abstract: We present a dequantization method that employs a phase-aware regularizer, originally successfully applied in an audio inpainting problem. The method maintains the temporal continuity of sinusoidal components in the audio signal time-frequency representation and avoids the energy loss artifacts commonly encountered with l1-based regularization approaches. The proposed method is called the Phase-Aware Audio Dequantizer (PHADQ). The methods are evaluated using the objective metric SDR and PEMO-Q ODG.

[974] SAC: Neural Speech Codec with Semantic-Acoustic Dual-Stream Quantization

Wenxi Chen, Xinsheng Wang, Ruiqi Yan, Yushen Chen, Zhikang Niu, Ziyang Ma, Xiquan Li, Yuzhe Liang, Hanlin Wen, Shunshun Yin, Ming Tao, Xie Chen

Main category: eess.AS

TL;DR: SAC is a neural speech codec with semantic-acoustic dual-stream quantization that balances high-quality reconstruction with semantically rich representations for speech language models.

DetailsMotivation: Existing speech codecs struggle to balance high-quality reconstruction with semantically rich representations, limiting their effectiveness in both generative and understanding tasks.

Method: Proposes SAC with semantic-acoustic dual-stream quantization that disentangles semantic and acoustic modeling into two dedicated streams, allowing each to be optimized for its respective role.

Result: SAC achieves strong reconstruction performance across diverse bitrates under clean and noisy conditions, with superior perceptual quality and intelligibility. It substantially outperforms state-of-the-art codecs in semantic representation, comparable to self-supervised learning continuous embeddings.

Conclusion: The dual-stream design effectively disentangles speech components, offering new potential for controllable speech applications.

Abstract: Speech codecs that convert continuous speech signals into discrete tokens have become essential for speech language models (SLMs). However, existing codecs struggle to balance high-quality reconstruction with semantically rich representations, limiting their effectiveness in both generative and understanding tasks. In this work, we propose SAC, a neural speech codec with semantic-acoustic dual-stream quantization. By disentangling semantic and acoustic modeling into two dedicated streams, SAC enables each to be optimized for its respective role. Comprehensive evaluations show that SAC achieves strong reconstruction performance across diverse bitrates under both clean and noisy conditions, with particularly high scores on UTMOS and WER, demonstrating superior perceptual quality and intelligibility. Moreover, SAC substantially outperforms state-of-the-art codecs in semantic representation, achieving a level comparable to that of self-supervised learning (SSL) continuous embeddings. Finally, our analysis of speech disentanglement highlights the effectiveness of the dual-stream design, offering new potential for controllable speech applications.

[975] Adaptive Deterministic Flow Matching for Target Speaker Extraction

Tsun-An Hsieh, Minje Kim

Main category: eess.AS

TL;DR: AD-FlowTSE introduces adaptive step size flow matching for target speaker extraction, using mixture ratio-aware initialization for more efficient and accurate speech extraction.

DetailsMotivation: Existing generative TSE methods with fixed step sizes are inefficient across varying noise conditions. The paper aims to improve efficiency by adapting to mixture composition and noise levels.

Method: Formulates TSE within flow matching paradigm with flow between background and source distributions, using mixture ratio-aware initialization and adaptive step sizes based on noise conditions.

Result: Achieves strong TSE performance with as few as one reverse step, and incorporating auxiliary mixture ratio estimation further improves target speech accuracy.

Conclusion: Aligning transport path with mixture composition and adapting step size to noise conditions enables efficient and accurate target speaker extraction.

Abstract: Generative target speaker extraction (TSE) methods often produce more natural outputs than predictive models. Recent work based on diffusion or flow matching (FM) typically relies on a small, fixed number of reverse steps with a fixed step size. We introduce Adaptive Discriminative Flow Matching TSE (AD-FlowTSE), which extracts the target speech using an adaptive step size. We formulate TSE within the FM paradigm but, unlike prior FM-based speech enhancement and TSE approaches that transport between the mixture (or a normal prior) and the clean-speech distribution, we define the flow between the background and the source, governed by the mixing ratio (MR) of the source and background that creates the mixture. This design enables MR-aware initialization, where the model starts at an adaptive point along the background-source trajectory rather than applying the same reverse schedule across all noise levels. Experiments show that AD-FlowTSE achieves strong TSE with as few as a single step, and that incorporating auxiliary MR estimation further improves target speech accuracy. Together, these results highlight that aligning the transport path with the mixture composition and adapting the step size to noise conditions yields efficient and accurate TSE.

[976] Towards Real-Time Generative Speech Restoration with Flow-Matching

Tsun-An Hsieh, Sebastian Braun

Main category: eess.AS

TL;DR: A low-latency real-time generative speech restoration system using flow-matching achieves 20ms latency for denoising, dereverberation, and restoration tasks with only 5 sampling steps.

DetailsMotivation: Most generative models for speech enhancement operate offline with high latency, making them unsuitable for streaming applications that require real-time processing.

Method: Proposes a causal flow-matching architecture without time-downsampling, exploring architectural variations and sampling strategies for efficient training and inference.

Result: Achieves 20ms total latency suitable for real-time communication, maintains high quality with only 5 NFEs, and shows causal FM models favor few-step reverse sampling.

Conclusion: Flow-matching enables low-latency real-time generative speech restoration with high quality and efficient inference, outperforming adversarial-loss-based approaches.

Abstract: Generative models have shown robust performance on speech enhancement and restoration tasks, but most prior approaches operate offline with high latency, making them unsuitable for streaming applications. In this work, we investigate the feasibility of a low-latency, real-time generative speech restoration system based on flow-matching (FM). Our method tackles diverse real-world tasks, including denoising, dereverberation, and generative restoration. The proposed causal architecture without time-downsampling achieves introduces an total latency of only 20 ms, suitable for real-time communication. In addition, we explore a broad set of architectural variations and sampling strategies to ensure effective training and efficient inference. Notably, our flow-matching model maintains high enhancement quality with only 5 number of function evaluations (NFEs) during sampling, achieving similar performance as when using ~20 NFEs under the same conditions. Experimental results indicate that causal FM-based models favor few-step reverse sampling, and smaller backbones degrade with longer reverse trajectories. We further show a side-by-side comparison of FM to typical adversarial-loss-based training for the same model architecture.

[977] AnyRIR: Robust Non-intrusive Room Impulse Response Estimation in the Wild

Kyung Yun Lee, Nils Meyer-Kahlen, Karolina Prawda, Vesa Välimäki, Sebastian J. Schlecht

Main category: eess.AS

TL;DR: AnyRIR is a non-intrusive method that uses music as excitation signal to estimate room impulse responses in noisy environments, formulated as L1-norm regression in time-frequency domain with IRLS and LSMR optimization.

DetailsMotivation: To address the problem of RIR estimation in noisy, uncontrolled environments where conventional deconvolution fails due to non-stationary sounds like speech or footsteps.

Method: Uses music as excitation signal instead of dedicated test signal, formulates RIR estimation as L1-norm regression in time-frequency domain, solved with Iterative Reweighted Least Squares (IRLS) and Least-Squares Minimal Residual (LSMR) methods.

Result: Outperforms L2-based and frequency-domain deconvolution methods under noisy scenarios and codec mismatch in both simulated and measured data experiments.

Conclusion: Enables robust RIR estimation for AR/VR and related applications by exploiting sparsity of non-stationary noise to suppress its influence.

Abstract: We address the problem of estimating room impulse responses (RIRs) in noisy, uncontrolled environments where non-stationary sounds such as speech or footsteps corrupt conventional deconvolution. We propose AnyRIR, a non-intrusive method that uses music as the excitation signal instead of a dedicated test signal, and formulate RIR estimation as an L1-norm regression in the time-frequency domain. Solved efficiently with Iterative Reweighted Least Squares (IRLS) and Least-Squares Minimal Residual (LSMR) methods, this approach exploits the sparsity of non-stationary noise to suppress its influence. Experiments on simulated and measured data show that AnyRIR outperforms L2-based and frequency-domain deconvolution, under in-the-wild noisy scenarios and codec mismatch, enabling robust RIR estimation for AR/VR and related applications.

[978] A Self-Attention-Driven Deep Denoiser Model for Real Time Lung Sound Denoising in Noisy Environments

Samiul Based Shuvo, Syed Samiul Alam, Taufiq Hasan

Main category: eess.AS

TL;DR: Proposed Uformer model for lung sound denoising using CNN encoder, Transformer encoder, and CNN decoder architecture, achieving significant SNR improvement in noisy clinical settings.

DetailsMotivation: Lung auscultation is valuable for respiratory disease diagnosis but sounds are contaminated in real-world clinical settings, and conventional denoising models are impractical due to spectral overlap complexities from diverse noise sources.

Method: Uformer model with three modules: CNN encoder for feature extraction, Transformer encoder to capture long-range dependencies and enhance LS features, and CNN decoder to generate denoised signals. Ablation study performed for optimal architecture.

Result: Model evaluated on lung sounds with synthetic and real-world noises (-12 dB to 15 dB SNR). Achieved average SNR improvement of 16.51 dB with -12 dB signals, and 19.31 dB average improvement outperforming existing models with ambient noise and fewer parameters.

Conclusion: Uformer is robust and generalized for assisting respiratory condition monitoring based on qualitative and quantitative findings.

Abstract: Objective: Lung auscultation is a valuable tool in diagnosing and monitoring various respiratory diseases. However, lung sounds (LS) are significantly affected by numerous sources of contamination, especially when recorded in real-world clinical settings. Conventional denoising models prove impractical for LS denoising, primarily owing to spectral overlap complexities arising from diverse noise sources. To address this issue, we propose a specialized deep-learning model (Uformer) for lung sound denoising. Methods: The proposed Uformer model is constituted of three modules: a Convolutional Neural Network (CNN) encoder module, dedicated to extracting latent features; a Transformer encoder module, employed to further enhance the encoding of unique LS features and effectively capture intricate long-range dependencies; and a CNN decoder module, employed to generate the denoised signals. An ablation study was performed in order to find the most optimal architecture. Results: The performance of the proposed Uformer model was evaluated on lung sounds induced with different types of synthetic and real-world noises. Lung sound signals of -12 dB to 15 dB signal-to-noise ratio (SNR) were considered in testing experiments. The proposed model showed an average SNR improvement of 16.51 dB when evaluated with -12 dB LS signals. Our end-to-end model, with an average SNR improvement of 19.31 dB, outperforms the existing model when evaluated with ambient noise and fewer parameters. Conclusion: Based on the qualitative and quantitative findings in this study, it can be stated that Uformer is robust and generalized to be used in assisting the monitoring of respiratory conditions.

[979] BINAQUAL: A Full-Reference Objective Localization Similarity Metric for Binaural Audio

Davoud Shariat Panah, Dan Barry, Alessandro Ragano, Jan Skoglund, Andrew Hines

Main category: eess.AS

TL;DR: BINAQUAL is a full-reference objective metric for assessing localization similarity in binaural audio that adapts AMBIQUAL from ambisonics to binaural domain, showing strong correlation with subjective tests.

DetailsMotivation: Spatial audio enhances immersion in VR/AR, gaming, and cinema, but processes like compression can alter localization cues. Subjective tests are costly and time-consuming, creating need for objective metrics.

Method: Adapts AMBIQUAL metric from ambisonics audio format to binaural domain, evaluated across five research questions covering sound source locations, angle interpolations, speaker layouts, audio degradations, and content diversity.

Result: BINAQUAL effectively differentiates subtle spatial variations and correlates strongly with subjective listening tests, demonstrating reliable performance for binaural localization quality assessment.

Conclusion: BINAQUAL provides a robust benchmark for ensuring spatial accuracy in binaural audio processing, enabling improved objective evaluations in immersive audio applications.

Abstract: Spatial audio enhances immersion in applications such as virtual reality, augmented reality, gaming, and cinema by creating a three-dimensional auditory experience. Ensuring the spatial fidelity of binaural audio is crucial, given that processes such as compression, encoding, or transmission can alter localization cues. While subjective listening tests like MUSHRA remain the gold standard for evaluating spatial localization quality, they are costly and time-consuming. This paper introduces BINAQUAL, a full-reference objective metric designed to assess localization similarity in binaural audio recordings. BINAQUAL adapts the AMBIQUAL metric, originally developed for localization quality assessment in ambisonics audio format to the binaural domain. We evaluate BINAQUAL across five key research questions, examining its sensitivity to variations in sound source locations, angle interpolations, surround speaker layouts, audio degradations, and content diversity. Results demonstrate that BINAQUAL effectively differentiates between subtle spatial variations and correlates strongly with subjective listening tests, making it a reliable metric for binaural localization quality assessment. The proposed metric provides a robust benchmark for ensuring spatial accuracy in binaural audio processing, paving the way for improved objective evaluations in immersive audio applications.

[980] SongBloom: Coherent Song Generation via Interleaved Autoregressive Sketching and Diffusion Refinement

Chenyu Yang, Shuai Wang, Hangting Chen, Wei Tan, Jianwei Yu, Haizhou Li

Main category: eess.AS

TL;DR: SongBloom is a novel framework for full-length song generation that uses an interleaved paradigm of autoregressive sketching and diffusion-based refinement to create coherent, high-quality music with balanced global structure and local fidelity.

DetailsMotivation: Existing methods struggle to balance global coherence with local fidelity in song generation, resulting in outputs that lack musicality or suffer from incoherent progression and mismatched lyrics.

Method: Uses an autoregressive diffusion model that gradually extends musical sketches from short to long and refines details from coarse to fine-grained through an interleaved generation paradigm that integrates semantic and acoustic context.

Result: Outperforms existing methods across both subjective and objective metrics and achieves performance comparable to state-of-the-art commercial music generation platforms.

Conclusion: SongBloom effectively addresses the challenge of generating coherent, high-quality full-length songs by combining the strengths of autoregressive and diffusion models through an interleaved generation approach.

Abstract: Generating music with coherent structure, harmonious instrumental and vocal elements remains a significant challenge in song generation. Existing language models and diffusion-based methods often struggle to balance global coherence with local fidelity, resulting in outputs that lack musicality or suffer from incoherent progression and mismatched lyrics. This paper introduces $\textbf{SongBloom}$, a novel framework for full-length song generation that leverages an interleaved paradigm of autoregressive sketching and diffusion-based refinement. SongBloom employs an autoregressive diffusion model that combines the high fidelity of diffusion models with the scalability of language models. Specifically, it gradually extends a musical sketch from short to long and refines the details from coarse to fine-grained. The interleaved generation paradigm effectively integrates prior semantic and acoustic context to guide the generation process. Experimental results demonstrate that SongBloom outperforms existing methods across both subjective and objective metrics and achieves performance comparable to the state-of-the-art commercial music generation platforms. Audio samples are available on our demo page: https://cypress-yang.github.io/SongBloom_demo. The code and model weights have been released on https://github.com/Cypress-Yang/SongBloom .

[981] Post-training for Deepfake Speech Detection

Wanying Ge, Xin Wang, Xuechen Liu, Junichi Yamagishi

Main category: eess.AS

TL;DR: Post-training approach adapts SSL models for deepfake speech detection using large multilingual dataset, achieving strong robustness and outperforming state-of-the-art detectors.

DetailsMotivation: To bridge the gap between general pre-training and domain-specific fine-tuning for deepfake speech detection.

Method: Post-training SSL models on large-scale multilingual dataset (56K+ hours genuine speech, 18K+ hours artifacts) across 100+ languages, then fine-tuning on Deepfake-Eval-2024.

Result: Post-trained models show strong robustness and generalization to unseen deepfake speech, consistently surpassing existing state-of-the-art detectors.

Conclusion: Post-training effectively adapts SSL models for deepfake detection, with models and code publicly available.

Abstract: We introduce a post-training approach that adapts self-supervised learning (SSL) models for deepfake speech detection by bridging the gap between general pre-training and domain-specific fine-tuning. We present AntiDeepfake models, a series of post-trained models developed using a large-scale multilingual speech dataset containing over 56,000 hours of genuine speech and 18,000 hours of speech with various artifacts in over one hundred languages. Experimental results show that the post-trained models already exhibit strong robustness and generalization to unseen deepfake speech. When they are further fine-tuned on the Deepfake-Eval-2024 dataset, these models consistently surpass existing state-of-the-art detectors that do not leverage post-training. Model checkpoints and source code are available online.

[982] Test-Time Training for Speech Enhancement

Avishkar Behera, Riya Ann Easow, Venkatesh Parvathala, K. Sri Rama Murty

Main category: eess.AS

TL;DR: This paper applies Test-Time Training (TTT) to Speech Enhancement using a Y-shaped architecture with self-supervised auxiliary tasks, enabling dynamic adaptation to new noise conditions without labeled data during inference.

DetailsMotivation: To address challenges of unpredictable noise conditions and domain shifts in speech enhancement that traditional methods struggle with.

Method: A Y-shaped architecture combining main speech enhancement with self-supervised auxiliary tasks (noise-augmented signal reconstruction, masked spectrogram prediction) that adapts during inference time without labeled data.

Result: Consistent improvements across speech quality metrics on both synthetic and real-world datasets, outperforming baseline models.

Conclusion: TTT is effective for speech enhancement, providing insights for future research in adaptive and robust speech processing systems.

Abstract: This paper introduces a novel application of Test-Time Training (TTT) for Speech Enhancement, addressing the challenges posed by unpredictable noise conditions and domain shifts. This method combines a main speech enhancement task with a self-supervised auxiliary task in a Y-shaped architecture. The model dynamically adapts to new domains during inference time by optimizing the proposed self-supervised tasks like noise-augmented signal reconstruction or masked spectrogram prediction, bypassing the need for labeled data. We further introduce various TTT strategies offering a trade-off between adaptation and efficiency. Evaluations across synthetic and real-world datasets show consistent improvements across speech quality metrics, outperforming the baseline model. This work highlights the effectiveness of TTT in speech enhancement, providing insights for future research in adaptive and robust speech processing.

[983] Guitar Tone Morphing by Diffusion-based Model

Kuan-Yu Chen, Kuan-Lin Chen, Yu-Chieh Yu, Jian-Jiun Ding

Main category: eess.AS

TL;DR: This paper explores learning-based approaches for guitar tone morphing, introducing a simpler spherical interpolation method using Music2Latent that outperforms LoRA fine-tuning.

DetailsMotivation: Modeling and transforming electric guitar tones is important in MIR due to the instrument's rich tone and expressive flexibility, enabling musicians to explore new textures and personalize performances through smooth tone transitions.

Method: The study compares LoRA fine-tuning for limited data with a simpler spherical interpolation approach using Music2Latent for tone morphing.

Result: The spherical interpolation method using Music2Latent yields significantly better results than LoRA fine-tuning, generating smoother and more natural tone transitions.

Conclusion: The proposed spherical interpolation architecture is a practical and efficient tool for music production and real-time audio effects, offering superior performance over more complex fine-tuning approaches.

Abstract: In Music Information Retrieval (MIR), modeling and transforming the tone of musical instruments, particularly electric guitars, has gained increasing attention due to the richness of the instrument tone and the flexibility of expression. Tone morphing enables smooth transitions between different guitar sounds, giving musicians greater freedom to explore new textures and personalize their performances. This study explores learning-based approaches for guitar tone morphing, beginning with LoRA fine-tuning to improve the model performance on limited data. Moreover, we introduce a simpler method, named spherical interpolation using Music2Latent. It yields significantly better results than the more complex fine-tuning approach. Experiments show that the proposed architecture generates smoother and more natural tone transitions, making it a practical and efficient tool for music production and real-time audio effects.

eess.IV

[984] Lung Cancer Classification from CT Images Using ResNet

Olajumoke O. Adekunle, Joseph D. Akinyemi, Khadijat T. Ladoja, Olufade F. W. Onifade

Main category: eess.IV

TL;DR: A novel deep learning approach using ResNet50 achieves 98.8% accuracy for multi-class lung cancer classification from CT images, significantly outperforming previous methods.

DetailsMotivation: Current automated systems for lung cancer classification from CT images have insufficient predictive efficacy for clinical adoption, and existing research mainly focuses on binary classification rather than distinguishing between different cancer subtypes.

Method: Used a pre-trained ResNet50 model with custom layers added on top, trained on 10,200 lung CT images from LC25000 dataset, validated on 2,550 images, and tested on 2,250 images. Applied meticulous hyperparameter fine-tuning for optimization.

Result: Achieved remarkable test accuracy of 98.8% for three-class classification (two malignant subtypes and one benign), representing a notable enhancement over prior models on the same dataset.

Conclusion: The proposed deep learning approach demonstrates superior performance in multi-class lung cancer classification from CT images, showing potential for improved clinical adoption of automated diagnostic systems.

Abstract: Lung cancer, a malignancy originating in lung tissues, is commonly diagnosed and classified using medical imaging techniques, particularly computed tomography (CT). Despite the integration of machine learning and deep learning methods, the predictive efficacy of automated systems for lung cancer classification from CT images remains below the desired threshold for clinical adoption. Existing research predominantly focuses on binary classification, distinguishing between malignant and benign lung nodules. In this study, a novel deep learning-based approach is introduced, aimed at an improved multi-class classification, discerning various subtypes of lung cancer from CT images. Leveraging a pre-trained ResNet model, lung tissue images were classified into three distinct classes, two of which denote malignancy and one benign. Employing a dataset comprising 15,000 lung CT images sourced from the LC25000 histopathological images, the ResNet50 model was trained on 10,200 images, validated on 2,550 images, and tested on the remaining 2,250 images. Through the incorporation of custom layers atop the ResNet architecture and meticulous hyperparameter fine-tuning, a remarkable test accuracy of 98.8% was recorded. This represents a notable enhancement over the performance of prior models on the same dataset.

[985] Time-Embedded Algorithm Unrolling for Computational MRI

Junno Yun, Yaşar Utku Alçalar, Mehmet Akçakaya

Main category: eess.IV

TL;DR: Proposes time-embedded algorithm unrolling for MRI reconstruction that uses time-dependent neural networks and learnable parameters instead of shared networks across iterations, improving reconstruction quality without significantly increasing computational complexity.

DetailsMotivation: Address limitations of traditional algorithm unrolling methods where shared proximal operator networks across iterations can cause artifacts/blurring, while using distinct networks increases parameters and risks overfitting.

Method: Introduces time-embedded unrolling by framing iteration-dependent proximal operations and Onsager corrections as time-embedded neural networks, with time-dependent learnable parameters for data fidelity operations.

Result: Extensive experiments on fastMRI dataset show effective reduction of aliasing artifacts and noise amplification, achieving state-of-the-art performance across various acceleration rates and datasets.

Conclusion: Time-embedding strategy enhances reconstruction quality in algorithm unrolling approaches without significant computational complexity increase, and extends to existing methods.

Abstract: Algorithm unrolling methods have proven powerful for solving the regularized least squares problem in computational magnetic resonance imaging (MRI). These approaches unfold an iterative algorithm with a fixed number of iterations, typically alternating between a neural network-based proximal operator for regularization, a data fidelity operation and auxiliary updates with learnable parameters. While the connection to optimization methods dictate that the proximal operator network should be shared across unrolls, this can introduce artifacts or blurring. Heuristically, practitioners have shown that using distinct networks may be beneficial, but this significantly increases the number of learnable parameters, making it challenging to prevent overfitting. To address these shortcomings, by taking inspirations from proximal operators with varying thresholds in approximate message passing (AMP) and the success of time-embedding in diffusion models, we propose a time-embedded algorithm unrolling scheme for inverse problems. Specifically, we introduce a novel perspective on the iteration-dependent proximal operation in vector AMP (VAMP) and the subsequent Onsager correction in the context of algorithm unrolling, framing them as a time-embedded neural network. Similarly, the scalar weights in the data fidelity operation and its associated Onsager correction are cast as time-dependent learnable parameters. Our extensive experiments on the fastMRI dataset, spanning various acceleration rates and datasets, demonstrate that our method effectively reduces aliasing artifacts and mitigates noise amplification, achieving state-of-the-art performance. Furthermore, we show that our time-embedding strategy extends to existing algorithm unrolling approaches, enhancing reconstruction quality without increasing the computational complexity significantly.

[986] Computer Navigated Spinal Surgery Using Magnetic Resonance Imaging and Augmented Reality

Songyuan Lu, Jingwen Hui, Jake Weeks, David B. Berry, Fanny Chapelin, Frank Talke

Main category: eess.IV

TL;DR: A radiation-free surgical navigation system using MRI and AR with ArUco markers for spinal pain management procedures, showing comparable accuracy to conventional fluoroscopy techniques.

DetailsMotivation: Current spinal pain management procedures like RFA and ESI use fluoroscopy for needle placement, exposing patients and physicians to harmful ionizing radiation.

Method: Combines MRI scans with fiducial ArUco marker-based AR tracking using a stereo camera, overlaying MRI images onto patients for real-time anatomical visualization during needle insertion.

Result: Dual-ArUco marker tracking improved needle insertion accuracy and reduced average misplacement distance compared to single-marker procedures, achieving comparable accuracy (2 mm average deviation) to conventional fluoroscopy techniques.

Conclusion: The radiation-free system demonstrates promise as an alternative to fluoroscopy for spinal navigation procedures by eliminating radiation exposure while maintaining procedural accuracy.

Abstract: Current spinal pain management procedures, such as radiofrequency ablation (RFA) and epidural steroid injection (ESI), rely on fluoroscopy for needle placement which exposes patients and physicians to ionizing radiation. In this paper, we investigate a radiation-free surgical navigation system for spinal pain management procedures that combines magnetic resonance imaging (MRI) with fiducial ArUco marker-based augmented reality (AR). High-resolution MRI scans of a lumbar spinal phantom were obtained and assembled as a surface mesh. Laplacian smoothing algorithms were then applied to smoothen the surface and improve the model fidelity. A commercially available stereo camera (ZED2) was used to track single or dual fiducial ArUco markers on the patient to determine the patient’s real-time pose. Custom AR software was applied to overlay the MRI image onto the patient, allowing the physician to see not only the outer surface of the patient but also the complete anatomy of the patient below the surface. Needle-insertion trials on a 3D-printed 3-vertebra phantom showed that dual-ArUco marker tracking increased the accuracy of needle insertions and reduced the average needle misplacement distance compared to single-ArUco marker procedures. The average needle misplacement is comparable to the average deviation of 2 mm for conventional epidural techniques using fluoroscopy. Our radiation-free system demonstrates promise to serve as an alternative to fluoroscopy by improving image-guided spinal navigation.

[987] FSAR-Cap: A Fine-Grained Two-Stage Annotated Dataset for SAR Image Captioning

Jinqi Zhang, Lamei Zhang, Bin Zou

Main category: eess.IV

TL;DR: FSAR-Cap is a large-scale SAR image captioning dataset with 14,480 images and 72,400 image-text pairs, built through a two-stage annotation strategy to address the scarcity of high-quality datasets in SAR image understanding.

DetailsMotivation: The development of Synthetic Aperture Radar (SAR) image captioning is limited by the scarcity of high-quality datasets, despite its crucial role in applications like military intelligence and urban planning.

Method: Built on the FAIR-CSAR detection dataset using a two-stage annotation strategy combining hierarchical template-based representation, manual verification and supplementation, and prompt standardization.

Result: FSAR-Cap provides richer fine-grained annotations, broader category coverage, and higher annotation quality compared to existing resources. Benchmarking with multiple encoder-decoder architectures verifies its effectiveness.

Conclusion: FSAR-Cap establishes a foundation for future research in SAR captioning and intelligent image interpretation by addressing the dataset scarcity problem with high-quality annotations.

Abstract: Synthetic Aperture Radar (SAR) image captioning enables scene-level semantic understanding and plays a crucial role in applications such as military intelligence and urban planning, but its development is limited by the scarcity of high-quality datasets. To address this, we present FSAR-Cap, a large-scale SAR captioning dataset with 14,480 images and 72,400 image-text pairs. FSAR-Cap is built on the FAIR-CSAR detection dataset and constructed through a two-stage annotation strategy that combines hierarchical template-based representation, manual verification and supplementation, prompt standardization. Compared with existing resources, FSAR-Cap provides richer fine-grained annotations, broader category coverage, and higher annotation quality. Benchmarking with multiple encoder-decoder architectures verifies its effectiveness, establishing a foundation for future research in SAR captioning and intelligent image interpretation.

[988] Dictionary-Based Deblurring for Unpaired Data

Alok Panigrahi, Jayaprakash Katual, Satish Mulleti

Main category: eess.IV

TL;DR: A dictionary learning approach for image deblurring that works across full, partial, and unsupervised settings by jointly estimating structured blur matrices and image dictionaries, requiring fewer training samples.

DetailsMotivation: Real-world paired blur-sharp image datasets are difficult to obtain, limiting existing deblurring methods' effectiveness and generalizability due to data scarcity.

Method: Joint estimation of structured blur matrix and high-resolution image dictionary using dictionary learning, evaluated across full supervision (paired data), partial supervision (unpaired data), and unsupervised learning (non-correspondence data).

Result: Superior performance compared to conventional coupled dictionary learning approaches on synthetically blurred CMU-Cornell iCoseg dataset and real-world FocusPath dataset, with accurate blur modeling and adaptive dictionary representation using fewer training samples.

Conclusion: The framework provides an efficient and robust solution for image deblurring in data-constrained scenarios by enabling accurate blur modeling and adaptive dictionary representation with notably smaller training requirements.

Abstract: Effective image deblurring typically relies on large and fully paired datasets of blurred and corresponding sharp images. However, obtaining such accurately aligned data in the real world poses a number of difficulties, limiting the effectiveness and generalizability of existing deblurring methods. To address this scarcity of data dependency, we present a novel dictionary learning based deblurring approach for jointly estimating a structured blur matrix and a high resolution image dictionary. This framework enables robust image deblurring across different degrees of data supervision. Our method is thoroughly evaluated across three distinct experimental settings: (i) full supervision involving paired data with explicit correspondence, (ii) partial supervision employing unpaired data with implicit relationships, and (iii) unsupervised learning using non-correspondence data where direct pairings are absent. Extensive experimental validation, performed on synthetically blurred subsets of the CMU-Cornell iCoseg dataset and the real-world FocusPath dataset, consistently shows that the proposed framework has superior performance compared to conventional coupled dictionary learning approaches. The results validate that our approach provides an efficient and robust solution for image deblurring in data-constrained scenarios by enabling accurate blur modeling and adaptive dictionary representation with a notably smaller number of training samples.

[989] AV1 Motion Vector Fidelity and Application for Efficient Optical Flow

Julien Zouein, Vibhoothi Vibhoothi, Anil Kokaram

Main category: eess.IV

TL;DR: AV1 motion vectors can replace traditional optical flow with high quality and efficiency, and using them as warm-start for RAFT reduces computation time by 4x with minimal accuracy loss.

DetailsMotivation: To leverage motion vectors from AV1 video codec as a computationally efficient substitute for resource-intensive traditional optical flow in computer vision pipelines.

Method: Compared motion vectors from AV1 and HEVC against ground-truth optical flow, analyzed impact of encoder settings, and used AV1 motion vectors as warm-start for RAFT deep learning optical flow method.

Result: AV1 motion vectors show high fidelity to ground-truth optical flow, and using them as warm-start for RAFT achieves 4x speedup in computation time with only minor increase in end-point error.

Conclusion: Motion vectors from compressed video can be practically reused for efficient motion-aware computer vision applications, offering significant computational benefits.

Abstract: This paper presents a comprehensive analysis of motion vectors extracted from AV1-encoded video streams and their application in accelerating optical flow estimation. We demonstrate that motion vectors from AV1 video codec can serve as a high-quality and computationally efficient substitute for traditional optical flow, a critical but often resource-intensive component in many computer vision pipelines. Our primary contributions are twofold. First, we provide a detailed comparison of motion vectors from both AV1 and HEVC against ground-truth optical flow, establishing their fidelity. In particular we show the impact of encoder settings on motion estimation fidelity and make recommendations about the optimal settings. Second, we show that using these extracted AV1 motion vectors as a “warm-start” for a state-of-the-art deep learning-based optical flow method, RAFT, significantly reduces the time to convergence while achieving comparable accuracy. Specifically, we observe a four-fold speedup in computation time with only a minor trade- off in end-point error. These findings underscore the potential of reusing motion vectors from compressed video as a practical and efficient method for a wide range of motion-aware computer vision applications.

[990] A Low-Complexity View Synthesis Distortion Estimation Method for 3D Video with Large Baseline Considerations

Chongyuan Bi, Jie Liang

Main category: eess.IV

TL;DR: A low-complexity, training-free method for estimating view synthesis distortion without actual rendering, using joint texture-depth classification and baseline distance compensation for large baseline configurations.

DetailsMotivation: Existing view synthesis distortion estimation methods are computationally intensive, require parameter training, or perform poorly in large baseline configurations, limiting real-time applications like interactive free-viewpoint video.

Method: Joint texture-depth classification separates images into stationary/non-stationary regions, uses baseline distance indicator for compensation, and employs region-based blending estimation to identify overlapping/single-view/disocclusion regions.

Result: Experiments on MPEG 3D video sequences show high accuracy and efficiency, especially in large baseline configurations, enabling more flexible camera arrangements.

Conclusion: The proposed method provides accurate synthesis quality prediction under challenging geometric configurations without requiring actual rendering or parameter training.

Abstract: Depth-image-based rendering is a key view synthesis algorithm in 3D video systems, which enables the synthesis of virtual views from texture images and depth maps. An efficient view synthesis distortion estimation model is critical for optimizing resource allocation in real-time applications such as interactive free-viewpoint video and 3D video streaming services. However, existing estimation methods are often computationally intensive, require parameter training, or performance poorly in challenging large baseline configurations. This paper presents a novel, low-complexity, and training-free method to accurately estimate the distortion of synthesized views without performing the actual rendering process. Key contributions include: (1) A joint texture-depth classification method that accurately separates texture image into locally stationary and non-stationary regions, which mitigates misclassifications by using texture-only methods. (2) A novel baseline distance indicator is designed for the compensation scheme for distortions caused by large baseline configurations. (3) A region-based blending estimation strategy that geometrically identifies overlapping, single-view, and mutual disocclusion regions, predicting distortion in synthesized views from two reference views with differing synthesis conditions. Experiments on standard MPEG 3D video sequences validate the proposed method’s high accuracy and efficiency, especially in large baseline configurations. This method enables more flexible camera arrangements in 3D content acquisition by accurately predicting synthesis quality under challenging geometric configurations.

[991] Segmenting infant brains across magnetic fields: Domain randomization and annotation curation in ultra-low field MRI

Vladyslav Zalevskyi, Dondu-Busra Bulut, Thomas Sanchez, Meritxell Bach Cuadra

Main category: eess.IV

TL;DR: A domain randomization framework improves segmentation of brain structures in ultra-low-field MRI by bridging the domain gap with high-field MRI through robust augmentation and annotation quality control.

DetailsMotivation: Early identification of neurodevelopmental disorders requires accurate brain structure segmentation in infants, which is challenging in ultra-low-field MRI due to poor image quality, rapid brain growth, and motion artifacts, despite its affordability and portability advantages.

Method: Proposed a domain randomization framework to bridge the domain gap between high-field and ultra-low-field MRI, using pre-training on whole-brain HF segmentations with DR, careful curation of training labels by removing misregistered annotations, and model fusion through majority voting.

Result: The approach achieved competitive performance in hippocampi and basal ganglia segmentation for the LISA challenge, demonstrating significant improvement in generalization to ULF data.

Conclusion: Combining robust augmentation with annotation quality control enables accurate segmentation in ultra-low-field MRI data, making it viable for neurodevelopmental disorder screening in low-resource settings.

Abstract: Early identification of neurodevelopmental disorders relies on accurate segmentation of brain structures in infancy, a task complicated by rapid brain growth, poor tissue contrast, and motion artifacts in pediatric MRI. These challenges are further exacerbated in ultra-low-field (ULF, 0.064~T) MRI, which, despite its lower image quality, offers an affordable, portable, and sedation-free alternative for use in low-resource settings. In this work, we propose a domain randomization (DR) framework to bridge the domain gap between high-field (HF) and ULF MRI in the context of the hippocampi and basal ganglia segmentation in the LISA challenge. We show that pre-training on whole-brain HF segmentations using DR significantly improves generalization to ULF data, and that careful curation of training labels, by removing misregistered HF-to-ULF annotations from training, further boosts performance. By fusing the predictions of several models through majority voting, we are able to achieve competitive performance. Our results demonstrate that combining robust augmentation with annotation quality control can enable accurate segmentation in ULF data. Our code is available at https://github.com/Medical-Image-Analysis-Laboratory/lisasegm

[992] AutoLungDx: A Hybrid Deep Learning Approach for Early Lung Cancer Diagnosis Using 3D Res-U-Net, YOLOv5, and Vision Transformers

Samiul Based Shuvo, Tasnia Binte Mamun

Main category: eess.IV

TL;DR: Proposed an automated end-to-end deep learning framework for early lung cancer detection in low-resource settings, achieving high accuracy in lung segmentation (98.82% dice score), nodule detection (0.76 mAP@50), and classification (93.57% accuracy).

DetailsMotivation: Early lung cancer detection is crucial but challenging in low-resource settings with limited access to medical resources and radiologists. Need for automated solutions to improve screening accuracy and efficiency.

Method: Three-stage framework: 1) Lung segmentation using 3D Res-U-Net, 2) Nodule detection using YOLO-v5, 3) Classification with Vision Transformer architecture. Evaluated on LUNA16 dataset.

Result: Achieved 98.82% lung segmentation dice score, 0.76 mAP@50 for nodule detection at low false-positive rate, and 93.57% classification accuracy (1.21% higher than state-of-the-art). Outperformed existing studies in all evaluation metrics.

Conclusion: The framework effectively segments lungs and detects/classifies nodules, particularly suitable for low-resource settings. Has potential to improve lung cancer screening accuracy and efficiency, leading to better patient outcomes.

Abstract: Lung cancer is a leading cause of cancer-related deaths worldwide, and early detection is crucial for improving patient outcomes. Nevertheless, early diagnosis of cancer is a major challenge, particularly in low-resource settings where access to medical resources and trained radiologists is limited. The objective of this study is to propose an automated end-to-end deep learning-based framework for the early detection and classification of lung nodules, specifically for low-resource settings. The proposed framework consists of three stages: lung segmentation using a modified 3D U-Net named 3D Res-U-Net, nodule detection using YOLO-v5, and classification with a Vision Transformer-based architecture. We evaluated the proposed framework on a publicly available dataset, LUNA16. The proposed framework’s performance was measured using the respective domain’s evaluation matrices. The proposed framework achieved a 98.82% lung segmentation dice score while detecting the lung nodule with 0.76 mAP@50 from the segmented lung, at a low false-positive rate. The performance of both networks of the proposed framework was compared with other studies and found to outperform them regarding segmentation and detection accuracy. Additionally, our proposed Vision transformer network obtained an accuracy of 93.57%, which is 1.21% higher than the state-of-the-art networks. Our proposed end-to-end deep learning-based framework can effectively segment lungs, and detect and classify lung nodules, specifically in low-resource settings with limited access to radiologists. The proposed framework outperforms existing studies regarding all the respective evaluation metrics. The proposed framework can potentially improve the accuracy and efficiency of lung cancer screening in low-resource settings, ultimately leading to better patient outcomes.

[993] Predicting Patient Recovery or Mortality Using Deep Neural Decision Tree and Forest

Mohammad Dehghani, Mohadeseh Zarei Ghobadi, Mobin Mohammadi, Diyana Tehrany Dehkordy

Main category: eess.IV

TL;DR: Deep neural decision forest model achieved 80% accuracy in predicting COVID-19 patient mortality using clinical data, outperforming other machine learning methods.

DetailsMotivation: Need to identify high-risk COVID-19 patients for effective resource allocation in emergency departments, especially during global health crises with limited medical services.

Method: Used patient data including COVID-19 diagnosis, demographics, health indicators, and occupational factors. Dataset split 80/20 for training/testing with stratified sampling. Compared 9 ML/DL methods including deep neural decision forest and deep neural decision tree.

Result: Deep neural decision forest consistently outperformed other models with 80% accuracy using only clinical data. Clinical data alone proved most accurate for mortality prediction.

Conclusion: Deep neural decision forest is a reliable predictor of COVID-19 patient mortality, and clinical data alone may be the most effective diagnostic tool for mortality prediction.

Abstract: Objective: Identifying patients at high risk of mortality is crucial for emergency physicians to allocate hospital resources effectively, particularly in regions with limited medical services. This need becomes even more pressing during global health crises that lead to significant morbidity and mortality. This study aimed to present the usability deep neural decision forest and deep neural decision tree to predict mortality among Coronavirus disease 2019 (COVID-19) patients. To this end, We used patient data encompassing Coronavirus disease 2019 diagnosis, demographics, health indicators, and occupational risk factors to analyze disease severity and outcomes. The dataset was partitioned using a stratified sampling method, ensuring that 80% was allocated for training and 20% for testing. Nine machine learning and deep learning methods were employed to build predictive models. The models were evaluated across all stages to determine their effectiveness in predicting patient outcomes. Results: Among the models, the deep neural decision forest consistently outperformed others. Results indicated that using only clinical data yielded an accuracy of 80% by deep neural decision forest, demonstrating it as a reliable predictor of patient mortality. Moreover, the results suggest that clinical data alone may be the most accurate diagnostic tool for predicting mortality.

[994] Adaptive Convolutional Neural Network for Image Super-resolution

Ziang Wu, Jinwei Xie, Xuanyu Zhang, Tao Wang, Yongjun Zhang, Qi Zhu, Chunwei Tian

Main category: eess.IV

TL;DR: ADSRNet is a heterogeneous parallel network for image super-resolution that uses two complementary sub-networks to capture diverse structural information and improve model robustness across different scenes.

DetailsMotivation: Convolutional neural networks can learn features automatically but face robustness challenges in varying scenes. Bigger architectural differences help extract more diversified structural information to strengthen super-resolution model robustness.

Method: ADSRNet uses a heterogeneous parallel network with two components: an upper network that enhances context information relations, kernel mapping salient information, and shallow-deep layer relations; and a lower symmetric network that enhances inter-layer relations to mine structural information complementary to the upper network.

Result: Experimental results show that ADSRNet is effective for image super-resolution, with codes available on GitHub.

Conclusion: The proposed ADSRNet successfully improves adaptability of super-resolution models for different scenes through its heterogeneous parallel architecture that captures complementary structural information.

Abstract: Convolutional neural networks can automatically learn features via deep network architectures and given input samples. However, the robustness of obtained models may face challenges in varying scenes. Bigger differences in network architecture are beneficial to extract more diversified structural information to strengthen the robustness of an obtained super-resolution model. In this paper, we proposed a adaptive convolutional neural network for image super-resolution (ADSRNet). To capture more information, ADSRNet is implemented by a heterogeneous parallel network. The upper network can enhance relation of context information, salient information relation of a kernel mapping and relations of shallow and deep layers to improve performance of image super-resolution. That can strengthen adaptability of an obtained super-resolution model for different scenes. The lower network utilizes a symmetric architecture to enhance relations of different layers to mine more structural information, which is complementary with a upper network for image super-resolution. The relevant experimental results show that the proposed ADSRNet is effective to deal with image resolving. Codes are obtained at https://github.com/hellloxiaotian/ADSRNet.

[995] JND-Guided Light-Weight Neural Pre-Filter for Perceptual Image Coding

Chenlong He, Zhijian Hao, Leilei Huang, Xiaoyang Zeng, Yibo Fan

Main category: eess.IV

TL;DR: This paper introduces FJNDF-Pytorch, a unified benchmark for frequency-domain JND-guided pre-filters, and proposes a lightweight CNN framework that achieves state-of-the-art compression efficiency with significantly reduced computational cost.

DetailsMotivation: Existing JND-guided pre-filter methods are computationally expensive and lack standardized benchmarks for fair comparison, limiting their practical application and evaluation.

Method: Developed FJNDF-Pytorch as a unified benchmark platform and proposed a complete learning framework for a novel lightweight CNN architecture optimized for frequency-domain JND-guided pre-filtering.

Result: The proposed method achieves state-of-the-art compression efficiency, outperforming competitors across multiple datasets and encoders. It requires only 7.15 GFLOPs for 1080p images (14.1% of recent lightweight networks’ cost).

Conclusion: This work provides a robust, state-of-the-art solution that excels in both performance and efficiency, supported by an open-source reproducible research platform.

Abstract: Just Noticeable Distortion (JND)-guided pre-filter is a promising technique for improving the perceptual compression efficiency of image coding. However, existing methods are often computationally expensive, and the field lacks standardized benchmarks for fair comparison. To address these challenges, this paper introduces a twofold contribution. First, we develop and open-source FJNDF-Pytorch, a unified benchmark for frequency-domain JND-Guided pre-filters. Second, leveraging this platform, we propose a complete learning framework for a novel, lightweight Convolutional Neural Network (CNN). Experimental results demonstrate that our proposed method achieves state-of-the-art compression efficiency, consistently outperforming competitors across multiple datasets and encoders. In terms of computational cost, our model is exceptionally lightweight, requiring only 7.15 GFLOPs to process a 1080p image, which is merely 14.1% of the cost of recent lightweight network. Our work presents a robust, state-of-the-art solution that excels in both performance and efficiency, supported by a reproducible research platform. The open-source implementation is available at https://github.com/viplab-fudan/FJNDF-Pytorch.

[996] Principled Feature Disentanglement for High-Fidelity Unified Brain MRI Synthesis

Jihoon Cho, Jonghye Woo, Jinah Park

Main category: eess.IV

TL;DR: HF-GAN is a unified framework for synthesizing missing MR sequences using principled feature disentanglement and dynamic fusion, achieving state-of-the-art performance and improving downstream segmentation tasks.

DetailsMotivation: Missing MR sequences in clinical practice lead to inconsistent analysis results, creating a need for reliable synthesis of multisequence MR images.

Method: Uses feature disentanglement with many-to-one stream for complementary features and parallel one-to-one streams for modality-specific information, integrated via channel attention-based fusion module (CAFF) and modality infuser.

Result: HF-GAN achieves state-of-the-art performance, with 2D slice-based framework outperforming leading 3D volumetric models, and substantially improves brain tumor segmentation when used for data imputation.

Conclusion: The framework demonstrates clinical relevance by effectively synthesizing missing MR sequences and enhancing downstream diagnostic tasks.

Abstract: Multisequence Magnetic Resonance Imaging (MRI) provides a more reliable diagnosis in clinical applications through complementary information across sequences. However, in practice, the absence of certain MR sequences is a common problem that can lead to inconsistent analysis results. In this work, we propose a novel unified framework for synthesizing multisequence MR images, called hybrid-fusion GAN (HF-GAN). The fundamental mechanism of this work is principled feature disentanglement, which aligns the design of the architecture with the complexity of the features. A powerful many-to-one stream is constructed for the extraction of complex complementary features, while utilizing parallel, one-to-one streams to process modality-specific information. These disentangled features are dynamically integrated into a common latent space by a channel attention-based fusion module (CAFF) and then transformed via a modality infuser to generate the target sequence. We validated our framework on public datasets of both healthy and pathological brain MRI. Quantitative and qualitative results show that HF-GAN achieves state-of-the-art performance, with our 2D slice-based framework notably outperforming a leading 3D volumetric model. Furthermore, the utilization of HF-GAN for data imputation substantially improves the performance of the downstream brain tumor segmentation task, demonstrating its clinical relevance.

[997] Accelerating MRI with Longitudinally-informed Latent Posterior Sampling

Yonatan Urman, Zachary Shah, Ashwin Kumar, Bruno P. Soares, Kawin Setsompop

Main category: eess.IV

TL;DR: A diffusion-model-based MRI reconstruction framework that uses prior scans to accelerate acquisition, eliminating the need for paired longitudinal training data.

DetailsMotivation: To leverage previous MRI scans for accelerated reconstruction while handling anatomical changes between sessions, addressing the lack of open-access longitudinal datasets with raw k-space data.

Method: Uses diffusion models trained on standalone images, treating all timepoints as same distribution. Integrates prior DICOM scans during inference to guide follow-up reconstruction without requiring paired training data.

Result: Outperforms both longitudinal and non-longitudinal baselines across accelerated Cartesian acquisitions, achieving up to 10% higher SSIM and 2 dB higher PSNR in similar regions without degrading dissimilar areas. Robust to anatomical changes and misregistration.

Conclusion: Prior scans can be effectively integrated with diffusion-based reconstruction to improve image quality and enable greater acceleration, without needing extensive paired training datasets.

Abstract: Purpose: To accelerate MRI acquisition by incorporating the previous scans of a subject during reconstruction. Although longitudinal imaging constitutes much of clinical MRI, leveraging previous scans is challenging due to the complex relationship between scan sessions, potentially involving substantial anatomical or pathological changes, and the lack of open-access datasets with both longitudinal pairs and raw k-space needed for training deep learning-based reconstruction models. Methods: We propose a diffusion-model-based reconstruction framework that eliminates the need for longitudinally paired training data. During training, we treat all scan timepoints as samples from the same distribution, therefore requiring only standalone images. At inference, our framework integrates a subject’s prior scan in magnitude DICOM format, which is readily available in clinical workflows, to guide reconstruction of the follow-up. To support future development, we introduce an open-access clinical dataset containing multi-session pairs including prior DICOMs and follow-up k-space. Results: Our method consistently outperforms both longitudinal and non-longitudinal baseline reconstruction methods across various accelerated Cartesian acquisition strategies. In imaging regions highly similar to the prior scan, we observe up to 10% higher SSIM and 2 dB higher PSNR, without degradation in dissimilar areas. Compared to longitudinal reconstruction baselines, our method demonstrates robustness to varying degrees of anatomical change and misregistration. Conclusion: We demonstrate that prior scans can be effectively integrated with state-of-the-art diffusion-based reconstruction methods to improve image quality and enable greater scan acceleration, without requiring an extensive longitudinally-paired training dataset.

[998] Beyond Uncertainty Quantification: Learning Uncertainty for Trust-Informed Neural Network Decisions - A Case Study in COVID-19 Classification

Hassan Gharoun, Mohammad Sadegh Khorshidi, Fang Chen, Amir H. Gandomi

Main category: eess.IV

TL;DR: Proposes an uncertainty-aware stacked neural network that reduces confidently incorrect predictions in high-stakes applications like medical diagnosis by learning when to trust predictions rather than relying on fixed confidence thresholds.

DetailsMotivation: Traditional uncertainty quantification methods use predefined confidence thresholds but fail to assess correctness of high-confidence predictions, leading to confidently incorrect predictions that erode trust in automated systems.

Method: Two-tier model: base model generates predictions with uncertainty estimates, meta-model learns to assign trust flags to distinguish confidently correct cases from those needing expert review.

Result: Significantly reduces confidently incorrect predictions compared to traditional threshold-based methods across multiple confidence thresholds and pre-trained architectures on COVIDx CXR-4 dataset.

Conclusion: Provides more trustworthy and efficient decision-support system for high-stakes domains by learning when predictions should be trusted rather than relying on fixed confidence thresholds.

Abstract: Reliable uncertainty quantification is critical in high-stakes applications, such as medical diagnosis, where confidently incorrect predictions can erode trust in automated decision-making systems. Traditional uncertainty quantification methods rely on a predefined confidence threshold to classify predictions as confident or uncertain. However, this approach assumes that predictions exceeding the threshold are trustworthy, while those below it are uncertain, without explicitly assessing the correctness of high-confidence predictions. As a result, confidently incorrect predictions may still occur, leading to misleading uncertainty assessments. To address this limitation, this study proposed an uncertainty-aware stacked neural network, which extends conventional uncertainty quantification by learning when predictions should be trusted. The framework consists of a two-tier model: the base model generates predictions with uncertainty estimates, while the meta-model learns to assign a trust flag, distinguishing confidently correct cases from those requiring expert review. The proposed approach is evaluated against the traditional threshold-based method across multiple confidence thresholds and pre-trained architectures using the COVIDx CXR-4 dataset. Results demonstrate that the proposed framework significantly reduces confidently incorrect predictions, offering a more trustworthy and efficient decision-support system for high-stakes domains.

[999] Using Randomized Nyström Preconditioners to Accelerate Variational Image Reconstruction

Tao Hong, Zhaoyi Xu, Jason Hu, Jeffrey A. Fessler

Main category: eess.IV

TL;DR: This paper proposes using randomized Nyström approximation to compute effective preconditioners for accelerating iterative image reconstruction, without requiring explicit matrix representations of forward models.

DetailsMotivation: Model-based iterative reconstruction faces challenges with large-scale, nonsmooth, and sometimes nonconvex minimization problems. Preconditioning can accelerate convergence but is difficult when explicit matrices are unavailable and computational efficiency is crucial.

Method: Adapts randomized Nyström approximation to compute preconditioners, leverages GPU platforms for on-the-fly computation, and proposes efficient approaches for applying preconditioners to problems with nonsmooth regularizers (wavelet, total variation, Hessian Schatten-norm).

Result: Numerical results on image deblurring, super-resolution with impulsive noise, and 2D computed tomography reconstruction demonstrate the efficiency and effectiveness of the proposed preconditioner.

Conclusion: The randomized Nyström-based preconditioner successfully accelerates image reconstruction while being computationally efficient and not requiring explicit matrix representations.

Abstract: Model-based iterative reconstruction plays a key role in solving inverse problems. However, the associated minimization problems are generally large-scale, nonsmooth, and sometimes even nonconvex, which present challenges in designing efficient iterative solvers. Preconditioning methods can significantly accelerate the convergence of iterative methods. In some applications, computing preconditioners on-the-fly is beneficial. Moreover, forward models in image reconstruction are typically represented as operators, and the corresponding explicit matrices are often unavailable, which brings additional challenges in designing preconditioners. Therefore, for practical use, computing and applying preconditioners should be computationally inexpensive. This paper adapts the randomized Nystr"om approximation to compute effective preconditioners that accelerate image reconstruction without requiring an explicit matrix for the forward model. We leverage modern GPU computational platforms to compute the preconditioner on-the-fly. Moreover, we propose efficient approaches for applying the preconditioners to problems with classical nonsmooth regularizers, i.e., wavelet, total variation, and Hessian Schatten-norm. Our numerical results on image deblurring, super-resolution with impulsive noise, and 2D computed tomography reconstruction illustrate the efficiency and effectiveness of the proposed preconditioner.

[1000] A Synthetic Data-Driven Radiology Foundation Model for Pan-tumor Clinical Diagnosis

Wenhui Lei, Hanyu Chen, Zitian Zhang, Luyang Luo, Qiong Xiao, Yannian Gu, Peng Gao, Yankai Jiang, Ci Wang, Guangtao Wu, Tongjia Xu, Yingjie Zhang, Pranav Rajpurkar, Xiaofan Zhang, Shaoting Zhang, Zhenning Wang

Main category: eess.IV

TL;DR: PASTA is a pan-tumor radiology foundation model trained on synthetic CT data that achieves state-of-the-art performance on 45/46 oncology tasks and improves radiologists’ efficiency and accuracy in clinical workflows.

DetailsMotivation: Overcome the scarcity of large-scale annotated medical imaging datasets due to privacy restrictions and high labeling costs for developing robust oncology foundation models.

Method: Developed PASTA-Gen synthetic data framework generating 30,000 3D CT scans with pixel-level lesion masks and structured reports across ten organ systems, then trained PASTA foundation model on this data.

Result: Achieved SOTA on 45/46 oncology tasks; PASTA-AID clinical system increased radiologists’ throughput by 11.1-25.1%, improved sensitivity by 17.0-31.4% and precision by 10.5-24.9%, reduced segmentation time by up to 78.2% and reporting time by up to 36.5%.

Conclusion: Established an end-to-end synthetic data-driven pipeline for pan-tumor research and clinical translation that narrows expertise gaps and enables less-experienced radiologists to approach expert-level performance.

Abstract: AI-assisted imaging made substantial advances in tumor diagnosis and management. However, a major barrier to developing robust oncology foundation models is the scarcity of large-scale, high-quality annotated datasets, which are limited by privacy restrictions and the high cost of manual labeling. To address this gap, we present PASTA, a pan-tumor radiology foundation model built on PASTA-Gen, a synthetic data framework that generated 30,000 3D CT scans with pixel-level lesion masks and structured reports of tumors across ten organ systems. Leveraging this resource, PASTA achieves state-of-the-art performance on 45 of 46 oncology tasks, including non-contrast CT tumor screening, lesion segmentation, structured reporting, tumor staging, survival prediction, and MRI-modality transfer. To assess clinical applicability, we developed PASTA-AID, a clinical decision support system, and ran a retrospective simulated clinical trial across two scenarios. For pan-tumor screening on plain CT with fixed reading time, PASTA-AID increased radiologists’ throughput by 11.1-25.1% and improved sensitivity by 17.0-31.4% and precision by 10.5-24.9%; additionally, in a diagnosis-aid workflow, it reduced segmentation time by up to 78.2% and reporting time by up to 36.5%. Beyond gains in accuracy and efficiency, PASTA-AID narrowed the expertise gap, enabling less-experienced radiologists to approach expert-level performance. Together, this work establishes an end-to-end, synthetic data-driven pipeline spanning data generation, model development, and clinical validation, thereby demonstrating substantial potential for pan-tumor research and clinical translation.

[1001] FetalCLIP: A Visual-Language Foundation Model for Fetal Ultrasound Image Analysis

Fadillah Maani, Numan Saeed, Tausifa Saleem, Zaid Farooq, Hussain Alasmawi, Werner Diehl, Ameera Mohammad, Gareth Waring, Saudabi Valappi, Leanne Bricker, Mohammad Yaqub

Main category: eess.IV

TL;DR: FetalCLIP is a vision-language foundation model for fetal ultrasound images that was pre-trained on 210,035 image-text pairs - the largest dataset of its kind. It outperforms baselines across multiple tasks including classification, gestational age estimation, CHD detection, and segmentation.

DetailsMotivation: Fetal ultrasound images are challenging for foundation models due to their complexity and scarcity of paired multimodal data. Existing models require substantial additional training and face limitations from data scarcity.

Method: Multimodal pre-training approach using a diverse dataset of 210,035 fetal ultrasound images paired with text, representing the largest paired dataset of its kind for foundation model development.

Result: FetalCLIP outperformed all baselines across key fetal ultrasound applications including classification, gestational age estimation, CHD detection, and fetal structure segmentation. It demonstrated remarkable generalizability and strong performance even with limited labeled data.

Conclusion: FetalCLIP effectively learns intricate anatomical features in fetal ultrasound images, producing robust representations for various downstream applications. The model will be released publicly to benefit the scientific community.

Abstract: Foundation models are becoming increasingly effective in the medical domain, offering pre-trained models on large datasets that can be readily adapted for downstream tasks. Despite progress, fetal ultrasound images remain a challenging domain for foundation models due to their inherent complexity, often requiring substantial additional training and facing limitations due to the scarcity of paired multimodal data. To overcome these challenges, here we introduce FetalCLIP, a vision-language foundation model capable of generating universal representation of fetal ultrasound images. FetalCLIP was pre-trained using a multimodal learning approach on a diverse dataset of 210,035 fetal ultrasound images paired with text. This represents the largest paired dataset of its kind used for foundation model development to date. This unique training approach allows FetalCLIP to effectively learn the intricate anatomical features present in fetal ultrasound images, resulting in robust representations that can be used for a variety of downstream applications. In extensive benchmarking across a range of key fetal ultrasound applications, including classification, gestational age estimation, congenital heart defect (CHD) detection, and fetal structure segmentation, FetalCLIP outperformed all baselines while demonstrating remarkable generalizability and strong performance even with limited labeled data. We plan to release the FetalCLIP model publicly for the benefit of the broader scientific community.

[1002] Geodesic Diffusion Models for Efficient Medical Image Enhancement

Teng Zhang, Hongxu Jiang, Kuang Gong, Wei Shao

Main category: eess.IV

TL;DR: Geodesic Diffusion Models (GDM) use optimal geodesic noise schedules based on Fisher-Rao metric to significantly improve training and sampling efficiency in diffusion models, achieving state-of-the-art results with as few as 6 steps.

DetailsMotivation: Traditional diffusion models use empirically chosen noise schedules that are inefficient, requiring many intermediate steps and resulting in high computational costs during training and sampling.

Method: Derived geodesic noise schedules corresponding to shortest paths in probability space under Fisher-Rao metric, enabling more efficient transformation between probability distributions.

Result: Achieved 20-30x faster training than DDPMs and 4-6x faster than Fast-DDPM, with 160-170x sampling speedup over DDPMs. State-of-the-art performance on CT denoising and MRI super-resolution with only 6 sampling steps.

Conclusion: GDM enables efficient model development and real-time clinical applications through optimal geometric scheduling, significantly reducing computational costs while maintaining high performance.

Abstract: Diffusion models generate data by learning to reverse a forward process, where samples are progressively perturbed with Gaussian noise according to a predefined noise schedule. From a geometric perspective, each noise schedule corresponds to a unique trajectory in probability space from the data distribution to a Gaussian prior. However, prior diffusion models rely on empirically chosen schedules that may not be optimal. This inefficiency necessitates many intermediate time steps, resulting in high computational costs during both training and sampling. To address this, we derive a family of geodesic noise schedules corresponding to the shortest paths in probability space under the Fisher-Rao metric. Based on these schedules, we propose Geodesic Diffusion Models (GDMs), which significantly improve training and sampling efficiency by minimizing the energy required to transform between probability distributions. This efficiency further enables sampling to start from an intermediate distribution in conditional image generation, achieving state-of-the-art results with as few as 6 steps. We evaluated GDM on two medical image enhancement tasks: CT image denoising and MRI image super-resolution. Experimental results show that GDM achieved state-of-the-art performance while reducing training time by 20- to 30-fold compared to Denoising Diffusion Probabilistic Models (DDPMs) and 4- to 6-fold compared to Fast-DDPM, and accelerating sampling by 160- to 170-fold and 1.6-fold, respectively. These gains support the use of GDM for efficient model development and real-time clinical applications. Our code is publicly available at: https://github.com/mirthAI/GDM-VE.

[1003] Convergent Complex Quasi-Newton Proximal Methods for Gradient-Driven Denoisers in Compressed Sensing MRI Reconstruction

Tao Hong, Zhaoyi Xu, Se Young Chun, Luis Hernandez-Garcia, Jeffrey A. Fessler

Main category: eess.IV

TL;DR: Proposes a complex quasi-Newton proximal method for faster CS MRI reconstruction using gradient-driven denoisers, with modified Hessian estimation for complex domain and rigorous convergence analysis.

DetailsMotivation: Existing PnP/RED methods with CNN denoisers lack theoretical guarantees for practical CNNs, while gradient-driven denoisers bridge this gap but have slow numerical solvers for CS MRI reconstruction.

Method: Develops a complex quasi-Newton proximal method with modified Hessian estimation that guarantees Hermitian positive definiteness in complex domain, enabling faster convergence than existing approaches.

Result: Numerical experiments on Cartesian and non-Cartesian sampling trajectories demonstrate the method’s effectiveness and efficiency in CS MRI reconstruction.

Conclusion: The proposed complex quasi-Newton proximal method achieves faster convergence with theoretical guarantees for nonconvex settings, bridging the performance-theory gap in CS MRI reconstruction.

Abstract: In compressed sensing (CS) MRI, model-based methods are pivotal to achieving accurate reconstruction. One of the main challenges in model-based methods is finding an effective prior to describe the statistical distribution of the target image. Plug-and-Play (PnP) and REgularization by Denoising (RED) are two general frameworks that use denoisers as the prior. While PnP/RED methods with convolutional neural networks (CNNs) based denoisers outperform classical hand-crafted priors in CS MRI, their convergence theory relies on assumptions that do not hold for practical CNNs. The recently developed gradient-driven denoisers offer a framework that bridges the gap between practical performance and theoretical guarantees. However, the numerical solvers for the associated minimization problem remain slow for CS MRI reconstruction. This paper proposes a complex quasi-Newton proximal method that achieves faster convergence than existing approaches. To address the complex domain in CS MRI, we propose a modified Hessian estimation method that guarantees Hermitian positive definiteness. Furthermore, we provide a rigorous convergence analysis of the proposed method for nonconvex settings. Numerical experiments on both Cartesian and non-Cartesian sampling trajectories demonstrate the effectiveness and efficiency of our approach.

[1004] SV-DRR: High-Fidelity Novel View X-Ray Synthesis Using Diffusion Model

Chun Xie, Yuichi Yoshii, Itaru Kitahara

Main category: eess.IV

TL;DR: A view-conditioned diffusion model using Diffusion Transformer to synthesize multi-view X-ray images from a single view, enabling high-resolution generation with improved angular control.

DetailsMotivation: Multi-view X-ray imaging provides complementary diagnostic information but increases radiation exposure and complicates clinical workflows. Need for method to generate multiple views from single acquisition.

Method: Proposes view-conditioned diffusion model leveraging Diffusion Transformer to preserve fine details, using weak-to-strong training strategy for stable high-resolution image generation from single X-ray view.

Result: Method generates higher-resolution outputs with improved control over viewing angles compared to prior methods limited in angular range, resolution, and image quality.

Conclusion: The approach has significant implications for clinical applications, medical education, and data extension, enabling creation of diverse, high-quality datasets for training and analysis.

Abstract: X-ray imaging is a rapid and cost-effective tool for visualizing internal human anatomy. While multi-view X-ray imaging provides complementary information that enhances diagnosis, intervention, and education, acquiring images from multiple angles increases radiation exposure and complicates clinical workflows. To address these challenges, we propose a novel view-conditioned diffusion model for synthesizing multi-view X-ray images from a single view. Unlike prior methods, which are limited in angular range, resolution, and image quality, our approach leverages the Diffusion Transformer to preserve fine details and employs a weak-to-strong training strategy for stable high-resolution image generation. Experimental results demonstrate that our method generates higher-resolution outputs with improved control over viewing angles. This capability has significant implications not only for clinical applications but also for medical education and data extension, enabling the creation of diverse, high-quality datasets for training and analysis. Our code is available at https://github.com/xiechun298/SV-DRR.

[1005] Multimodal Fusion at Three Tiers: Physics-Driven Data Generation and Vision-Language Guidance for Brain Tumor Segmentation

Mingda Zhang

Main category: eess.IV

TL;DR: A three-tier fusion architecture for brain tumor segmentation that combines pixel-level multimodal data (MRI, simulated US, synthetic CT), feature-level Transformer-based cross-modal fusion, and semantic-level clinical text guidance from GPT-4V.

DetailsMotivation: To address challenges in automatic brain tumor segmentation including tumor morphological heterogeneity and complex 3D spatial relationships, which current deep learning methods struggle with.

Method: Three-tier fusion: 1) Pixel-level: physical modeling extends MRI to multimodal data (simulated US, synthetic CT); 2) Feature-level: Transformer-based cross-modal fusion via multi-teacher collaborative distillation; 3) Semantic-level: GPT-4V clinical text transformed to spatial guidance using CLIP contrastive learning and FiLM.

Result: Achieved average Dice coefficients of 0.8665, 0.9014, and 0.8912 on BraTS 2020, 2021, and 2023 datasets respectively, with average HD95 reduction of 6.57mm compared to baseline.

Conclusion: The method provides a new paradigm for precise tumor segmentation and boundary localization through comprehensive processing from data augmentation to semantic guidance.

Abstract: Accurate brain tumor segmentation is crucial for neuro-oncology diagnosis and treatment planning. Deep learning methods have made significant progress, but automatic segmentation still faces challenges, including tumor morphological heterogeneity and complex three-dimensional spatial relationships. This paper proposes a three-tier fusion architecture that achieves precise brain tumor segmentation. The method processes information progressively at the pixel, feature, and semantic levels. At the pixel level, physical modeling extends magnetic resonance imaging (MRI) to multimodal data, including simulated ultrasound and synthetic computed tomography (CT). At the feature level, the method performs Transformer-based cross-modal feature fusion through multi-teacher collaborative distillation, integrating three expert teachers (MRI, US, CT). At the semantic level, clinical textual knowledge generated by GPT-4V is transformed into spatial guidance signals using CLIP contrastive learning and Feature-wise Linear Modulation (FiLM). These three tiers together form a complete processing chain from data augmentation to feature extraction to semantic guidance. We validated the method on the Brain Tumor Segmentation (BraTS) 2020, 2021, and 2023 datasets. The model achieves average Dice coefficients of 0.8665, 0.9014, and 0.8912 on the three datasets, respectively, and reduces the 95% Hausdorff Distance (HD95) by an average of 6.57 millimeters compared with the baseline. This method provides a new paradigm for precise tumor segmentation and boundary localization.

[1006] SpectraLift: Physics-Guided Spectral-Inversion Network for Self-Supervised Hyperspectral Image Super-Resolution

Ritik Shah, Marco F. Duarte

Main category: eess.IV

TL;DR: SpectraLift is a self-supervised framework that fuses low-resolution hyperspectral images with high-resolution multispectral images using only the MSI’s spectral response function, without requiring PSF calibration or ground truth data.

DetailsMotivation: Existing HSI-MSI fusion methods require impractical PSF calibration or ground truth HR-HSI, which are difficult to obtain in real-world settings. There's a need for a self-supervised approach that can work with readily available data.

Method: Trains a lightweight per-pixel MLP network using synthetic LR-MSI from LR-HSI as input, LR-HSI as output, and L1 spectral reconstruction loss. At inference, maps HR-MSI pixels to HR-HSI estimates.

Result: Outperforms state-of-the-art methods on PSNR, SAM, SSIM, and RMSE benchmarks. Converges in minutes and is agnostic to spatial blur and resolution.

Conclusion: SpectraLift provides an effective self-supervised solution for HSI-MSI fusion that works with practical real-world constraints and achieves superior performance compared to existing methods.

Abstract: High-spatial-resolution hyperspectral images (HSI) are essential for applications such as remote sensing and medical imaging, yet HSI sensors inherently trade spatial detail for spectral richness. Fusing high-spatial-resolution multispectral images (HR-MSI) with low-spatial-resolution hyperspectral images (LR-HSI) is a promising route to recover fine spatial structures without sacrificing spectral fidelity. Most state-of-the-art methods for HSI-MSI fusion demand point spread function (PSF) calibration or ground truth high resolution HSI (HR-HSI), both of which are impractical to obtain in real world settings. We present SpectraLift, a fully self-supervised framework that fuses LR-HSI and HR-MSI inputs using only the MSI’s Spectral Response Function (SRF). SpectraLift trains a lightweight per-pixel multi-layer perceptron (MLP) network using ($i$)~a synthetic low-spatial-resolution multispectral image (LR-MSI) obtained by applying the SRF to the LR-HSI as input, ($ii$)~the LR-HSI as the output, and ($iii$)~an $\ell_1$ spectral reconstruction loss between the estimated and true LR-HSI as the optimization objective. At inference, SpectraLift uses the trained network to map the HR-MSI pixel-wise into a HR-HSI estimate. SpectraLift converges in minutes, is agnostic to spatial blur and resolution, and outperforms state-of-the-art methods on PSNR, SAM, SSIM, and RMSE benchmarks.

[1007] REACT-KD: Region-Aware Cross-modal Topological Knowledge Distillation for Interpretable Medical Image Classification

Hongzhao Chen, Hexiao Ding, Yufeng Jiang, Jing Lan, Ka Chun Li, Gerald W. Y. Cheng, Nga-Chun Ng, Yao Pu, Jing Cai, Liang-ting Lin, Jung Sun Yoo

Main category: eess.IV

TL;DR: REACT-KD is a region-aware cross-modal knowledge distillation framework that transfers supervision from multi-modal imaging sources to a lightweight CT-based model for reliable tumor classification, achieving high performance even with degraded inputs.

DetailsMotivation: To address challenges in clinical tumor classification including heterogeneous modality quality, limited annotations, and lack of structured anatomical guidance by leveraging multi-modal supervision.

Method: Uses dual teacher design: one branch captures structure-function relationships via PET/CT, another models dose-aware features via synthetically degraded CT. Employs logits distillation for semantic alignment and region graph distillation for anatomical topology. Includes CBAM3D for cross-modal attention and modality dropout for robust inference.

Result: Achieved 93.5% AUC on internal PET/CT cohort and maintained 76.6%-81.5% AUC across varying dose degradation levels in external CT testing. Decision curve analysis showed highest net clinical benefit across all thresholds.

Conclusion: REACT-KD provides reliable and interpretable tumor classification with robust performance under partial or noisy inputs, demonstrating practical value for real-world diagnostic applications.

Abstract: Reliable and interpretable tumor classification from clinical imaging remains a core challenge. The main difficulties arise from heterogeneous modality quality, limited annotations, and the absence of structured anatomical guidance. We present REACT-KD, a Region-Aware Cross-modal Topological Knowledge Distillation framework that transfers supervision from high-fidelity multi-modal sources into a lightweight CT-based student model. The framework employs a dual teacher design. One branch captures structure-function relationships through dual-tracer PET/CT, while the other models dose-aware features using synthetically degraded low-dose CT. These branches jointly guide the student model through two complementary objectives. The first achieves semantic alignment through logits distillation, and the second models anatomical topology through region graph distillation. A shared CBAM3D module ensures consistent attention across modalities. To improve reliability in deployment, REACT-KD introduces modality dropout during training, which enables robust inference under partial or noisy inputs. As a case study, we applied REACT-KD to hepatocellular carcinoma staging. The framework achieved an average AUC of 93.5% on an internal PET/CT cohort and maintained 76.6% to 81.5% AUC across varying levels of dose degradation in external CT testing. Decision curve analysis further shows that REACT-KD consistently provides the highest net clinical benefit across all thresholds, confirming its value in real-world diagnostic practice. Code is available at: https://github.com/Kinetics-JOJO/REACT-KD

[1008] Robust Pan-Cancer Mitotic Figure Detection with YOLOv12

Raphaël Bourgade, Guillaume Balezo, Hana Feki, Lily Monier, Matthieu Blons, Alice Blondel, Delphine Loussouarn, Anne Vincent-Salomon, Thomas Walter

Main category: eess.IV

TL;DR: A YOLOv12-based approach for mitosis detection achieved second place in the MIDOG 2025 challenge with F1-scores of 0.801 on hotspots and 0.7216 on whole-slide regions.

DetailsMotivation: Mitotic figures are crucial for tumor pathology but challenging to identify consistently, with high inter-observer variability among pathologists.

Method: Used state-of-the-art YOLOv12 object detection architecture for mitosis detection without external data.

Result: Achieved F1-score of 0.801 on preliminary test set (hotspots) and ranked second on final test with F1-score of 0.7216 on complex whole-slide regions.

Conclusion: The YOLOv12-based approach demonstrates robust mitosis detection performance across heterogeneous tissue regions, ranking second in the international MIDOG 2025 challenge.

Abstract: Mitotic figures represent a key histoprognostic feature in tumor pathology, providing crucial insights into tumor aggressiveness and proliferation. However, their identification remains challenging, subject to significant inter-observer variability, even among experienced pathologists. To address this issue, the MItosis DOmain Generalization (MIDOG) 2025 challenge marks the third edition of an international competition aiming to develop robust mitosis detection algorithms. In this paper, we present a mitotic figure detection approach based on the state-of-the-art YOLOv12 object detection architecture. Our method achieved an F1-score of 0.801 on the preliminary test set (hotspots only) and ranked second on the final test leaderboard with an F1-score of 0.7216 across complex and heterogeneous whole-slide regions, without relying on external data.

Last updated: 2025-11-05
Built with Hugo, theme modified on Stack