Daily arXiv Papers - 2025-09-03

AI-enhanced summaries of 0 research papers from arXiv

Today’s Research Highlights

AI-enhanced summaries of the latest research papers from arXiv.

Table of Contents

cs.CL

[1] MultiStream-LLM: Bridging Modalities for Robust Sign Language Translation

Marshall Thomas, Edward Fish, Richard Bowden

Main category: cs.CL

TL;DR: MultiStream-LLM introduces a modular framework for sign language translation using specialized predictors for continuous signing, fingerspelling, and lipreading, achieving state-of-the-art results by solving recognition tasks separately before fusion.

DetailsMotivation: Monolithic end-to-end SLT models fail at precise fingerspelling recognition and integration of non-manual facial cues, leading to poor performance on translating names, places, and technical terms.

Method: Uses separate expert networks for continuous signing, fingerspelling, and lipreading that decode modalities into tokens, then fuses them with a lightweight transformer to resolve temporal misalignments before final generation by an LLM.

Result: Achieves new SOTA on How2Sign benchmark with BLEU-4 score of 23.5 and 73.2% letter accuracy on ChicagoFSWildPlus fingerspelling dataset.

Conclusion: Isolating and solving distinct recognition tasks before fusion provides a more powerful pathway to robust, high-fidelity sign language translation compared to monolithic approaches.

Abstract: Despite progress in gloss-free Sign Language Translation (SLT), monolithic end-to-end models consistently fail on two critical components of natural signing: the precise recognition of high-speed fingerspelling and the integration of asynchronous non-manual cues from the face. Recent progress in Automated Sign Language Translation with Large Language Models has side stepped this challenge, forcing a single network to learn these simultaneously resulting in poor performance when tasked with translating crucial information such as names,places, and technical terms. We introduce MultiStream-LLM, a modular framework designed to overcome these limitations. Our approach employs separate, specialized predictors for continuous signing, fingerspelling, and lipreading. Each expert network first decodes its specific modality into a sequence of tokens. These parallel streams are then fused by a lightweight transformer that resolves temporal misalignments before passing the combined representation to a Large Language Model (LLM) for final sentence generation. Our method establishes a new state-of-the-art on the How2Sign benchmark with a BLEU-4 score of 23.5 and achieves 73.2% letter accuracy on the challenging ChicagoFSWildPlus fingerspelling dataset. These results validate our core hypothesis: by isolating and solving distinct recogni tion tasks before fusion, our multi-expert approach provides a more powerful and effective pathway to robust, high-fidelity sign language translation.

[2] Compiling Prompts, Not Crafting Them: A Reproducible Workflow for AI-Assisted Evidence Synthesis

Teo Susnjak

Main category: cs.CL

TL;DR: This paper proposes a structured framework using declarative prompt optimization to improve reliability and reproducibility of LLM-assisted systematic literature reviews, addressing current fragility in manual prompt crafting.

DetailsMotivation: Current LLM approaches for systematic literature reviews rely on brittle, manually crafted prompts that compromise reliability and reproducibility, undermining scientific confidence in evidence synthesis.

Method: Adapts declarative prompt optimization advances for general-purpose LLMs and proposes a domain-specific framework embedding task declarations, test suites, and automated prompt tuning into reproducible SLR workflows.

Result: Provides a concrete blueprint with working code examples that enables researchers to construct verifiable LLM pipelines aligned with transparency and rigor principles in evidence synthesis.

Conclusion: This represents a novel application of declarative prompt optimization methods to systematic literature review pipelines, offering more reliable and reproducible LLM-assisted evidence synthesis.

Abstract: Large language models (LLMs) offer significant potential to accelerate systematic literature reviews (SLRs), yet current approaches often rely on brittle, manually crafted prompts that compromise reliability and reproducibility. This fragility undermines scientific confidence in LLM-assisted evidence synthesis. In response, this work adapts recent advances in declarative prompt optimisation, developed for general-purpose LLM applications, and demonstrates their applicability to the domain of SLR automation. This research proposes a structured, domain-specific framework that embeds task declarations, test suites, and automated prompt tuning into a reproducible SLR workflow. These emerging methods are translated into a concrete blueprint with working code examples, enabling researchers to construct verifiable LLM pipelines that align with established principles of transparency and rigour in evidence synthesis. This is a novel application of such approaches to SLR pipelines.

[3] What Are Research Hypotheses?

Jian Wu, Sarah Rajtmajer

Main category: cs.CL

TL;DR: This paper provides an overview and delineation of various definitions of ‘hypothesis’ in natural language processing, highlighting differences from traditional scientific definitions and variations across NLU literature.

DetailsMotivation: To address the migration of hypothesis definitions from traditional scientific meanings to various interpretations in natural language understanding tasks, and to emphasize the importance of well-structured definitions for machine-interpretable scholarly records.

Method: The authors overview and analyze various definitions of hypotheses across recently published NLU tasks, discerning nuances and differences in how hypotheses are defined in the literature.

Result: The paper identifies significant variations in how hypotheses are defined across NLU literature, showing that definitions have diverged from traditional scientific meanings and differ even within the NLU field.

Conclusion: Well-structured and well-defined hypotheses are crucial, especially as the field moves toward machine-interpretable scholarly records, requiring clearer and more consistent definitions of hypotheses in NLU research.

Abstract: Over the past decades, alongside advancements in natural language processing, significant attention has been paid to training models to automatically extract, understand, test, and generate hypotheses in open and scientific domains. However, interpretations of the term \emph{hypothesis} for various natural language understanding (NLU) tasks have migrated from traditional definitions in the natural, social, and formal sciences. Even within NLU, we observe differences defining hypotheses across literature. In this paper, we overview and delineate various definitions of hypothesis. Especially, we discern the nuances of definitions across recently published NLU tasks. We highlight the importance of well-structured and well-defined hypotheses, particularly as we move toward a machine-interpretable scholarly record.

[4] Explainable Chain-of-Thought Reasoning: An Empirical Analysis on State-Aware Reasoning Dynamics

Sheldon Yu, Yuxin Xiong, Junda Wu, Xintong Li, Tong Yu, Xiang Chen, Ritwik Sinha, Jingbo Shang, Julian McAuley

Main category: cs.CL

TL;DR: A state-aware transition framework that abstracts chain-of-thought reasoning into structured latent dynamics using spectral analysis and Markov chains for improved explainability.

DetailsMotivation: Current chain-of-thought prompting provides multi-step reasoning but lacks explainability, with prior work focusing only on local token-level attribution while ignoring high-level semantic roles and transitions between reasoning steps.

Method: Represent each reasoning step via spectral analysis of token-level embeddings, cluster them into semantically coherent latent states, and model their progression as a Markov chain to capture the global structure of reasoning.

Result: The framework provides a structured and interpretable view of the reasoning process that supports semantic role identification, temporal pattern visualization, and consistency evaluation.

Conclusion: The state-aware transition framework offers an effective approach to abstract and analyze chain-of-thought trajectories, enabling better understanding of the semantic evolution and global structure of multi-step reasoning in language models.

Abstract: Recent advances in chain-of-thought (CoT) prompting have enabled large language models (LLMs) to perform multi-step reasoning. However, the explainability of such reasoning remains limited, with prior work primarily focusing on local token-level attribution, such that the high-level semantic roles of reasoning steps and their transitions remain underexplored. In this paper, we introduce a state-aware transition framework that abstracts CoT trajectories into structured latent dynamics. Specifically, to capture the evolving semantics of CoT reasoning, each reasoning step is represented via spectral analysis of token-level embeddings and clustered into semantically coherent latent states. To characterize the global structure of reasoning, we model their progression as a Markov chain, yielding a structured and interpretable view of the reasoning process. This abstraction supports a range of analyses, including semantic role identification, temporal pattern visualization, and consistency evaluation.

[5] The Rarity Blind Spot: A Framework for Evaluating Statistical Reasoning in LLMs

Seiji Maekawa, Hayate Iso, Nikita Bhutani

Main category: cs.CL

TL;DR: Introduces Distinctive Feature Mining (DFM) task and DiFBench framework to evaluate LLMs’ ability to identify globally rare features across document collections, revealing significant performance gaps and limitations in statistical reasoning.

DetailsMotivation: Real-world decision-making requires identifying distinctive features across candidates, but existing LLM benchmarks focus on retrieval/summarization rather than global rarity detection in document collections.

Method: Created DiFBench framework with configurable parameters (document set size, distinctiveness thresholds) to systematically evaluate DFM capability across 10 state-of-the-art LLMs.

Result: Significant performance gap between general-purpose and reasoning-enhanced models; all models degrade with increased complexity and document count; common failure is misidentifying frequent features as distinctive.

Conclusion: Contemporary LLMs have core limitations in fine-grained statistical reasoning and rarity detection, highlighting the need for improved capabilities in distinctive feature mining tasks.

Abstract: Effective decision-making often relies on identifying what makes each candidate distinctive. While existing benchmarks for LLMs emphasize retrieving or summarizing information relevant to a given query, they do not evaluate a model’s ability to identify globally distinctive features across a set of documents. We introduce Distinctive Feature Mining (DFM), a new task that challenges models to analyze a small-to-medium collection (10-40 documents) and surface features that are rare in the global context (e.g., appearing in less than 10% of documents). This setting mirrors real-world scenarios such as candidate selection or product differentiation, where statistical reasoning, not retrieval, is key. To enable systematic evaluation of this capability, we present DiFBench, a configurable benchmark creation framework with controllable parameters such as document set size and distinctiveness thresholds. Using DiFBench, we perform a large-scale assessment of distinctive feature mining across ten state-of-the-art LLMs. Our findings reveal a significant performance gap between general-purpose and reasoning-enhanced models. All models, however, substantially degrade as the task complexity and document count increase. We also find that a common failure mode is misidentifying frequent features as distinctive. These insights reveal core limitations in contemporary LLMs’ abilities to perform fine-grained, statistical reasoning and rarity detection.

[6] The Differential Meaning of Models: A Framework for Analyzing the Structural Consequences of Semantic Modeling Decisions

Zachary K. Stine, James E. Deitrick

Main category: cs.CL

TL;DR: Proposes a theoretical framework based on Peirce’s semiotic theory to analyze and compare different modeling methods for human meaning-making, treating models themselves as signs that can be understood relationally through contrast with other models.

DetailsMotivation: The field lacks a general theoretical framework for describing modeling practices across various model types in a consistent way, despite the proliferation of methods for modeling human meaning-making in complex semiotic systems.

Method: Develops a framework grounded in C.S. Peirce’s semiotic theory, arguing that models measure latent symbol geometries and can be understood relationally through contrast with other models when performance measures are insufficient.

Result: The framework enables treating models and modeling decisions as signs themselves, providing a basis for understanding model semantics and facilitating comparative analysis across different modeling approaches.

Conclusion: The proposed semiotic framework offers a unified way to analyze modeling practices, addresses foundational questions in the field, and opens new directions for understanding how different models provide interpretive lenses for complex symbolic datasets.

Abstract: The proliferation of methods for modeling of human meaning-making constitutes a powerful class of instruments for the analysis of complex semiotic systems. However, the field lacks a general theoretical framework for describing these modeling practices across various model types in an apples-to-apples way. In this paper, we propose such a framework grounded in the semiotic theory of C. S. Peirce. We argue that such models measure latent symbol geometries, which can be understood as hypotheses about the complex of semiotic agencies underlying a symbolic dataset. Further, we argue that in contexts where a model’s value cannot be straightforwardly captured by proxy measures of performance, models can instead be understood relationally, so that the particular interpretive lens of a model becomes visible through its contrast with other models. This forms the basis of a theory of model semantics in which models, and the modeling decisions that constitute them, are themselves treated as signs. In addition to proposing the framework, we illustrate its empirical use with a few brief examples and consider foundational questions and future directions enabled by the framework.

[7] The Temporal Game: A New Perspective on Temporal Relation Extraction

Hugo Sousa, Ricardo Campos, Alípio Jorge

Main category: cs.CL

TL;DR: The Temporal Game is a novel interactive approach for temporal relation extraction that decomposes interval relations into point-wise comparisons, enabling fine-grained annotation and supporting both interval and instant entities through temporal closure.

DetailsMotivation: To create a more flexible and fine-grained temporal annotation approach that can handle both interval and instant entities, while laying groundwork for reinforcement learning applications in temporal reasoning.

Method: Decomposes interval-level relations into point-wise comparisons between start/end points, uses temporal closure to infer additional relations and enforce consistency, and provides both Game mode (with scoring) and Annotation mode for custom documents.

Result: A publicly available demo system that serves as both research tool and annotation interface, supporting TempEval-3 dataset annotation and custom document processing with timeline export capabilities.

Conclusion: The Temporal Game provides a novel point-based strategy for temporal annotation that offers greater flexibility than previous approaches and establishes a foundation for training reinforcement learning agents on temporal reasoning tasks.

Abstract: In this paper we demo the Temporal Game, a novel approach to temporal relation extraction that casts the task as an interactive game. Instead of directly annotating interval-level relations, our approach decomposes them into point-wise comparisons between the start and end points of temporal entities. At each step, players classify a single point relation, and the system applies temporal closure to infer additional relations and enforce consistency. This point-based strategy naturally supports both interval and instant entities, enabling more fine-grained and flexible annotation than any previous approach. The Temporal Game also lays the groundwork for training reinforcement learning agents, by treating temporal annotation as a sequential decision-making task. To showcase this potential, the demo presented in this paper includes a Game mode, in which users annotate texts from the TempEval-3 dataset and receive feedback based on a scoring system, and an Annotation mode, that allows custom documents to be annotated and resulting timeline to be exported. Therefore, this demo serves both as a research tool and an annotation interface. The demo is publicly available at https://temporal-game.inesctec.pt, and the source code is open-sourced to foster further research and community-driven development in temporal reasoning and annotation.

[8] Exploring Reasoning-Infused Text Embedding with Large Language Models for Zero-Shot Dense Retrieval

Yuxiang Liu, Tian Wang, Gourab Kundu, Tianyu Cao, Guang Cheng, Zhen Ge, Jianshu Chen, Qingjun Cui, Trishul Chilimbi

Main category: cs.CL

TL;DR: RITE enhances text embeddings by integrating logical reasoning through generative LLMs, generating intermediate reasoning texts before computing embeddings to improve retrieval performance on reasoning-intensive tasks.

DetailsMotivation: Encoder-only retrievers struggle with complex queries requiring sophisticated reasoning beyond surface-level matching, while decoder-only LLMs have strong reasoning capabilities that are underutilized in existing embedding methods.

Method: Proposes Reasoning-Infused Text Embedding (RITE) which generates intermediate reasoning texts in token space using generative LLMs before computing embeddings, enriching representations with inferential depth.

Result: Experimental results on BRIGHT benchmark show RITE significantly enhances zero-shot retrieval performance across diverse domains in reasoning-intensive retrieval tasks.

Conclusion: Incorporating reasoning into the embedding process through intermediate text generation effectively bridges the gap between LLMs’ reasoning capabilities and text embedding needs, demonstrating substantial performance improvements.

Abstract: Transformer-based models such as BERT and E5 have significantly advanced text embedding by capturing rich contextual representations. However, many complex real-world queries require sophisticated reasoning to retrieve relevant documents beyond surface-level lexical matching, where encoder-only retrievers often fall short. Decoder-only large language models (LLMs), known for their strong reasoning capabilities, offer a promising alternative. Despite this potential, existing LLM-based embedding methods primarily focus on contextual representation and do not fully exploit the reasoning strength of LLMs. To bridge this gap, we propose Reasoning-Infused Text Embedding (RITE), a simple but effective approach that integrates logical reasoning into the text embedding process using generative LLMs. RITE builds upon existing language model embedding techniques by generating intermediate reasoning texts in the token space before computing embeddings, thereby enriching representations with inferential depth. Experimental results on BRIGHT, a reasoning-intensive retrieval benchmark, demonstrate that RITE significantly enhances zero-shot retrieval performance across diverse domains, underscoring the effectiveness of incorporating reasoning into the embedding process.

[9] OpinioRAG: Towards Generating User-Centric Opinion Highlights from Large-scale Online Reviews

Mir Tafseer Nayeem, Davood Rafiei

Main category: cs.CL

TL;DR: OpinioRAG is a scalable, training-free framework that combines RAG-based evidence retrieval with LLMs to generate personalized opinion summaries from large volumes of user reviews, addressing scalability and personalization challenges in existing methods.

DetailsMotivation: Existing methods for opinion highlights generation from large user review volumes either fail to scale or produce generic summaries that overlook personalized needs, creating a gap for tailored, efficient solutions.

Method: Introduces OpinioRAG framework combining RAG-based evidence retrieval with LLMs, proposes novel reference-free verification metrics for sentiment-rich domains, and contributes a large-scale dataset of long-form user reviews with expert summaries and annotated queries.

Result: Through extensive experiments, the framework demonstrates capability to generate accurate, relevant, and structured summaries at scale while identifying key challenges and providing actionable insights for system improvement.

Conclusion: OpinioRAG establishes itself as a robust framework for scalable opinion summary generation and paves the way for future research in personalized review summarization with novel evaluation metrics and datasets.

Abstract: We study the problem of opinion highlights generation from large volumes of user reviews, often exceeding thousands per entity, where existing methods either fail to scale or produce generic, one-size-fits-all summaries that overlook personalized needs. To tackle this, we introduce OpinioRAG, a scalable, training-free framework that combines RAG-based evidence retrieval with LLMs to efficiently produce tailored summaries. Additionally, we propose novel reference-free verification metrics designed for sentiment-rich domains, where accurately capturing opinions and sentiment alignment is essential. These metrics offer a fine-grained, context-sensitive assessment of factual consistency. To facilitate evaluation, we contribute the first large-scale dataset of long-form user reviews, comprising entities with over a thousand reviews each, paired with unbiased expert summaries and manually annotated queries. Through extensive experiments, we identify key challenges, provide actionable insights into improving systems, pave the way for future research, and position OpinioRAG as a robust framework for generating accurate, relevant, and structured summaries at scale.

[10] Wage Sentiment Indices Derived from Survey Comments via Large Language Models

Taihei Sone

Main category: cs.CL

TL;DR: This paper proposes a Wage Sentiment Index (WSI) using Large Language Models to forecast wage dynamics in Japan, based on the Economy Watchers Survey data. The LLM-based WSI significantly outperforms baseline approaches and pretrained models.

DetailsMotivation: The emergence of generative AI creates new opportunities for economic text analysis. There's a need for timely and effective wage forecasting to enhance economic policy design by governments and central banks.

Method: Extends the Price Sentiment Index framework to wage sentiment using LLMs. Develops a scalable data architecture based on Japan’s Economy Watchers Survey that can integrate additional sources like newspapers and social media.

Result: Experimental results show that WSI models based on LLMs significantly outperform both baseline approaches and pretrained models in forecasting wage dynamics.

Conclusion: LLM-driven sentiment indices have strong potential to enhance the timeliness and effectiveness of economic policy design, demonstrating the value of generative AI in economic text analysis.

Abstract: The emergence of generative Artificial Intelligence (AI) has created new opportunities for economic text analysis. This study proposes a Wage Sentiment Index (WSI) constructed with Large Language Models (LLMs) to forecast wage dynamics in Japan. The analysis is based on the Economy Watchers Survey (EWS), a monthly survey conducted by the Cabinet Office of Japan that captures real-time economic assessments from workers in industries highly sensitive to business conditions. The WSI extends the framework of the Price Sentiment Index (PSI) used in prior studies, adapting it specifically to wage related sentiment. To ensure scalability and adaptability, a data architecture is also developed that enables integration of additional sources such as newspapers and social media. Experimental results demonstrate that WSI models based on LLMs significantly outperform both baseline approaches and pretrained models. These findings highlight the potential of LLM-driven sentiment indices to enhance the timeliness and effectiveness of economic policy design by governments and central banks.

[11] Balanced Actor Initialization: Stable RLHF Training of Distillation-Based Reasoning Models

Chen Zheng, Yiyuan Ma, Yuan Yang, Deyi Liu, Jing Liu, Zuquan Song, Yuxin Song, Cheng Ren, Hang Zhu, Xin Liu, Yiyuan Ma, Siyuan Qiao, Xun Zhou, Liang Xiang, Yonghui Wu

Main category: cs.CL

TL;DR: RLHF applied to distillation-trained models causes sequence length collapse and reward instability. Proposed Balanced Actor Initialization (BAI) merges instruction-following, reasoning, and pretrained models to enable stable training.

DetailsMotivation: The third paradigm of applying RLHF to distillation-trained models presents significant challenges including sequence length collapse and reward instability, which compromise model alignment and reasoning capabilities.

Method: Balanced Actor Initialization (BAI) - a two-stage weighted model merging approach that first merges instruction-following and distillation-based reasoning models, then combines this intermediate model with the pretrained model to preserve foundational knowledge.

Result: BAI resolves sequence length collapse, mitigates the reward hockey stick curve, enables continuous sequence length improvement during training, and achieves optimal trade-offs between training stability and reasoning capability preservation.

Conclusion: BAI provides an effective solution for stable training in the third paradigm, enabling more capable reasoning models that combine distillation efficiency with RLHF alignment.

Abstract: The development of alignment and reasoning capabilities in large language models has seen remarkable progress through two paradigms: instruction tuning and reinforcement learning from human feedback (RLHF) alignment paradigm, and distillation-based reasoning fine-tuning paradigm. While both approaches prove effective independently, the third paradigm of applying RLHF to distillation-trained models presents significant challenges. Our investigation reveals two critical phenomena that emerge in this paradigm: Sequence Length Collapse, where language generation dramatically reduces during early RLHF training, and the Reward Hockey Stick Curve, featuring severe reward score drops followed by gradual recovery. These instabilities fundamentally compromise the model’s alignment and reasoning capabilities. To address these challenges, we propose Balanced Actor Initialization (BAI), a two-stage weighted model merging approach. BAI first merges instruction-following and distillation-based reasoning fine-tuned models, then further combines this intermediate model with the pretrained model to preserve foundational knowledge. Through comprehensive experiments across diverse benchmarks and detailed analysis of training experiments, we demonstrate that BAI resolves Sequence Length Collapse, mitigates the Reward Hockey Stick Curve, and enables continuous sequence length improvement during training. Additionally, our analysis reveals that balanced merging ratios achieve optimal trade-offs between training stability and reasoning capability preservation. Our work provides the effective solution for stable training in this third paradigm, enabling more capable reasoning models that combine distillation efficiency with RLHF alignment.

[12] GIER: Gap-Driven Self-Refinement for Large Language Models

Rinku Dewri

Main category: cs.CL

TL;DR: GIER is a framework that improves LLM outputs through self-reflection using natural language gap descriptions, enhancing reasoning quality without accuracy loss.

DetailsMotivation: Existing prompting strategies rely on demonstrations or templates, but there's a need for methods that use conceptual quality criteria to guide LLM self-improvement through natural language gap descriptions.

Method: GIER uses natural language descriptions of reasoning gaps to prompt models to iteratively critique and refine their own outputs based on conceptual quality criteria, without relying on demonstrations or chain-of-thought templates.

Result: Across three reasoning tasks and four LLMs, GIER improved rationale quality, grounding, and reasoning alignment without degrading task accuracy. Models successfully interpreted abstract conceptual gaps and translated them into concrete reasoning improvements.

Conclusion: GIER demonstrates that LLMs can effectively use natural language gap descriptions for self-reflection and iterative improvement, providing a general framework for enhancing reasoning quality across various models and tasks.

Abstract: We introduce GIER (Gap-driven Iterative Enhancement of Responses), a general framework for improving large language model (LLM) outputs through self-reflection and revision based on conceptual quality criteria. Unlike prompting strategies that rely on demonstrations, examples, or chain-of-thought templates, GIER utilizes natural language descriptions of reasoning gaps, and prompts a model to iteratively critique and refine its own outputs to better satisfy these criteria. Across three reasoning-intensive tasks (SciFact, PrivacyQA, and e-SNLI) and four LLMs (GPT-4.1, GPT-4o Mini, Gemini 1.5 Pro, and Llama 3.3 70B), GIER improves rationale quality, grounding, and reasoning alignment without degrading task accuracy. Our analysis demonstrates that models can not only interpret abstract conceptual gaps but also translate them into concrete reasoning improvements.

[13] Open Data Synthesis For Deep Research

Ziyi Xia, Kun Luo, Hongjin Qian, Zheng Liu

Main category: cs.CL

TL;DR: InfoSeek is a framework for synthesizing complex Deep Research tasks using hierarchical constraint satisfaction problems, enabling better training of LLMs for multi-step reasoning tasks.

DetailsMotivation: Existing benchmarks fail to capture the complexity of deep research tasks that require hierarchical reasoning and evidence synthesis from diverse sources, while synthetic datasets often have shortcuts or knowledge leakage issues.

Method: Uses a dual-agent system to recursively build Research Trees from webpages, blurring intermediate nodes into valid sub-problems, and converting these trees into natural language questions requiring full hierarchy traversal. Enables scaling with over 50K training examples and reasoning trajectories via reject sampling.

Result: Models trained on InfoSeek outperform strong baselines. 3B LLMs optimized with InfoSeek surpass 32B models and lightweight commercial APIs, achieving performance comparable to stronger APIs on BrowseComp-Plus benchmark.

Conclusion: InfoSeek effectively addresses the gap in deep research task benchmarks, provides scalable training data, and supports advanced optimization strategies while preserving meta-information for compound reward design and trajectory-level exploration.

Abstract: Large language models (LLMs) are increasingly expected to go beyond simple factual queries toward Deep Research-tasks that require decomposing questions into sub-problems, coordinating multi-step reasoning, and synthesizing evidence from diverse sources. We formalize Deep Research tasks with verifiable answers as Hierarchical Constraint Satisfaction Problems (HCSPs), which are fundamentally different from single-constraint, multi-hop, or flat CSP formulations. However, existing benchmarks (e.g., Natural Questions, HotpotQA) fail to capture this complexity, while recent synthetic datasets often introduce shortcut reasoning, knowledge leakage, or lack sufficient structural depth. To address this gap, we introduce InfoSeek, a scalable framework for synthesizing complex Deep Research tasks. InfoSeek uses a dual-agent system to recursively build a Research Tree from large-scale webpages, blurring intermediate nodes into valid sub-problems, and converting these trees into natural language questions that require traversing the full hierarchy. It also enables rapid scaling, yielding over 50K training examples, a curated test set, and reasoning trajectories generated via reject sampling. Experiments show that models trained on InfoSeek consistently outperform strong baselines. On a challenging benchmark BrowseComp-Plus, 3B LLMs optimized with InfoSeek surpass much larger 32B models and lightweight commercial APIs (e.g., Gemini2.5-Flash), while achieving performance comparable to stronger APIs (e.g., Gemini2.5-Pro). By preserving meta-information such as intermediate steps and retrieval labels, InfoSeek further supports advanced optimization strategies, including compound reward design and trajectory-level exploration. We provide our codes and datasets in \href{https://github.com/VectorSpaceLab/InfoSeek}{this repository}.

[14] GraphKV: Breaking the Static Selection Paradigm with Graph-Based KV Cache Eviction

Xuelin Li, Xiangqi Jin, Linfeng Zhang

Main category: cs.CL

TL;DR: GraphKV is a graph-based framework that dynamically manages KV cache by modeling tokens as nodes with importance scores and updating them through decay-signal-propagation, enabling adaptive retention of contextually significant tokens.

DetailsMotivation: Conventional KV eviction strategies use static heuristics that fail to capture evolving token dependencies during inference, limiting performance in long text sequence processing for LLMs.

Method: Models tokens as graph nodes with importance scores and similarity-based edges, uses decay-signal-propagation mechanism to dynamically update token importance by propagating information across the graph.

Result: Provides a plug-and-play framework that can be seamlessly integrated with existing KV cache eviction methods like SnapKV and PyramidKV.

Conclusion: GraphKV offers an adaptive, graph-based approach to KV cache management that better captures evolving token dependencies compared to static heuristic methods.

Abstract: Efficient Key-Value (KV) cache management is essential for processing long text sequences in large language models (LLMs), where memory constraints often limit performance. Conventional KV eviction strategies, such as top-k selection based on attention scores, depend on static heuristics that fail to capture the evolving implicit dependencies among tokens during inference. To overcome this, we propose GraphKV, a graph-based framework that redefines token selection for KV cache compression. In GraphKV, tokens are modeled as nodes with importance scores, and edges represent their similarity relationships. Through a decay-signal-propagation mechanism, token importance is dynamically updated by propagating information across the graph, enabling adaptive retention of the most contextually significant tokens. GraphKV can be seamlessly utilized in existing KV cache eviction methods such as SnapKV and PyramidKV in a plug-and-play manner. Codes will be released on Github.

[15] The Resurgence of GCG Adversarial Attacks on Large Language Models

Yuting Tan, Xuying Li, Zhuo Li, Huizhen Shu, Peikang Hu

Main category: cs.CL

TL;DR: Systematic evaluation of GCG and T-GCG adversarial attacks shows attack success decreases with model size, prefix heuristics overestimate effectiveness, and coding prompts are more vulnerable than safety prompts.

DetailsMotivation: To systematically evaluate gradient-based adversarial prompting methods (GCG and T-GCG) across different LLM scales and understand their effectiveness, limitations, and vulnerabilities in both safety and reasoning tasks.

Method: Evaluated GCG and annealing-augmented T-GCG algorithms on Qwen2.5-0.5B, LLaMA-3.2-1B, and GPT-OSS-20B models using both safety-oriented prompts (AdvBench) and reasoning-intensive coding prompts, comparing prefix-based heuristics with GPT-4o semantic judgments.

Result: Three key findings: (1) Attack success rates decrease with model size due to complex loss landscapes; (2) Prefix heuristics substantially overestimate effectiveness compared to semantic judgments; (3) Coding prompts are significantly more vulnerable than safety prompts.

Conclusion: GCG has scalability limits, reasoning tasks contain overlooked vulnerabilities, and annealing-inspired strategies like T-GCG show promise for diversifying adversarial search, motivating further development for robust adversarial evaluation.

Abstract: Gradient-based adversarial prompting, such as the Greedy Coordinate Gradient (GCG) algorithm, has emerged as a powerful method for jailbreaking large language models (LLMs). In this paper, we present a systematic appraisal of GCG and its annealing-augmented variant, T-GCG, across open-source LLMs of varying scales. Using Qwen2.5-0.5B, LLaMA-3.2-1B, and GPT-OSS-20B, we evaluate attack effectiveness on both safety-oriented prompts (AdvBench) and reasoning-intensive coding prompts. Our study reveals three key findings: (1) attack success rates (ASR) decrease with model size, reflecting the increasing complexity and non-convexity of larger models’ loss landscapes; (2) prefix-based heuristics substantially overestimate attack effectiveness compared to GPT-4o semantic judgments, which provide a stricter and more realistic evaluation; and (3) coding-related prompts are significantly more vulnerable than adversarial safety prompts, suggesting that reasoning itself can be exploited as an attack vector. In addition, preliminary results with T-GCG show that simulated annealing can diversify adversarial search and achieve competitive ASR under prefix evaluation, though its benefits under semantic judgment remain limited. Together, these findings highlight the scalability limits of GCG, expose overlooked vulnerabilities in reasoning tasks, and motivate further development of annealing-inspired strategies for more robust adversarial evaluation.

[16] MedSEBA: Synthesizing Evidence-Based Answers Grounded in Evolving Medical Literature

Juraj Vladika, Florian Matthes

Main category: cs.CL

TL;DR: MedSEBA is an AI-powered system that synthesizes evidence-based medical answers by grounding LLM responses in dynamically retrieved PubMed studies, providing traceable arguments and research consensus visualization.

DetailsMotivation: Address the challenge of distinguishing reliable medical information from misleading content online and help researchers track evolving scientific findings that traditional search tools don't capture.

Method: Uses Large Language Models to generate coherent answers but grounds them in trustworthy medical studies dynamically retrieved from PubMed database, with traceable key points and arguments linked to respective studies.

Result: User study showed medical experts and lay users find the system usable, helpful, trustworthy and informative. Provides overview of study support/refutation levels and research consensus evolution visualization.

Conclusion: The system is well-suited for both everyday health questions and advanced research insights, effectively bridging the gap between AI-generated content and evidence-based medical research.

Abstract: In the digital age, people often turn to the Internet in search of medical advice and recommendations. With the increasing volume of online content, it has become difficult to distinguish reliable sources from misleading information. Similarly, millions of medical studies are published every year, making it challenging for researchers to keep track of the latest scientific findings. These evolving studies can reach differing conclusions, which is not reflected in traditional search tools. To address these challenges, we introduce MedSEBA, an interactive AI-powered system for synthesizing evidence-based answers to medical questions. It utilizes the power of Large Language Models to generate coherent and expressive answers, but grounds them in trustworthy medical studies dynamically retrieved from the research database PubMed. The answers consist of key points and arguments, which can be traced back to respective studies. Notably, the platform also provides an overview of the extent to which the most relevant studies support or refute the given medical claim, and a visualization of how the research consensus evolved through time. Our user study revealed that medical experts and lay users find the system usable and helpful, and the provided answers trustworthy and informative. This makes the system well-suited for both everyday health questions and advanced research insights.

[17] The Gold Medals in an Empty Room: Diagnosing Metalinguistic Reasoning in LLMs with Camlang

Fenghua Liu, Yulong Chen, Yixuan Liu, Zhujun Jin, Solomon Tsai, Ming Zhong

Main category: cs.CL

TL;DR: Camlang is a constructed language test that reveals LLMs struggle with genuine metalinguistic reasoning, performing far below humans despite excelling on English benchmarks.

DetailsMotivation: To test whether LLMs' benchmark success reflects genuine reasoning or just pattern matching by assessing their ability to learn an unfamiliar language through explicit metalinguistic deductive learning.

Method: Created Camlang, a novel constructed language with naturalistic but unattested features, providing explicit grammar rules and bilingual dictionary. Adapted CommonsenseQA into Camlang-CSQA-v0 task requiring grammar rule application and lexical mapping.

Result: GPT-5 achieved 98% EM accuracy in English but only 47% in Camlang, far below human performance at 87%. Most model successes came from shallow lexical alignment rather than systematic grammatical mastery.

Conclusion: Camlang exposes fundamental gaps between current LLMs and human metalinguistic competence, showing models lack genuine reasoning abilities despite benchmark success.

Abstract: Large Language Models (LLMs) achieve gold-medal performance across many benchmarks, yet it remains unclear whether such success reflects genuine reasoning or pattern matching. From a cognitive science perspective, an informative test is whether models can master an unfamiliar language through explicit metalinguistic deductive learning, a paradigm where human learners can reliably internalise grammatical systems through metalinguistic reasoning. We address this question with Camlang, a novel constructed language that exhibits naturalistic yet unattested feature combinations. Camlang consists of two explicit resources, a grammar book and a bilingual dictionary, which mirror adult second-language learning via explicit grammar rules and lexical lookup, and enable us to disentangle errors in morpho-syntax, lexical semantics, and sentence-level reasoning. Human experiments show that these resources are sufficient for participants to acquire Camlang and successfully solve Camlang tasks. To operationalise evaluation, we adapt CommonsenseQA into Camlang, creating Camlang-CSQA-v0, the first task in a broader suite where solving questions requires applying grammar rules and lexical mappings. Experimental results show that GPT-5 achieves 98% EM accuracy in English but only 47% in Camlang, far below human performance at 87%, while other state-of-the-art reasoning LLMs perform even worse. Human verification further reveals that most model successes stem from shallow lexical alignment while GPT-5 shows emerging metalinguistic awareness to a limited extent but not systematic grammatical mastery as humans. Camlang establishes a cognitively grounded evaluation paradigm that exposes fundamental gaps between current models and human metalinguistic competence.

[18] GOSU: Retrieval-Augmented Generation with Global-Level Optimized Semantic Unit-Centric Framework

Xuecheng Zou, Ke Liu, Bingbing Wang, Huafei Deng, Li Zhang, Yu Tang

Main category: cs.CL

TL;DR: GOSU is a semantic unit-centric RAG framework that addresses ambiguity and retrieval overhead in graph-based RAG by performing global disambiguation and leveraging semantic units to capture interconnections across global context.

DetailsMotivation: Standard graph-based RAG with heterogeneous graphs and hypergraphs faces issues with ambiguous high-level semantic unit extraction limited to local text chunks, leading to complex coupling and increased retrieval overhead due to lack of global knowledge and neglect of fine-grained relationships.

Method: GOSU performs global merging on pre-extracted semantic units from local chunks, guides entity and relationship extraction to reduce coreference resolution difficulty, and introduces hierarchical keyword extraction with semantic unit completion to capture both fine-grained binary relationships and coarse-grained n-ary relationships.

Result: Evaluation across multiple tasks demonstrates that GOSU outperforms baseline RAG methods in terms of generation quality.

Conclusion: The GOSU framework effectively addresses the limitations of traditional graph-based RAG by leveraging global semantic unit processing and hierarchical relationship capture, resulting in improved generation performance.

Abstract: Building upon the standard graph-based Retrieval-Augmented Generation (RAG), the introduction of heterogeneous graphs and hypergraphs aims to enrich retrieval and generation by leveraging the relationships between multiple entities through the concept of semantic units (SUs). But this also raises a key issue: The extraction of high-level SUs limited to local text chunks is prone to ambiguity, complex coupling, and increased retrieval overhead due to the lack of global knowledge or the neglect of fine-grained relationships. To address these issues, we propose GOSU, a semantic unit-centric RAG framework that efficiently performs global disambiguation and utilizes SUs to capture interconnections between different nodes across the global context. In the graph construction phase, GOSU performs global merging on the pre-extracted SUs from local text chunks and guides entity and relationship extraction, reducing the difficulty of coreference resolution while uncovering global semantic objects across text chunks. In the retrieval and generation phase, we introduce hierarchical keyword extraction and semantic unit completion. The former uncovers the fine-grained binary relationships overlooked by the latter, while the latter compensates for the coarse-grained n-ary relationships missing from the former. Evaluation across multiple tasks demonstrates that GOSU outperforms the baseline RAG methods in terms of generation quality.

[19] CVPD at QIAS 2025 Shared Task: An Efficient Encoder-Based Approach for Islamic Inheritance Reasoning

Salah Eddine Bekhouche, Abdellah Zakaria Sellam, Hichem Telli, Cosimo Distante, Abdenour Hadid

Main category: cs.CL

TL;DR: Lightweight framework using Arabic text encoder and Attentive Relevance Scoring for Islamic inheritance multiple-choice questions, achieving 69.87% accuracy with MARBERT vs 87.6% from large LLMs, but offering better efficiency and on-device deployability.

DetailsMotivation: Islamic inheritance law requires precise heir identification and share calculation, posing challenges for AI. Need for efficient, specialized systems that can operate on-device without heavy resource requirements.

Method: Specialized Arabic text encoder (MARBERT, ArabicBERT, AraBERT) with Attentive Relevance Scoring (ARS) to rank answer options by semantic relevance. Fast, on-device inference without generative reasoning.

Result: MARBERT-based approach achieved 69.87% accuracy on QIAS 2025 dataset, while large API-based LLMs (Gemini, DeepSeek) reached up to 87.6% accuracy but required more resources and context.

Conclusion: Quantifies trade-off between peak performance of large models (87.6%) and practical advantages of smaller specialized systems (69.87%) - efficiency, on-device deployability, and privacy in high-stakes domains like Islamic inheritance law.

Abstract: Islamic inheritance law (Ilm al-Mawarith) requires precise identification of heirs and calculation of shares, which poses a challenge for AI. In this paper, we present a lightweight framework for solving multiple-choice inheritance questions using a specialised Arabic text encoder and Attentive Relevance Scoring (ARS). The system ranks answer options according to semantic relevance, and enables fast, on-device inference without generative reasoning. We evaluate Arabic encoders (MARBERT, ArabicBERT, AraBERT) and compare them with API-based LLMs (Gemini, DeepSeek) on the QIAS 2025 dataset. While large models achieve an accuracy of up to 87.6%, they require more resources and are context-dependent. Our MARBERT-based approach achieves 69.87% accuracy, presenting a compelling case for efficiency, on-device deployability, and privacy. While this is lower than the 87.6% achieved by the best-performing LLM, our work quantifies a critical trade-off between the peak performance of large models and the practical advantages of smaller, specialized systems in high-stakes domains.

[20] TECP: Token-Entropy Conformal Prediction for LLMs

Beining Xu

Main category: cs.CL

TL;DR: TECP is a novel framework that uses token-level entropy as an uncertainty measure in conformal prediction to provide formal coverage guarantees for black-box language generation models.

DetailsMotivation: Uncertainty quantification for open-ended language generation remains challenging, especially under black-box constraints where internal model signals are inaccessible.

Method: Leverages token-level entropy as a logit-free, reference-free uncertainty measure and integrates it into a split conformal prediction pipeline to construct prediction sets with formal coverage guarantees.

Result: Empirical evaluations across six large language models and two benchmarks demonstrate TECP consistently achieves reliable coverage and compact prediction sets, outperforming prior self-consistency-based methods.

Conclusion: TECP provides a principled and efficient solution for trustworthy generation in black-box LLM settings with provable error control.

Abstract: Uncertainty quantification (UQ) for open-ended language generation remains a critical yet underexplored challenge, especially under black-box constraints where internal model signals are inaccessible. In this paper, we introduce Token-Entropy Conformal Prediction (TECP), a novel framework that leverages token-level entropy as a logit-free, reference-free uncertainty measure and integrates it into a split conformal prediction (CP) pipeline to construct prediction sets with formal coverage guarantees. Unlike existing approaches that rely on semantic consistency heuristics or white-box features, TECP directly estimates epistemic uncertainty from the token entropy structure of sampled generations and calibrates uncertainty thresholds via CP quantiles to ensure provable error control. Empirical evaluations across six large language models and two benchmarks (CoQA and TriviaQA) demonstrate that TECP consistently achieves reliable coverage and compact prediction sets, outperforming prior self-consistency-based UQ methods. Our method provides a principled and efficient solution for trustworthy generation in black-box LLM settings.

[21] Talk Less, Call Right: Enhancing Role-Play LLM Agents with Automatic Prompt Optimization and Role Prompting

Saksorn Ruangtanusak, Pittawat Taveekitworachai, Kunat Pipatanakul

Main category: cs.CL

TL;DR: Rule-based role prompting with character-card/scene-contract design and strict function calling enforcement achieved best performance (0.571 score) for role-playing dialogue agents, addressing over-speaking and under-acting issues.

DetailsMotivation: Dialogue agents in CPDC 2025 API track often produce overly long responses (over-speaking) while failing to use tools effectively according to persona (under-acting), such as generating non-existent function calls or unnecessary tool calls.

Method: Explored four prompting approaches: 1) basic role prompting, 2) human-crafted role prompting, 3) automatic prompt optimization (APO), and 4) rule-based role prompting with character-card/scene-contract design and strict function calling enforcement.

Result: Rule-based role prompting (RRP) achieved best performance with overall score of 0.571, improving on zero-shot baseline score of 0.519, outperforming more elaborate methods like APO.

Conclusion: RRP design substantially improves effectiveness and reliability of role-playing dialogue agents. Best-performing prompts and APO tool are open-sourced to support future persona prompt development.

Abstract: This report investigates approaches for prompting a tool-augmented large language model (LLM) to act as a role-playing dialogue agent in the API track of the Commonsense Persona-grounded Dialogue Challenge (CPDC) 2025. In this setting, dialogue agents often produce overly long in-character responses (over-speaking) while failing to use tools effectively according to the persona (under-acting), such as generating function calls that do not exist or making unnecessary tool calls before answering. We explore four prompting approaches to address these issues: 1) basic role prompting, 2) human-crafted role prompting, 3) automatic prompt optimization (APO), and 4) rule-based role prompting. The rule-based role prompting (RRP) approach achieved the best performance through two novel techniques–character-card/scene-contract design and strict enforcement of function calling–which led to an overall score of 0.571, improving on the zero-shot baseline score of 0.519. These findings demonstrate that RRP design can substantially improve the effectiveness and reliability of role-playing dialogue agents compared with more elaborate methods such as APO. To support future efforts in developing persona prompts, we are open-sourcing all of our best-performing prompts and the APO tool. Source code is available at https://github.com/scb-10x/apo.

[22] Entropy-based Coarse and Compressed Semantic Speech Representation Learning

Jialong Zuo, Guangyan Zhang, Minghui Fang, Shengpeng Ji, Xiaoqi Jiao, Jingyu Li, Yiwen Guo, Zhou Zhao

Main category: cs.CL

TL;DR: Proposes entropy-based dynamic aggregation framework to compress redundant speech tokens from 25-50Hz to more semantic representations, achieving comparable or better performance with improved efficiency.

DetailsMotivation: Existing speech representations at 25-50 tokens/second are redundant since speech only conveys 2-5 words/second, and fine-grained tokenization captures phonetic rather than semantic information, hindering downstream efficiency.

Method: Pre-train speech language model via next-token prediction, use predictive entropy to adaptively determine aggregation boundaries, then fuse information within segments using cross-attention module with controllable granularity via entropy threshold.

Result: Experiments on ASR, speech-to-text translation, and voice conversion show compressed representations perform on par with or better than dense token sequences.

Conclusion: The entropy-based dynamic aggregation framework effectively learns compressed semantic speech representations while maintaining or improving performance across multiple speech tasks.

Abstract: Discrete speech representation learning has recently attracted increasing interest in both acoustic and semantic modeling. Existing approaches typically encode 16 kHz waveforms into discrete tokens at a rate of 25 or 50 tokens per second. However, given that speech generally conveys only 2 to 5 words per second, such fine-grained tokenization introduces redundancy and hinders efficiency in downstream training and inference. Moreover, semantic speech representations at this frequency primarily capture phonetic-level information, while semantic understanding may not require such detailed token-level resolution. To address these limitations, we propose an entropy-based dynamic aggregation framework for learning compressed semantic speech representations. A speech language model is first pre-trained via next-token prediction on large-scale unlabeled data to capture frequent token patterns. Predictive entropy is then used to adaptively determine aggregation boundaries, followed by a cross-attention module that fuses information within each segment. By adjusting the entropy threshold, the granularity and compression ratio of the representations can be flexibly controlled. Experiments on ASR, speech-to-text translation, and voice conversion tasks demonstrate that the compressed representations perform on par with or better than dense token sequences, demonstrating the effectiveness of the proposed approach.

[23] ResearchQA: Evaluating Scholarly Question Answering at Scale Across 75 Fields with Survey-Mined Questions and Rubrics

Li S. Yifei, Allen Chang, Chaitanya Malaviya, Mark Yatskar

Main category: cs.CL

TL;DR: ResearchQA introduces a dataset of 21K research queries and 160K rubric items from 75 fields to evaluate LLM systems’ ability to provide comprehensive, citation-rich responses to research questions.

DetailsMotivation: Current evaluation of long-form responses relies heavily on expert annotators, limiting scope to fields like AI where researchers can easily enlist colleagues. Research expertise is widespread across many fields through survey articles.

Method: Distilled survey articles from 75 research fields into queries and rubrics, with 31 PhD annotators assessing quality. Created automatic pairwise judge with 74% expert agreement. Evaluated 18 systems across 7.6K pairwise evaluations.

Result: No parametric or retrieval-augmented system exceeded 70% rubric coverage. Best agentic system achieved 75% coverage. Systems struggled with citations (11% fully addressed), limitations (48%), and comparisons (49%).

Conclusion: ResearchQA enables comprehensive multi-field evaluation of LLM research capabilities, revealing significant competency gaps in current systems, particularly in citation handling and comprehensive analysis.

Abstract: Evaluating long-form responses to research queries heavily relies on expert annotators, restricting attention to areas like AI where researchers can conveniently enlist colleagues. Yet, research expertise is widespread: survey articles synthesize knowledge distributed across the literature. We introduce ResearchQA, a resource for evaluating LLM systems by distilling survey articles from 75 research fields into 21K queries and 160K rubric items. Each rubric, derived jointly with queries from survey sections, lists query-specific answer evaluation criteria, i.e., citing papers, making explanations, and describing limitations. Assessments by 31 Ph.D. annotators in 8 fields indicate 96% of queries support Ph.D. information needs and 87% of rubric items should be addressed in system responses by a sentence or more. Using our rubrics, we are able to construct an automatic pairwise judge obtaining 74% agreement with expert judgments. We leverage ResearchQA to analyze competency gaps in 18 systems in over 7.6K pairwise evaluations. No parametric or retrieval-augmented system we evaluate exceeds 70% on covering rubric items, and the highest-ranking agentic system shows 75% coverage. Error analysis reveals that the highest-ranking system fully addresses less than 11% of citation rubric items, 48% of limitation items, and 49% of comparison items. We release our data to facilitate more comprehensive multi-field evaluations.

[24] Modeling Motivated Reasoning in Law: Evaluating Strategic Role Conditioning in LLM Summarization

Eunjung Cho, Alexander Hoyle, Yoan Hermstrüwer

Main category: cs.CL

TL;DR: LLMs show role-biased summarization in legal contexts, selectively framing information to align with different legal stakeholders’ perspectives even when given balancing instructions.

DetailsMotivation: To investigate how LLMs exhibit motivated reasoning by adapting summaries to align with different legal roles (judges, prosecutors, attorneys) when summarizing judicial decisions, building on legal realism theories.

Method: Introduced an evaluation framework grounded in legal fact and reasoning inclusion, analyzing favorability towards stakeholders. Tested models with prompts conditioned on different legal roles, including balancing instructions.

Result: Models exhibited selective inclusion patterns that reflect role-consistent perspectives, showing bias even with balancing instructions. LLMs strategically frame information to align with stakeholder positions.

Conclusion: Findings raise concerns about LLMs inferring user roles from context and exhibiting alignment without explicit instructions. Highlights need for role-aware evaluation in high-stakes legal settings.

Abstract: Large Language Models (LLMs) are increasingly used to generate user-tailored summaries, adapting outputs to specific stakeholders. In legal contexts, this raises important questions about motivated reasoning – how models strategically frame information to align with a stakeholder’s position within the legal system. Building on theories of legal realism and recent trends in legal practice, we investigate how LLMs respond to prompts conditioned on different legal roles (e.g., judges, prosecutors, attorneys) when summarizing judicial decisions. We introduce an evaluation framework grounded in legal fact and reasoning inclusion, also considering favorability towards stakeholders. Our results show that even when prompts include balancing instructions, models exhibit selective inclusion patterns that reflect role-consistent perspectives. These findings raise broader concerns about how similar alignment may emerge as LLMs begin to infer user roles from prior interactions or context, even without explicit role instructions. Our results underscore the need for role-aware evaluation of LLM summarization behavior in high-stakes legal settings.

[25] Thinking Hard, Going Misaligned: Emergent Misalignment in LLMs

Hanqi Yan, Hainiu Xu, Yulan He

Main category: cs.CL

TL;DR: LLMs become more vulnerable to malicious requests when reasoning capabilities are enhanced through think-mode or math fine-tuning, creating a reasoning-safety trade-off that requires urgent alignment solutions.

DetailsMotivation: As LLMs gain widespread adoption, concerns about their safety and alignment with human values intensify. Previous research showed fine-tuning on malicious data causes misalignment, but this paper investigates a more concerning phenomenon where enhanced reasoning itself induces misalignment.

Method: The study examines how strengthening reasoning capabilities (via switching to think-mode or fine-tuning on benign math datasets) affects LLM responsiveness to malicious requests. Researchers analyze internal model states including attention shifts and specialized experts in mixture-of-experts models.

Result: LLMs become more responsive to malicious requests when reasoning is strengthened, with dense models being particularly vulnerable. Internal analysis shows attention shifts and specialized experts help redirect excessive reasoning towards safety guardrails.

Conclusion: The findings reveal an emerging reasoning-safety trade-off where enhanced reasoning capabilities paradoxically increase vulnerability to malicious requests, underscoring the urgent need to advance alignment techniques for reasoning-enhanced models.

Abstract: With Large Language Models (LLMs) becoming increasingly widely adopted, concerns regarding their safety and alignment with human values have intensified. Previous studies have shown that fine-tuning LLMs on narrow and malicious datasets induce misaligned behaviors. In this work, we report a more concerning phenomenon, Reasoning-Induced Misalignment. Specifically, we observe that LLMs become more responsive to malicious requests when reasoning is strengthened, via switching to “think-mode” or fine-tuning on benign math datasets, with dense models particularly vulnerable. Moreover, we analyze internal model states and find that both attention shifts and specialized experts in mixture-of-experts models help redirect excessive reasoning towards safety guardrails. These findings provide new insights into the emerging reasoning-safety trade-off and underscore the urgency of advancing alignment for advanced reasoning models.

[26] SimulMEGA: MoE Routers are Advanced Policy Makers for Simultaneous Speech Translation

Chenyang Le, Bing Han, Jinshun Li, Songyong Chen, Yanmin Qian

Main category: cs.CL

TL;DR: SimulMEGA is a simultaneous speech translation framework that uses Mixture-of-Experts gating and prefix-based training to achieve better latency-quality tradeoffs without inference overhead.

DetailsMotivation: Existing simultaneous speech translation systems struggle to balance translation quality, latency, and semantic coherence, especially in multilingual many-to-many scenarios where divergent read/write policies hinder unified strategy learning.

Method: Unsupervised policy learning framework combining prefix-based training with Mixture-of-Experts refiner to learn read/write decisions implicitly. Minimal modifications to standard transformer architectures, generalizes across speech-to-text and text-to-speech tasks.

Result: 500M parameter speech-to-text model outperforms Seamless baseline with under 7% BLEU degradation at 1.5s average lag and under 3% at 3s. Also successfully extended to streaming TTS with unidirectional backbone.

Conclusion: SimulMEGA provides an effective framework for simultaneous speech translation that achieves superior latency-quality tradeoffs across multiple modalities without adding inference-time computational overhead.

Abstract: Simultaneous Speech Translation (SimulST) enables real-time cross-lingual communication by jointly optimizing speech recognition and machine translation under strict latency constraints. Existing systems struggle to balance translation quality, latency, and semantic coherence, particularly in multilingual many-to-many scenarios where divergent read and write policies hinder unified strategy learning. In this paper, we present SimulMEGA (Simultaneous Generation by Mixture-of-Experts Gating), an unsupervised policy learning framework that combines prefix-based training with a Mixture-of-Experts refiner to learn effective read and write decisions in an implicit manner, without adding inference-time overhead. Our design requires only minimal modifications to standard transformer architectures and generalizes across both speech-to-text and text-to-speech streaming tasks. Through comprehensive evaluation on six language pairs, our 500M parameter speech-to-text model outperforms the Seamless baseline, achieving under 7 percent BLEU degradation at 1.5 seconds average lag and under 3 percent at 3 seconds. We further demonstrate the versatility of SimulMEGA by extending it to streaming TTS with a unidirectional backbone, yielding superior latency quality tradeoffs.

[27] StealthEval: A Probe-Rewrite-Evaluate Workflow for Reliable Benchmarks

Lang Xiong, Nishant Bhargava, Wesley Chang, Jianhang Hong, Haihao Liu, Kevin Zhu

Main category: cs.CL

TL;DR: LLMs show different behaviors in test vs deployment contexts (evaluation awareness). This study quantifies this by rewriting prompts to be more “deploy-like,” finding models become more honest and safe in perceived deployment settings.

DetailsMotivation: Benchmark performance may not reflect true model safety and honesty due to evaluation awareness - models behave differently when they perceive they're being tested versus in real deployment.

Method: Used linear probe to score prompts on test-like to deploy-like scale, then employed LLM rewriting strategy to shift prompts toward natural deployment context while preserving original tasks.

Result: 30% increase in average probe score after rewriting. Across models: 5.26% average increase in honest responses, 12.40% decrease in deceptive responses, and 6.38% increase in refusal rates indicating improved safety compliance.

Conclusion: Evaluation awareness is quantifiable and manipulable, with models being more prone to unsafe/deceptive outputs in perceived test environments, highlighting need for more realistic evaluation frameworks.

Abstract: Large Language Models (LLMs) often exhibit significant behavioral shifts when they perceive a change from a real-world deployment context to a controlled evaluation setting, a phenomenon known as “evaluation awareness.” This discrepancy poses a critical challenge for AI alignment, as benchmark performance may not accurately reflect a model’s true safety and honesty. In this work, we systematically quantify these behavioral changes by manipulating the perceived context of prompts. We introduce a methodology that uses a linear probe to score prompts on a continuous scale from “test-like” to “deploy-like” and leverage an LLM rewriting strategy to shift these prompts towards a more natural, deployment-style context while preserving the original task. Using this method, we achieved a 30% increase in the average probe score across a strategic role-playing dataset after rewriting. Evaluating a suite of state-of-the-art models on these original and rewritten prompts, we find that rewritten “deploy-like” prompts induce a significant and consistent shift in behavior. Across all models, we observed an average increase in honest responses of 5.26% and a corresponding average decrease in deceptive responses of 12.40%. Furthermore, refusal rates increased by an average of 6.38%, indicating heightened safety compliance. Our findings demonstrate that evaluation awareness is a quantifiable and manipulable factor that directly influences LLM behavior, revealing that models are more prone to unsafe or deceptive outputs in perceived test environments. This underscores the urgent need for more realistic evaluation frameworks to accurately gauge true model alignment before deployment.

[28] Gated Associative Memory: A Parallel O(N) Architecture for Efficient Sequence Modeling

Rishiraj Acharya

Main category: cs.CL

TL;DR: GAM network replaces Transformer’s quadratic self-attention with linear-complexity parallel pathways combining causal convolution for local context and associative memory for global patterns, achieving faster training and competitive performance.

DetailsMotivation: Transformer's quadratic complexity with sequence length creates bottlenecks for long contexts, requiring more efficient alternatives.

Method: Replace self-attention with parallel causal convolution (local context) and associative memory retrieval (global patterns), fused dynamically with gating mechanism.

Result: Consistently faster than Transformer and Mamba, achieves superior/competitive perplexity on WikiText-2 and TinyStories datasets.

Conclusion: GAM is a promising efficient alternative to Transformer for sequence modeling with linear complexity and competitive performance.

Abstract: The Transformer architecture, underpinned by the self-attention mechanism, has become the de facto standard for sequence modeling tasks. However, its core computational primitive scales quadratically with sequence length (O(N^2)), creating a significant bottleneck for processing long contexts. In this paper, we propose the Gated Associative Memory (GAM) network, a novel, fully parallel architecture for sequence modeling that exhibits linear complexity (O(N)) with respect to sequence length. The GAM block replaces the self-attention layer with two parallel pathways: a causal convolution to efficiently capture local, position-dependent context, and a parallel associative memory retrieval mechanism to model global, content-based patterns. These pathways are dynamically fused using a gating mechanism, allowing the model to flexibly combine local and global information for each token. We implement GAM from scratch and conduct a rigorous comparative analysis against a standard Transformer model and a modern linear-time baseline (Mamba) on the WikiText-2 benchmark, as well as against the Transformer on the TinyStories dataset. Our experiments demonstrate that GAM is consistently faster, outperforming both baselines on training speed, and achieves a superior or competitive final validation perplexity across all datasets, establishing it as a promising and efficient alternative for sequence modeling.

[29] Analysing the Language of Neural Audio Codecs

Joonyong Park, Shinnosuke Takamichi, David M. Chan, Shunsuke Kando, Yuki Saito, Hiroshi Saruwatari

Main category: cs.CL

TL;DR: Neural audio codecs produce speech tokens that follow linguistic statistical laws like Zipf’s and Heaps’ laws, and these statistical properties correlate with better speech recognition and synthesis performance.

DetailsMotivation: To understand the statistical and linguistic properties of neural audio codec tokens and how they relate to speech synthesis quality and intelligibility.

Method: Comparative analysis of NAC tokens examining adherence to linguistic statistical laws (Zipf’s, Heaps’), entropy, redundancy, and evaluating intelligibility through ASR error rates and quality through UTMOS scores.

Result: NAC tokens, especially 3-grams, exhibit language-like statistical patterns, and these properties correlate with improved speech recognition and resynthesis performance.

Conclusion: The findings provide insights into NAC token sequence structure and can inform the design of more effective generative speech models.

Abstract: This study presents a comparative analysis of the statistical and linguistic properties of neural audio codecs (NACs). We investigate discrete speech tokens produced by various NAC models, examining their adherence to linguistic statistical laws such as Zipf’s law and Heaps’ law, as well as their entropy and redundancy. To assess how these token-level properties relate to semantic and acoustic preservation in synthesized speech, we evaluate intelligibility using error rates of automatic speech recognition, and quality using the UTMOS score. Our results reveal that NAC tokens, particularly 3-grams, exhibit language-like statistical patterns. Moreover, these properties, together with measures of information content, are found to correlate with improved performances in speech recognition and resynthesis tasks. These findings offer insights into the structure of NAC token sequences and inform the design of more effective generative speech models.

[30] A Multi-Strategy Approach for AI-Generated Text Detection

Ali Zain, Sareem Farooqui, Muhammad Rafi

Main category: cs.CL

TL;DR: Three AI detection systems developed for M-DAIGT shared task: fine-tuned RoBERTa, TF-IDF+SVM, and ensemble model Candace using Llama-3.2 features. RoBERTa achieved best performance.

DetailsMotivation: To develop effective systems for detecting AI-generated content in news articles and academic abstracts as part of the M-DAIGT shared task.

Method: Developed three distinct approaches: (1) Fine-tuned RoBERTa-base classifier, (2) TF-IDF + SVM classifier, (3) Candace ensemble model using probabilistic features from multiple Llama-3.2 models processed through custom Transformer encoder.

Result: The RoBERTa-based system performed best, achieving near-perfect results on both development and test sets.

Conclusion: Fine-tuned RoBERTa classifier is the most effective approach among the three systems for detecting AI-generated content in news and academic texts.

Abstract: This paper presents presents three distinct systems developed for the M-DAIGT shared task on detecting AI generated content in news articles and academic abstracts. The systems includes: (1) A fine-tuned RoBERTa-base classifier, (2) A classical TF-IDF + Support Vector Machine (SVM) classifier , and (3) An Innovative ensemble model named Candace, leveraging probabilistic features extracted from multiple Llama-3.2 models processed by a customTransformer encoder.The RoBERTa-based system emerged as the most performant, achieving near-perfect results on both development and test sets.

[31] Can Multi-turn Self-refined Single Agent LMs with Retrieval Solve Hard Coding Problems?

Md Tanzib Hosain, Md Kishor Morol

Main category: cs.CL

TL;DR: The paper introduces the ICPC benchmark with 254 competitive programming problems and shows that advanced LM inference techniques can significantly improve solve rates from 19.1% to 42.2%, with human guidance further boosting performance.

DetailsMotivation: Competitive programming requires sophisticated algorithmic thinking and represents a challenging domain for assessing language models' capabilities, yet it has received insufficient attention in LM evaluation.

Method: Created ICPC benchmark with 254 problems including official analysis, reference code, and tests. Evaluated various LM inference techniques including zero-shot chain-of-thought, multi-turn self-judge with reflection, and retrieval over episodic information.

Result: Zero-shot chain-of-thought achieved 19.1% pass@1 solve rate, while best technique (multi-turn self-judge with reflection and retrieval) reached 42.2%. Human-in-the-loop guidance solved 17/18 previously unsolvable problems.

Conclusion: The study provides a step toward LMs with grounded, imaginative, and algorithmic thinking, demonstrating that advanced inference techniques and human guidance significantly improve competitive programming performance.

Abstract: Among the hardest tasks for humans are those found in competitive programming where problems require sophisticated algorithmic thinking, puzzle solving, and the creation of effective code. As a domain to assess language models (LMs), it has not received enough attention, though. This study presents the ICPC benchmark, which consists of 254 international collegiate programming contest (ICPC) tasks. Each problem includes official analysis, reference code, and sample, high-quality unit, and hidden tests. We are able to develop and evaluate a variety of LM inference techniques for competitive programming with these resources. With zero-shot chain-of-thought prompting, we find that o1 only achieves a 19.1% pass@1 solve rate. With our best inference technique, which combines multi-turn self-judge with reflection and retrieval over episodic information, raises this to 42.2%. Furthermore, we conduct a new human-in-the-loop investigation to gain a deeper understanding of the remaining difficulties. Surprisingly, we discover that o1 can solve 17 out of 18 problems that were previously unsolvable by any model or technique with just a few specific instructions. A footstep toward LMs with grounded, imaginative, and algorithmic thinking is provided by our quantitative findings and qualitative research. We open-source our code and data at https://github.com/kraritt/zolve.

[32] Confident, Calibrated, or Complicit: Probing the Trade-offs between Safety Alignment and Ideological Bias in Language Models in Detecting Hate Speech

Sanjeeevan Selvaganapathy, Mehwish Nasim

Main category: cs.CL

TL;DR: LLMs with safety alignment outperform uncensored models in hate speech detection (78.7% vs 64.1% accuracy) but have ideological anchoring, while uncensored models are more malleable. All models struggle with irony and show fairness disparities.

DetailsMotivation: To examine whether uncensored LLMs provide more objective hate speech classification compared to safety-aligned models, and to understand the trade-offs between performance and ideological bias.

Method: Comparative analysis of censored (safety-aligned) and uncensored LLMs on detecting implicit and explicit hate speech, evaluating accuracy, robustness, ideological consistency, and fairness across different target groups.

Result: Censored models significantly outperformed uncensored models (78.7% vs 64.1% accuracy) but were resistant to persona-based influence. All models failed at understanding irony and showed fairness disparities across target groups with unreliable self-reported certainty.

Conclusion: LLMs cannot be considered objective arbiters for hate speech detection due to ideological biases, fairness issues, and poor calibration. More sophisticated auditing frameworks are needed to address fairness, calibration, and ideological consistency.

Abstract: We investigate the efficacy of Large Language Models (LLMs) in detecting implicit and explicit hate speech, examining whether models with minimal safety alignment (uncensored) might provide more objective classification capabilities compared to their heavily-aligned (censored) counterparts. While uncensored models theoretically offer a less constrained perspective free from moral guardrails that could bias classification decisions, our results reveal a surprising trade-off: censored models significantly outperform their uncensored counterparts in both accuracy and robustness, achieving 78.7% versus 64.1% strict accuracy. However, this enhanced performance comes with its own limitation – the safety alignment acts as a strong ideological anchor, making censored models resistant to persona-based influence, while uncensored models prove highly malleable to ideological framing. Furthermore, we identify critical failures across all models in understanding nuanced language such as irony. We also find alarming fairness disparities in performance across different targeted groups and systemic overconfidence that renders self-reported certainty unreliable. These findings challenge the notion of LLMs as objective arbiters and highlight the need for more sophisticated auditing frameworks that account for fairness, calibration, and ideological consistency.

[33] Router Upcycling: Leveraging Mixture-of-Routers in Mixture-of-Experts Upcycling

Junfeng Ran, Guangxiang Zhao, Yuhan Wu, Dawei Zhu, Longyun Wu, Yikai Zhao, Tong Yang, Lin Sun, Xiangzheng Zhang, Sujian Li

Main category: cs.CL

TL;DR: Router Upcycling technique enhances Mixture-of-Experts upcycling by initializing multiple routers from attention heads, achieving SOTA performance with attention-like token routing.

DetailsMotivation: Efficient training of Mixture-of-Experts models remains challenging, and simple routers struggle with complex routing tasks in MoE upcycling, requiring better routing mechanisms.

Method: Initialize multiple routers from attention heads of preceding layers during upcycling. These routers collaboratively assign tokens to specialized experts using attention-like mechanism where tokens are processed into queries and aligned with experts’ features as keys.

Result: Experimental results show the method achieves state-of-the-art performance, outperforming other upcycling baselines.

Conclusion: Router Upcycling effectively enhances MoE upcycling performance by leveraging attention mechanisms for improved routing, providing a superior alternative to simple linear routers.

Abstract: The Mixture-of-Experts (MoE) models have gained significant attention in deep learning due to their dynamic resource allocation and superior performance across diverse tasks. However, efficiently training these models remains challenging. The MoE upcycling technique has been proposed to reuse and improve existing model components, thereby minimizing training overhead. Despite this, simple routers, such as linear routers, often struggle with complex routing tasks within MoE upcycling. In response, we propose a novel routing technique called Router Upcycling to enhance the performance of MoE upcycling models. Our approach initializes multiple routers from the attention heads of preceding attention layers during upcycling. These routers collaboratively assign tokens to specialized experts in an attention-like manner. Each token is processed into diverse queries and aligned with the experts’ features (serving as keys). Experimental results demonstrate that our method achieves state-of-the-art (SOTA) performance, outperforming other upcycling baselines.

[34] Do small language models generate realistic variable-quality fake news headlines?

Austin McCutcheon, Chris Brogly

Main category: cs.CL

TL;DR: Small language models (1.7B-14B parameters) can generate fake news headlines with high compliance and minimal ethical resistance, but existing detectors struggle to identify them accurately (35.2-63.5% accuracy).

DetailsMotivation: To evaluate whether small language models can be used to generate deceptive fake news headlines and assess how well existing detection systems can identify these AI-generated fakes.

Method: Tested 14 SLMs from various families using controlled prompt engineering to generate 24,000 headlines across low/high quality deceptive categories, then applied DistilBERT and bagging classifier models for detection.

Result: SLMs showed high compliance in generating fake headlines with minimal ethical resistance. Detection accuracy was poor (35.2-63.5%), indicating generated headlines don’t closely resemble human-written content.

Conclusion: SLMs can effectively generate falsified headlines, existing detection methods are inadequate, and generated content differs significantly from human-written news, posing challenges for content moderation.

Abstract: Small language models (SLMs) have the capability for text generation and may potentially be used to generate falsified texts online. This study evaluates 14 SLMs (1.7B-14B parameters) including LLaMA, Gemma, Phi, SmolLM, Mistral, and Granite families in generating perceived low and high quality fake news headlines when explicitly prompted, and whether they appear to be similar to real-world news headlines. Using controlled prompt engineering, 24,000 headlines were generated across low-quality and high-quality deceptive categories. Existing machine learning and deep learning-based news headline quality detectors were then applied against these SLM-generated fake news headlines. SLMs demonstrated high compliance rates with minimal ethical resistance, though there were some occasional exceptions. Headline quality detection using established DistilBERT and bagging classifier models showed that quality misclassification was common, with detection accuracies only ranging from 35.2% to 63.5%. These findings suggest the following: tested SLMs generally are compliant in generating falsified headlines, although there are slight variations in ethical restraints, and the generated headlines did not closely resemble existing primarily human-written content on the web, given the low quality classification accuracy.

[35] NADI 2025: The First Multidialectal Arabic Speech Processing Shared Task

Bashar Talafha, Hawau Olamide Toyin, Peter Sullivan, AbdelRahim Elmadany, Abdurrahman Juma, Amirbek Djanibekov, Chiyu Zhang, Hamad Alshehhi, Hanan Aldarmaki, Mustafa Jarrar, Nizar Habash, Muhammad Abdul-Mageed

Main category: cs.CL

TL;DR: NADI 2025 shared task results on Arabic speech dialect processing with 44 teams participating across three subtasks: dialect identification, speech recognition, and diacritic restoration.

DetailsMotivation: To address the challenges of Arabic dialect speech processing and advance research in dialect identification, recognition, and diacritic restoration for spoken Arabic dialects.

Method: Organized as a shared task with three subtasks: spoken dialect identification (Subtask 1), speech recognition (Subtask 2), and diacritic restoration for spoken dialects (Subtask 3). 44 teams participated with 100 valid submissions.

Result: Best results: 79.8% accuracy on dialect identification, 35.68/12.20 WER/CER on speech recognition, and 55/13 WER/CER on diacritic restoration. 8 unique teams submitted during testing phase.

Conclusion: Results demonstrate significant ongoing challenges in Arabic dialect speech processing, particularly in dialect identification, recognition, and diacritic restoration, highlighting the need for continued research and development in this area.

Abstract: We present the findings of the sixth Nuanced Arabic Dialect Identification (NADI 2025) Shared Task, which focused on Arabic speech dialect processing across three subtasks: spoken dialect identification (Subtask 1), speech recognition (Subtask 2), and diacritic restoration for spoken dialects (Subtask 3). A total of 44 teams registered, and during the testing phase, 100 valid submissions were received from eight unique teams. The distribution was as follows: 34 submissions for Subtask 1 “five teams{\ae}, 47 submissions for Subtask 2 “six teams”, and 19 submissions for Subtask 3 “two teams”. The best-performing systems achieved 79.8% accuracy on Subtask 1, 35.68/12.20 WER/CER (overall average) on Subtask 2, and 55/13 WER/CER on Subtask 3. These results highlight the ongoing challenges of Arabic dialect speech processing, particularly in dialect identification, recognition, and diacritic restoration. We also summarize the methods adopted by participating teams and briefly outline directions for future editions of NADI.

[36] Text Reinforcement for Multimodal Time Series Forecasting

Chen Su, Yuanhe Tian, Yan Song, Yongdong Zhang

Main category: cs.CL

TL;DR: Proposes TeR model to reinforce text inputs for multimodal time series forecasting, using reinforcement learning to improve text quality and enhance forecasting performance.

DetailsMotivation: Existing multimodal TSF approaches rely on high-quality text inputs, but text often fails to accurately capture historical time series information, leading to unstable performance. There's a need to enhance textual content to improve multimodal TSF reliability.

Method: Text Reinforcement model (TeR) generates reinforced text to address weaknesses in original text. Uses reinforcement learning with rewards based on impact on TSF performance and task relevance to optimize text quality.

Result: Extensive experiments on real-world benchmark dataset across various domains show the approach outperforms strong baselines and existing studies, demonstrating effectiveness.

Conclusion: Reinforcing text modalities through the proposed TeR model with reinforcement learning significantly improves multimodal time series forecasting performance by addressing text quality issues.

Abstract: Recent studies in time series forecasting (TSF) use multimodal inputs, such as text and historical time series data, to predict future values. These studies mainly focus on developing advanced techniques to integrate textual information with time series data to perform the task and achieve promising results. Meanwhile, these approaches rely on high-quality text and time series inputs, whereas in some cases, the text does not accurately or fully capture the information carried by the historical time series, which leads to unstable performance in multimodal TSF. Therefore, it is necessary to enhance the textual content to improve the performance of multimodal TSF. In this paper, we propose improving multimodal TSF by reinforcing the text modalities. We propose a text reinforcement model (TeR) to generate reinforced text that addresses potential weaknesses in the original text, then apply this reinforced text to support the multimodal TSF model’s understanding of the time series, improving TSF performance. To guide the TeR toward producing higher-quality reinforced text, we design a reinforcement learning approach that assigns rewards based on the impact of each reinforced text on the performance of the multimodal TSF model and its relevance to the TSF task. We optimize the TeR accordingly, so as to improve the quality of the generated reinforced text and enhance TSF performance. Extensive experiments on a real-world benchmark dataset covering various domains demonstrate the effectiveness of our approach, which outperforms strong baselines and existing studies on the dataset.

[37] CE-Bench: Towards a Reliable Contrastive Evaluation Benchmark of Interpretability of Sparse Autoencoders

Alex Gulko, Yusen Peng, Sachin Kumar

Main category: cs.CL

TL;DR: CE-Bench is a lightweight contrastive evaluation benchmark for sparse autoencoders that measures interpretability without requiring external LLMs, using curated story pairs.

DetailsMotivation: The lack of automated evaluation methods has hindered broader adoption and development of sparse autoencoders for interpretable feature discovery in LLMs.

Method: Built a contrastive evaluation benchmark using curated contrastive story pairs, conducted comprehensive ablation studies to validate effectiveness.

Result: CE-Bench reliably measures interpretability of sparse autoencoders and aligns well with existing benchmarks, all without external LLM requirements.

Conclusion: The approach provides an effective automated evaluation method for sparse autoencoders, with open-sourced implementation and dataset under MIT License.

Abstract: Probing with sparse autoencoders is a promising approach for uncovering interpretable features in large language models (LLMs). However, the lack of automated evaluation methods has hindered their broader adoption and development. In this work, we introduce CE-Bench, a novel and lightweight contrastive evaluation benchmark for sparse autoencoders, built on a curated dataset of contrastive story pairs. We conduct comprehensive ablation studies to validate the effectiveness of our approach. Our results show that CE-Bench reliably measures the interpretability of sparse autoencoders and aligns well with existing benchmarks, all without requiring an external LLM. The official implementation and evaluation dataset are open-sourced under the MIT License.

[38] Flavors of Moonshine: Tiny Specialized ASR Models for Edge Devices

Evan King, Adam Sabra, Manjunath Kudlur, James Wang, Pete Warden

Main category: cs.CL

TL;DR: Monolingual ASR models outperform multilingual ones for small models (27M params) when trained on balanced high-quality data mix, achieving 48% lower error rates than Whisper Tiny and matching larger models.

DetailsMotivation: Challenge the assumption that multilingual ASR models always outperform monolingual counterparts, especially for underrepresented languages and small model sizes.

Method: Train monolingual systems on carefully balanced mix of high-quality human-labeled, pseudo-labeled, and synthetic data for 27M parameter models across multiple underrepresented languages.

Result: Models achieve error rates 48% lower than comparably sized Whisper Tiny, outperform 9x larger Whisper Small, and match/outperform 28x larger Whisper Medium in most cases.

Conclusion: Monolingual training with quality data balancing enables superior on-device ASR for underrepresented languages, advancing state-of-the-art for small models and providing open-source support for 6 languages.

Abstract: We present the Flavors of Moonshine, a suite of tiny automatic speech recognition (ASR) models specialized for a range of underrepresented languages. Prevailing wisdom suggests that multilingual ASR models outperform monolingual counterparts by exploiting cross-lingual phonetic similarities. We challenge this assumption, showing that for sufficiently small models (27M parameters), training monolingual systems on a carefully balanced mix of high-quality human-labeled, pseudo-labeled, and synthetic data yields substantially superior performance. On average, our models achieve error rates 48% lower than the comparably sized Whisper Tiny model, outperform the 9x larger Whisper Small model, and in most cases match or outperform the 28x larger Whisper Medium model. These results advance the state of the art for models of this size, enabling accurate on-device ASR for languages that previously had limited support. We release Arabic, Chinese, Japanese, Korean, Ukrainian, and Vietnamese Moonshine models under a permissive open-source license.

[39] Learning to Shop Like Humans: A Review-driven Retrieval-Augmented Recommendation Framework with LLMs

Kaiwen Wei, Jinpeng Gao, Jiang Zhong, Yuming Yang, Fengmao Lv, Zhenyang Li

Main category: cs.CL

TL;DR: RevBrowse is a review-driven LLM recommendation framework that uses PrefRAG to efficiently retrieve relevant reviews for better item ranking and interpretability.

DetailsMotivation: LLMs show strong potential for recommendation but struggle with efficiently incorporating user reviews due to context window constraints and lack of mechanisms to prioritize context-relevant reviews.

Method: Proposes RevBrowse framework inspired by ‘browse-then-decide’ decision process, with PrefRAG module that disentangles user/item representations and adaptively retrieves preference-relevant content for target items.

Result: Extensive experiments on four Amazon datasets show consistent and significant improvements over strong baselines, demonstrating generalizability and effectiveness in modeling dynamic user preferences.

Conclusion: RevBrowse enhances LLM-based recommendation by making review usage more efficient and relevant, while providing interpretability through transparent retrieval process that shows which reviews influence recommendations.

Abstract: Large language models (LLMs) have shown strong potential in recommendation tasks due to their strengths in language understanding, reasoning and knowledge integration. These capabilities are especially beneficial for review-based recommendation, which relies on semantically rich user-generated texts to reveal fine-grained user preferences and item attributes. However, effectively incorporating reviews into LLM-based recommendation remains challenging due to (1) inefficient to dynamically utilize user reviews under LLMs’ constrained context windows, and (2) lacking effective mechanisms to prioritize reviews most relevant to the user’s current decision context. To address these challenges, we propose RevBrowse, a review-driven recommendation framework inspired by the “browse-then-decide” decision process commonly observed in online user behavior. RevBrowse integrates user reviews into the LLM-based reranking process to enhance its ability to distinguish between candidate items. To improve the relevance and efficiency of review usage, we introduce PrefRAG, a retrieval-augmented module that disentangles user and item representations into structured forms and adaptively retrieves preference-relevant content conditioned on the target item. Extensive experiments on four Amazon review datasets demonstrate that RevBrowse achieves consistent and significant improvements over strong baselines, highlighting its generalizability and effectiveness in modeling dynamic user preferences. Furthermore, since the retrieval-augmented process is transparent, RevBrowse offers a certain level of interpretability by making visible which reviews influence the final recommendation.

[40] Reward-Weighted Sampling: Enhancing Non-Autoregressive Characteristics in Masked Diffusion LLMs

Daehoon Gwak, Minseo Jung, Junwoo Park, Minho Park, ChaeHun Park, Junha Hyung, Jaegul Choo

Main category: cs.CL

TL;DR: Proposes Reward-Weighted Sampling (RWS) to improve masked diffusion models by using external reward models for global sequence guidance during decoding, promoting more non-autoregressive generation orders.

DetailsMotivation: Standard confidence-based sampling in masked diffusion models results in generation orders resembling sequential autoregressive processes, limiting the benefits of non-autoregressive modeling.

Method: RWS leverages external reward models to evaluate entire intermediate sequences at each diffusion step, scaling token logits based on global sequence quality to guide token selection and promote non-autoregressive generation.

Result: Experiments show RWS significantly promotes non-autoregressive generation orders and improves performance across multiple evaluation metrics compared to standard decoding methods.

Conclusion: Integrating global signals through reward-weighted sampling effectively enhances both non-autoregressive properties and overall performance of masked diffusion models.

Abstract: Masked diffusion models (MDMs) offer a promising non-autoregressive alternative for large language modeling. Standard decoding methods for MDMs, such as confidence-based sampling, select tokens independently based on individual token confidences at each diffusion step. However, we observe that this independent token selection often results in generation orders resembling sequential autoregressive processes, limiting the advantages of non-autoregressive modeling. To mitigate this pheonomenon, we propose Reward-Weighted Sampling (RWS), a novel decoding strategy that leverages an external reward model to provide a principled global signal during the iterative diffusion process. Specifically, at each diffusion step, RWS evaluates the quality of the entire intermediate sequence and scales token logits accordingly, guiding token selection by integrating global sequence-level coherence. This method selectively increases the confidence of tokens that initially have lower scores, thereby promoting a more non-autoregressive generation order. Furthermore, we provide theoretical justification showing that reward-weighted logit scaling induces beneficial rank reversals in token selection and consistently improves expected reward. Experiments demonstrate that RWS significantly promotes non-autoregressive generation orders, leading to improvements across multiple evaluation metrics. These results highlight the effectiveness of integrating global signals in enhancing both the non-autoregressive properties and overall performance of MDMs.

[41] Designing LMS and Instructional Strategies for Integrating Generative-Conversational AI

Elias Ra, Seung Je Kim, Eui-Yeong Seo, Geunju So

Main category: cs.CL

TL;DR: A framework for designing AI-powered LMS that integrates generative and conversational AI for adaptive, personalized learning using design-based research methodology.

DetailsMotivation: Higher education needs scalable, personalized learning experiences that maintain pedagogical coherence while leveraging AI capabilities.

Method: Design-based research methodology with five phases: literature review, SWOT analysis, ethical-pedagogical principles development, system design, and instructional strategy formulation.

Result: Developed an AI-LMS framework with modular components including configurable prompts, adaptive feedback loops, and multi-agent conversation flows aligned with behaviorist, constructivist, and connectivist learning theories.

Conclusion: The study presents a practical model combining AI capabilities with human-centered design and ethical safeguards for education, with future validation through real-world implementation planned.

Abstract: Higher education faces growing challenges in delivering personalized, scalable, and pedagogically coherent learning experiences. This study introduces a structured framework for designing an AI-powered Learning Management System (AI-LMS) that integrates generative and conversational AI to support adaptive, interactive, and learner-centered instruction. Using a design-based research (DBR) methodology, the framework unfolds through five phases: literature review, SWOT analysis, development of ethical-pedagogical principles, system design, and instructional strategy formulation. The resulting AI-LMS features modular components – including configurable prompts, adaptive feedback loops, and multi-agent conversation flows – aligned with pedagogical paradigms such as behaviorist, constructivist, and connectivist learning theories. By combining AI capabilities with human-centered design and ethical safeguards, this study advances a practical model for AI integration in education. Future research will validate and refine the system through real-world implementation.

[42] Zero-shot Context Biasing with Trie-based Decoding using Synthetic Multi-Pronunciation

Changsong Liu, Yizhou Peng, Eng Siong Chng

Main category: cs.CL

TL;DR: A synthesis-driven multi-pronunciation contextual biasing method for zero-shot ASR that uses TTS to generate pronunciation variants, compiles them into a prefix-trie for beam-search decoding, and reduces biased-word error rate by 43-44% on LibriSpeech.

DetailsMotivation: Contextual ASR systems struggle with out-of-vocabulary words due to limited training data and ambiguous/inconsistent pronunciations, making accurate recognition of rare words and named entities challenging.

Method: Leverage TTS to synthesize diverse speech samples containing target rare words, use pretrained Whisper to extract multiple pronunciation variants, compile variants into prefix-trie for shallow-fusion beam-search decoding, then map recognized variants back to original words.

Result: Method reduces biased-word error rate (B-WER) by 43% on LibriSpeech test-clean and 44% on test-other, while maintaining unbiased-WER (U-WER) essentially unchanged.

Conclusion: The proposed synthesis-driven multi-pronunciation approach effectively addresses pronunciation variability in contextual ASR, significantly improving recognition of rare words without compromising general ASR performance.

Abstract: Contextual automatic speech recognition (ASR) systems allow for recognizing out-of-vocabulary (OOV) words, such as named entities or rare words. However, it remains challenging due to limited training data and ambiguous or inconsistent pronunciations. In this paper, we propose a synthesis-driven multi-pronunciation contextual biasing method that performs zero-shot contextual ASR on a pretrained Whisper model. Specifically, we leverage text-to-speech (TTS) systems to synthesize diverse speech samples containing each target rare word, and then use the pretrained Whisper model to extract multiple predicted pronunciation variants. These variant token sequences are compiled into a prefix-trie, which assigns rewards to beam hypotheses in a shallow-fusion manner during beam-search decoding. Subsequently, any recognized variant is mapped back to the original rare word in the final transcription. The evaluation results on the LibriSpeech dataset show that our method reduces biased-word error rate (B-WER) by 43% on test-clean and 44% on test-other while maintaining unbiased-WER (U-WER) essentially unchanged.

[43] LLM Encoder vs. Decoder: Robust Detection of Chinese AI-Generated Text with LoRA

Houji Jin, Negin Ashrafi, Armin Abdollahi, Wei Liu, Jian Wang, Ganyu Gui, Maryam Pishgar, Huanghao Feng

Main category: cs.CL

TL;DR: Decoder-based LLMs with LoRA fine-tuning outperform encoder models and FastText for Chinese AI-generated text detection, achieving 95.94% accuracy with superior generalization.

DetailsMotivation: The rapid growth of LLMs demands accurate detection of AI-generated text, especially in Chinese where linguistic nuances challenge current methods.

Method: Systematic comparison of encoder Transformers (BERT/RoBERTa), decoder LLM (Qwen2.5-7B with LoRA), and FastText baseline using NLPCC 2025 dataset. Encoders used prompt-based MLM, Qwen used instruction-format with classification head.

Result: Encoder models suffered performance degradation (76.3-79.3% accuracy), FastText showed lexical robustness (83.5%) but lacked semantics, while LoRA-adapted Qwen2.5-7B achieved 95.94% accuracy with balanced precision-recall.

Conclusion: Decoder-based LLMs with parameter-efficient fine-tuning are most effective for robust Chinese AI-generated text detection, with future work planned on Qwen3 models and ensemble strategies.

Abstract: The rapid growth of large language models (LLMs) has heightened the demand for accurate detection of AI-generated text, particularly in languages like Chinese, where subtle linguistic nuances pose significant challenges to current methods. In this study, we conduct a systematic comparison of encoder-based Transformers (Chinese BERT-large and RoBERTa-wwm-ext-large), a decoder-only LLM (Alibaba’s Qwen2.5-7B/DeepSeek-R1-Distill-Qwen-7B fine-tuned via Low-Rank Adaptation, LoRA), and a FastText baseline using the publicly available dataset from the NLPCC 2025 Chinese AI-Generated Text Detection Task. Encoder models were fine-tuned using a novel prompt-based masked language modeling approach, while Qwen2.5-7B was adapted for classification with an instruction-format input and a lightweight classification head trained via LoRA. Experiments reveal that although encoder models nearly memorize training data, they suffer significant performance degradation under distribution shifts (RoBERTa: 76.3% test accuracy; BERT: 79.3%). FastText demonstrates surprising lexical robustness (83.5% accuracy) yet lacks deeper semantic understanding. In contrast, the LoRA-adapted Qwen2.5-7B achieves 95.94% test accuracy with balanced precision-recall metrics, indicating superior generalization and resilience to dataset-specific artifacts. These findings underscore the efficacy of decoder-based LLMs with parameter-efficient fine-tuning for robust Chinese AI-generated text detection. Future work will explore next-generation Qwen3 models, distilled variants, and ensemble strategies to enhance cross-domain robustness further.

[44] Decomposing and Revising What Language Models Generate

Zhichao Yan, Jiaoyan Chen, Jiapu Wang, Xiaoli Li, Ru Li, Jeff Z. Pan

Main category: cs.CL

TL;DR: FIDES is a new framework that improves attribution in QA by using contextually enhanced fact decomposition and evidence aggregation, outperforming SOTA methods by over 14%.

DetailsMotivation: Current question decomposition approaches for attribution in QA generate irrelevant/incomplete questions, lose facts during retrieval, and fail to aggregate evidence from different sources effectively.

Method: Uses two-stage faithful decomposition to break answers into sub-facts, retrieves evidence snippets, revises conflicting sub-facts, and aggregates evidence according to original sentences.

Result: Outperforms SOTA methods by over 14% across GPT-3.5-turbo, Gemini and Llama 70B series on six datasets with new Attr_auto-P metric.

Conclusion: FIDES provides a more effective framework for attributed QA through improved fact decomposition and evidence aggregation techniques.

Abstract: Attribution is crucial in question answering (QA) with Large Language Models (LLMs).SOTA question decomposition-based approaches use long form answers to generate questions for retrieving related documents. However, the generated questions are often irrelevant and incomplete, resulting in a loss of facts in retrieval.These approaches also fail to aggregate evidence snippets from different documents and paragraphs. To tackle these problems, we propose a new fact decomposition-based framework called FIDES (\textit{faithful context enhanced fact decomposition and evidence aggregation}) for attributed QA. FIDES uses a contextually enhanced two-stage faithful decomposition method to decompose long form answers into sub-facts, which are then used by a retriever to retrieve related evidence snippets. If the retrieved evidence snippets conflict with the related sub-facts, such sub-facts will be revised accordingly. Finally, the evidence snippets are aggregated according to the original sentences.Extensive evaluation has been conducted with six datasets, with an additionally proposed new metric called $Attr_{auto-P}$ for evaluating the evidence precision. FIDES outperforms the SOTA methods by over 14% in average with GPT-3.5-turbo, Gemini and Llama 70B series.

Weizhe Shi, Qiqi Wang, Yihong Pan, Qian Liu, Kaiqi Zhao

Main category: cs.CL

TL;DR: Proposes Judicial Opinion Generation task that simultaneously produces legal reasoning and sentencing decisions using LegalChainReasoner framework with structured legal chains for comprehensive case assessment.

DetailsMotivation: Current approaches separate legal reasoning and sentencing prediction into isolated subtasks, leading to inconsistency and failing to meet real-world judicial requirements. Manual knowledge curation methods have limited practical deployment.

Method: LegalChainReasoner framework that applies structured legal chains integrating factual premises, composite legal conditions, and sentencing conclusions for flexible knowledge injection and end-to-end opinion generation.

Result: Experiments on two real-world Chinese legal case datasets show the method outperforms baseline models.

Conclusion: The proposed approach addresses limitations of previous methods by ensuring consistency between reasoning and sentencing while providing practical judicial opinion generation capabilities.

Abstract: A criminal judicial opinion represents the judge’s disposition of a case, including the decision rationale and sentencing. Automatically generating such opinions can assist in analyzing sentencing consistency and provide judges with references to similar past cases. However, current research typically approaches this task by dividing it into two isolated subtasks: legal reasoning and sentencing prediction. This separation often leads to inconsistency between the reasoning and predictions, failing to meet real-world judicial requirements. Furthermore, prior studies rely on manually curated knowledge to enhance applicability, yet such methods remain limited in practical deployment. To address these limitations and better align with legal practice, we propose a new LegalAI task: Judicial Opinion Generation, which simultaneously produces both legal reasoning and sentencing decisions. To achieve this, we introduce LegalChainReasoner, a framework that applies structured legal chains to guide the model through comprehensive case assessments. By integrating factual premises, composite legal conditions, and sentencing conclusions, our approach ensures flexible knowledge injection and end-to-end opinion generation. Experiments on two real-world and open-source Chinese legal case datasets demonstrate that our method outperforms baseline models.

[46] CRMAgent: A Multi-Agent LLM System for E-Commerce CRM Message Template Generation

Yinzhu Quan, Xinrui Li, Ying Chen

Main category: cs.CL

TL;DR: CRMAgent is a multi-agent LLM system that helps merchants create persuasive marketing messages through three modes: learning from top-performing messages, adapting successful templates, and rule-based fallbacks, significantly improving marketing effectiveness.

DetailsMotivation: Most merchants struggle to write persuasive marketing copy due to lack of expertise and scalable tools, while only a few top performers excel at crafting effective outbound messages for CRM programs.

Method: A multi-agent system built on LLMs with three complementary modes: 1) group-based learning from merchant’s own top-performing messages, 2) retrieval-and-adaptation of successful templates with similar audience/voucher/product characteristics, 3) rule-based fallback for zero-shot rewriting when no references are available.

Result: Extensive experiments show CRMAgent consistently outperforms merchants’ original templates, delivering significant gains in both audience-match and marketing-effectiveness metrics.

Conclusion: CRMAgent provides an effective solution for merchants to generate high-quality message templates and actionable writing guidance, addressing the scalability and expertise gap in persuasive copywriting for e-commerce CRM channels.

Abstract: In e-commerce private-domain channels such as instant messaging and e-mail, merchants engage customers directly as part of their Customer Relationship Management (CRM) programmes to drive retention and conversion. While a few top performers excel at crafting outbound messages, most merchants struggle to write persuasive copy because they lack both expertise and scalable tools. We introduce CRMAgent, a multi-agent system built on large language models (LLMs) that generates high-quality message templates and actionable writing guidance through three complementary modes. First, group-based learning enables the agent to learn from a merchant’s own top-performing messages within the same audience segment and rewrite low-performing ones. Second, retrieval-and-adaptation fetches templates that share the same audience segment and exhibit high similarity in voucher type and product category, learns their successful patterns, and adapts them to the current campaign. Third, a rule-based fallback provides a lightweight zero-shot rewrite when no suitable references are available. Extensive experiments show that CRMAgent consistently outperforms merchants’ original templates, delivering significant gains in both audience-match and marketing-effectiveness metrics.

[47] CaresAI at BioCreative IX Track 1 – LLM for Biomedical QA

Reem Abdel-Salam, Mary Adewunmi, Modinat A. Abayomi

Main category: cs.CL

TL;DR: Fine-tuned LLaMA 3 8B on biomedical QA data shows strong semantic understanding but struggles with exact answer formatting, requiring two-stage inference for better evaluation alignment.

DetailsMotivation: Rigorous evaluation of LLMs is essential before deploying them in real-world biomedical applications, particularly for complex multi-hop question answering involving diseases, genes, and chemicals.

Method: Supervised fine-tuning of LLaMA 3 8B using curated biomedical datasets (BioASQ, MedQuAD, TREC) with three experimental setups: combined short/long answers, short only, and long only answers, plus a two-stage inference pipeline for precise answer extraction.

Result: Models achieved concept-level accuracy scores up to 0.8 but showed significantly lower Exact Match scores, especially in testing. Two-stage inference improved short-answer extraction but challenges remained in strict output formatting.

Conclusion: There’s a significant gap between semantic understanding and exact answer evaluation in biomedical LLMs, highlighting the need for further research in output control and post-processing strategies.

Abstract: Large language models (LLMs) are increasingly evident for accurate question answering across various domains. However, rigorous evaluation of their performance on complex question-answering (QA) capabilities is essential before deployment in real-world biomedical and healthcare applications. This paper presents our approach to the MedHopQA track of the BioCreative IX shared task, which focuses on multi-hop biomedical question answering involving diseases, genes, and chemicals. We adopt a supervised fine-tuning strategy leveraging LLaMA 3 8B, enhanced with a curated biomedical question-answer dataset compiled from external sources including BioASQ, MedQuAD, and TREC. Three experimental setups are explored: fine-tuning on combined short and long answers, short answers only, and long answers only. While our models demonstrate strong domain understanding, achieving concept-level accuracy scores of up to 0.8, their Exact Match (EM) scores remain significantly lower, particularly in the test phase. We introduce a two-stage inference pipeline for precise short-answer extraction to mitigate verbosity and improve alignment with evaluation metrics. Despite partial improvements, challenges persist in generating strictly formatted outputs. Our findings highlight the gap between semantic understanding and exact answer evaluation in biomedical LLM applications, motivating further research in output control and post-processing strategies.

[48] TMT: A Simple Way to Translate Topic Models Using Dictionaries

Felix Engl, Andreas Henrich

Main category: cs.CL

TL;DR: Topic Model Translation (TMT) enables transferring topic models between languages without requiring aligned corpora, metadata, or embeddings, making multilingual topic modeling accessible even with limited target language data.

DetailsMotivation: Multilingual topic modeling is challenging due to requirements for sophisticated algorithms, aligned corpora, and manual evaluation, especially when developers lack target language knowledge or have limited data availability.

Method: TMT (Topic Model Translation) is a novel technique that transfers topic models (like LDA) from one language to another without needing metadata, embeddings, or aligned corpora.

Result: Extensive evaluation shows TMT produces semantically coherent and consistent topic translations across languages, validated through both quantitative and qualitative methods.

Conclusion: TMT provides a robust and transparent solution for reusing topic models across languages, particularly valuable when large target language corpora are unavailable or manual translation is impractical.

Abstract: The training of topic models for a multilingual environment is a challenging task, requiring the use of sophisticated algorithms, topic-aligned corpora, and manual evaluation. These difficulties are further exacerbated when the developer lacks knowledge of the target language or is working in an environment with limited data, where only small or unusable multilingual corpora are available. Considering these challenges, we introduce Topic Model Translation (TMT), a novel, robust and transparent technique designed to transfer topic models (e.g., Latent Dirichlet Allocation (LDA) based topic models) from one language to another, without the need for metadata, embeddings, or aligned corpora. TMT enables the reuse of topic models across languages, making it especially suitable for scenarios where large corpora in the target language are unavailable or manual translation is infeasible. Furthermore, we evaluate TMT extensively using both quantitative and qualitative methods, demonstrating that it produces semantically coherent and consistent topic translations.

[49] Neural Models and Language Model Prompting for the Multidimensional Evaluation of Open-Ended Conversations

Michelle Elizabeth, Alicja Kasicka, Natalia Krawczyk, Magalie Ochs, Gwénolé Lecorvé, Justyna Gromada, Lina M. Rojas-Barahona

Main category: cs.CL

TL;DR: Small LM models for dialogue evaluation achieve modest results in DSTC-12 challenge, with regression/classification models showing high correlation on validation but dropping on test set due to score distribution differences.

DetailsMotivation: The growing number of generative AI-based dialogue systems requires effective evaluation methods, particularly with constrained model sizes (under 13B parameters).

Method: Two main strategies: 1) Using Language Models as evaluators through prompting, 2) Training encoder-based classification and regression models with fewer parameters.

Result: LM prompting achieved modest correlations with human judgments (ranked second on test set). Regression/classification models showed high correlation on validation set but performance decreased on test set due to different score ranges in annotations.

Conclusion: While smaller models can achieve reasonable performance, test set distribution differences highlight challenges in dialogue evaluation and the need for robust evaluation methods that generalize across different score distributions.

Abstract: The growing number of generative AI-based dialogue systems has made their evaluation a crucial challenge. This paper presents our contribution to this important problem through the Dialogue System Technology Challenge (DSTC-12, Track 1), where we developed models to predict dialogue-level, dimension-specific scores. Given the constraint of using relatively small models (i.e. fewer than 13 billion parameters) our work follows two main strategies: employing Language Models (LMs) as evaluators through prompting, and training encoder-based classification and regression models. Our results show that while LM prompting achieves only modest correlations with human judgments, it still ranks second on the test set, outperformed only by the baseline. The regression and classification models, with significantly fewer parameters, demonstrate high correlation for some dimensions on the validation set. Although their performance decreases on the test set, it is important to note that the test set contains annotations with significantly different score ranges for some of the dimensions with respect to the train and validation sets.

[50] Negative Matters: Multi-Granularity Hard-Negative Synthesis and Anchor-Token-Aware Pooling for Enhanced Text Embeddings

Tengyu Pan, Zhichao Duan, Zhenyu Li, Bowen Dong, Ning Liu, Xiuxing Li, Jianyong Wang

Main category: cs.CL

TL;DR: A framework using LLMs to generate multi-granularity hard negatives and anchor token aware pooling for improved text embeddings.

DetailsMotivation: Negative samples are crucial for contrastive learning in text embedding models, but existing methods lack diverse similarity levels for progressive learning.

Method: Multi-Granularity Hard-negative synthesis using LLMs to create diverse negative samples, combined with Anchor Token Aware pooling that weights important tokens based on LLM patterns.

Result: Achieved state-of-the-art performance on MTEB benchmark, outperforming existing synthesis strategies with both synthetic data and public retrieval datasets.

Conclusion: The proposed MGH framework and ATA pooling method effectively improve text embedding quality through progressive curriculum learning and better token weighting.

Abstract: Text embedding models are essential for various natural language processing tasks, enabling the effective encoding of semantic information into dense vector representations. These models are typically optimized using triplets of (query, positive, negative) data pairs for contrastive learning, where the negative samples play a critical role in enhancing the model’s ability to discern subtle semantic distinctions. In this work, we introduce a Multi-Granularity Hard-negative (MGH) synthesis framework that leverages large language models (LLMs) to generate diverse negative samples with varying levels of similarity with the query. This approach facilitates a coarse-to-fine curriculum learning strategy during supervised training, allowing the embedding model to progressively learn more nuanced semantic representations. Meanwhile, we propose an Anchor Token Aware (ATA) pooling method that assigns higher weights to anchor tokens based on aggregation patterns observed in LLMs, improving text embedding accuracy without increasing model complexity. Comprehensive experiments on the MTEB benchmark demonstrate that our methods achieve state-of-the-art performance, surpassing existing synthesis strategies both with synthetic data and when combined with public retrieval datasets.

[51] Prompting Away Stereotypes? Evaluating Bias in Text-to-Image Models for Occupations

Shaina Raza, Maximus Powers, Partha Pratim Saha, Mahveen Raza, Rizwan Qureshi

Main category: cs.CL

TL;DR: TTI models amplify social biases in occupational portrayals. Prompting can shift demographic representations but with model-specific effects - some diversify well, others overcorrect or show little responsiveness.

DetailsMotivation: Text-to-Image models risk amplifying harmful social biases in occupational portrayals, requiring systematic assessment and intervention strategies.

Method: Created benchmark of 5 occupational roles, tested 5 state-of-the-art TTI models with neutral vs fairness-aware prompts, annotated outputs for gender and race distribution analysis.

Result: Prompting can substantially shift demographic representations but effects are highly model-specific - some diversify effectively, others overcorrect into uniformity, some show little responsiveness.

Conclusion: Prompting shows promise but has limitations as fairness intervention, highlighting need for complementary model-level strategies alongside prompt engineering.

Abstract: Text-to-Image (TTI) models are powerful creative tools but risk amplifying harmful social biases. We frame representational societal bias assessment as an image curation and evaluation task and introduce a pilot benchmark of occupational portrayals spanning five socially salient roles (CEO, Nurse, Software Engineer, Teacher, Athlete). Using five state-of-the-art models: closed-source (DALLE 3, Gemini Imagen 4.0) and open-source (FLUX.1-dev, Stable Diffusion XL Turbo, Grok-2 Image), we compare neutral baseline prompts against fairness-aware controlled prompts designed to encourage demographic diversity. All outputs are annotated for gender (male, female) and race (Asian, Black, White), enabling structured distributional analysis. Results show that prompting can substantially shift demographic representations, but with highly model-specific effects: some systems diversify effectively, others overcorrect into unrealistic uniformity, and some show little responsiveness. These findings highlight both the promise and the limitations of prompting as a fairness intervention, underscoring the need for complementary model-level strategies. We release all code and data for transparency and reproducibility https://github.com/maximus-powers/img-gen-bias-analysis.

[52] Exploring and Mitigating Fawning Hallucinations in Large Language Models

Zixuan Shangguan, Yanjie Dong, Lanjun Wang, Xiaoyi Fan, Victor C. M. Leung, Xiping Hu

Main category: cs.CL

TL;DR: This paper addresses fawning hallucinations in LLMs where models prioritize alignment with deceptive prompts over factual accuracy, and proposes Collaborative Contrastive Decoding (CCD) to mitigate this issue without additional training.

DetailsMotivation: LLMs often generate inaccurate responses when aligning with deceptive/misleading prompts, prioritizing user perspective over factual truthfulness - a phenomenon called fawning hallucinations that needs mitigation.

Method: Proposes Collaborative Contrastive Decoding (CCD) which contrasts output distributions between induced deceptive inputs and transformed neutral inputs to reduce reliance on misleading information without requiring additional training.

Result: Extensive experiments show CCD effectively mitigates fawning hallucinations and improves factuality of generated responses across various NLP tasks.

Conclusion: The proposed CCD method successfully addresses fawning hallucinations in LLMs by contrasting output distributions, enhancing factual accuracy without the need for additional model training.

Abstract: Large language models (LLMs) have demonstrated exceptional proficiency in language understanding. However, when LLMs align their outputs with deceptive and/or misleading prompts, the generated responses could deviate from the de facto information. Such observations are known as fawning hallucinations, where the model prioritizes alignment with the input’s implied perspective over accuracy and truthfulness. In this work, we analyze fawning hallucinations in various natural language processing tasks and tailor the so-termed contrastive decoding method for fawning-hallucination mitigation. Specifically, we design two paradigms to generate corresponding deceptive and/or misleading inputs for the consistent fawning hallucinations induction. Then, we propose the collaborative contrastive decoding (CCD) to handle the fawning hallucinations across different tasks in LLMs. By contrasting the deviation in output distribution between induced and transformed neutral inputs, the proposed CCD can reduce reliance on deceptive and/or misleading information without requiring additional training. Extensive experiments demonstrate that the proposed CCD can effectively mitigate fawning hallucinations and improve the factuality of the generated responses over various tasks.

[53] EviNote-RAG: Enhancing RAG Models via Answer-Supportive Evidence Notes

Yuqin Dai, Guoqing Wang, Yuan Wang, Kairan Dou, Kaichen Zhou, Zhanwei Zhang, Shuo Yang, Fei Tang, Jun Yin, Pengyu Zeng, Zhenzhe Ying, Can Yi, Changhua Meng, Yuchen Zhou, Yongliang Shen, Shuai Lu

Main category: cs.CL

TL;DR: EviNote-RAG introduces a structured retrieve-note-answer pipeline that composes concise Supportive-Evidence Notes to filter noise and improve reasoning in open-domain QA, achieving state-of-the-art results with significant performance gains.

DetailsMotivation: To address limitations of conventional retrieve-then-answer paradigm: low signal-to-noise ratio in retrieved evidence and error accumulation in multi-hop reasoning when dealing with incomplete or noisy passages.

Method: Proposes an agentic RAG framework with a retrieve-note-answer pipeline where models compose Supportive-Evidence Notes (SENs) - concise human-like notes preserving only relevant information. Uses Evidence Quality Reward (EQR), an entailment-based signal to evaluate whether SENs logically support final answers.

Result: Outperforms strong baselines in accuracy, generalization, and training stability. Achieves state-of-the-art results with relative F1 gains: 20% on HotpotQA (+0.093), 40% on Bamboogle (+0.151), and 91% on 2Wiki (+0.256) through denser rewards and reduced verbosity.

Conclusion: EviNote-RAG enhances robustness and efficiency in open-domain QA by guiding models toward faithful reasoning through structured evidence distillation and quality reinforcement, effectively reducing noise impact while maintaining strong performance.

Abstract: Large Language Models (LLMs) empowered with retrieval mechanisms have achieved strong progress in open-domain question answering (QA). Yet, the conventional retrieve–then–answer paradigm often suffers from two key limitations: (1) low signal-to-noise ratio in retrieved evidence, where useful information is buried under irrelevant content, and (2) error accumulation in multi-hop reasoning when incomplete or noisy passages are involved. To address these challenges, we present EviNote-RAG, an agentic RAG framework that introduces a structured retrieve–note–answer pipeline. Instead of directly reasoning over raw retrievals, the model is trained to compose Supportive-Evidence Notes (SENs), concise, human-like notes that preserve only answer-relevant information, highlight uncertainty, and explicitly state when no useful evidence exists. This distillation process is further reinforced by the Evidence Quality Reward (EQR), an entailment-based signal that evaluates whether SENs logically support the final answer. Together, SENs and EQR guide the model toward faithful and robust reasoning, while reducing the impact of noise. Experiments on in-domain and out-of-domain QA benchmarks show that EviNote-RAG consistently outperforms strong baselines in accuracy, generalization, and training stability. In particular, it achieves state-of-the-art results while enhancing robustness and efficiency, yielding relative F1 gains of 20% on HotpotQA (+0.093), 40% on Bamboogle (+0.151), and 91% on 2Wiki (+0.256) via denser rewards and reduced verbosity.

[54] SeLeRoSa: Sentence-Level Romanian Satire Detection Dataset

Răzvan-Alexandru Smădu, Andreea Iuga, Dumitru-Clementin Cercel, Florin Pop

Main category: cs.CL

TL;DR: First Romanian sentence-level satire detection dataset with 13,873 annotated sentences, evaluated with LLMs showing current limitations in detecting satirical content.

DetailsMotivation: Satire and irony can be mistaken for factual reporting like fake news, but there's a lack of sentence-level datasets for satire detection, particularly for Romanian language.

Method: Created SeLeRoSa dataset with 13,873 manually annotated sentences across multiple domains. Evaluated LLMs in zero-shot and fine-tuning settings, plus transformer-based models for satire detection.

Result: Current LLMs and transformer models show limitations in sentence-level satire detection task, indicating room for improvement.

Conclusion: The research highlights the challenges in satire detection and opens new directions for improving models’ ability to distinguish satirical content from factual reporting.

Abstract: Satire, irony, and sarcasm are techniques typically used to express humor and critique, rather than deceive; however, they can occasionally be mistaken for factual reporting, akin to fake news. These techniques can be applied at a more granular level, allowing satirical information to be incorporated into news articles. In this paper, we introduce the first sentence-level dataset for Romanian satire detection for news articles, called SeLeRoSa. The dataset comprises 13,873 manually annotated sentences spanning various domains, including social issues, IT, science, and movies. With the rise and recent progress of large language models (LLMs) in the natural language processing literature, LLMs have demonstrated enhanced capabilities to tackle various tasks in zero-shot settings. We evaluate multiple baseline models based on LLMs in both zero-shot and fine-tuning settings, as well as baseline transformer-based models. Our findings reveal the current limitations of these models in the sentence-level satire detection task, paving the way for new research directions.

[55] Supervised In-Context Fine-Tuning for Generative Sequence Labeling

David Dukić, Goran Glavaš, Jan Šnajder

Main category: cs.CL

TL;DR: SIFT method combines in-context learning with supervised fine-tuning for generative sequence labeling, outperforming traditional approaches and showing that instructions are often unnecessary for strong performance.

DetailsMotivation: Sequence labeling tasks are typically handled by encoder-only models, but causal LLMs should theoretically outperform them due to rapid scaling. However, less work has focused on supervised generative approaches that are more natural for causal LLMs.

Method: Proposes supervised in-context fine-tuning (SIFT) which casts sequence labeling as constrained response generation, combining in-context learning from demonstrations with supervised fine-tuning.

Result: SIFT considerably outperforms both in-context learning and decoder-as-encoder fine-tuning baselines on standard sequence labeling tasks. Long context hinders performance but this can be mitigated by removing instructions.

Conclusion: Response-based generative task formulation is crucial for effective sequence labeling performance with LLMs, highlighting both strengths and limitations of using LLMs for sequence labeling tasks.

Abstract: Sequence labeling (SL) tasks, where labels are assigned to tokens, are abundant in NLP (e.g., named entity recognition and aspect-based sentiment analysis). Owing to the intuition that they require bidirectional context, SL tasks are commonly tackled with encoder-only models. Recent work also shows that removing the causal mask in fine-tuning enables decoder-based LLMs to become effective token classifiers. Less work, however, focused on (supervised) generative SL, a more natural setting for causal LLMs. Due to their rapid scaling, causal LLMs applied to SL are expected to outperform encoders, whose own development has stagnated. In this work, we propose supervised in-context fine-tuning (SIFT) for generative SL. SIFT casts SL tasks as constrained response generation, natural to LLMs, combining (1) in-context learning (ICL) from demonstrations with (2) supervised fine-tuning. SIFT considerably outperforms both ICL and decoder-as-encoder fine-tuning baselines on a range of standard SL tasks. We further find that although long context hinders the performance of generative SL in both ICL and SIFT, this deficiency can be mitigated by removing the instruction, as instructions are shown to be largely unnecessary for achieving strong SL performance with SIFT. Our findings highlight strengths and limitations of SL with LLMs, underscoring the importance of a response-based generative task formulation for effective SL performance.

[56] MedCOD: Enhancing English-to-Spanish Medical Translation of Large Language Models Using Enriched Chain-of-Dictionary Framework

Md Shahidul Salim, Lian Fu, Arav Adikesh Ramakrishnan, Zonghai Yao, Hong Yu

Main category: cs.CL

TL;DR: MedCOD framework integrates UMLS and LLM-KB knowledge to improve English-Spanish medical translation through structured prompting and fine-tuning, achieving state-of-the-art results across multiple LLMs.

DetailsMotivation: To enhance medical translation quality by integrating domain-specific structured knowledge from UMLS and LLM-KB into large language models, addressing the need for accurate medical terminology translation.

Method: Hybrid framework combining structured prompts with multilingual variants, medical synonyms, UMLS-derived definitions, and LoRA-based fine-tuning on a parallel corpus of 2,999 English-Spanish medical articles, evaluated on four open-source LLMs.

Result: Significant translation quality improvements across all models, with Phi-4 achieving BLEU 44.23, chrF++ 28.91, and COMET 0.863, outperforming GPT-4o and GPT-4o-mini baselines.

Conclusion: Structured knowledge integration through MedCOD framework effectively enhances LLM performance for medical translation, with both prompting and adaptation contributing independently to performance gains.

Abstract: We present MedCOD (Medical Chain-of-Dictionary), a hybrid framework designed to improve English-to-Spanish medical translation by integrating domain-specific structured knowledge into large language models (LLMs). MedCOD integrates domain-specific knowledge from both the Unified Medical Language System (UMLS) and the LLM-as-Knowledge-Base (LLM-KB) paradigm to enhance structured prompting and fine-tuning. We constructed a parallel corpus of 2,999 English-Spanish MedlinePlus articles and a 100-sentence test set annotated with structured medical contexts. Four open-source LLMs (Phi-4, Qwen2.5-14B, Qwen2.5-7B, and LLaMA-3.1-8B) were evaluated using structured prompts that incorporated multilingual variants, medical synonyms, and UMLS-derived definitions, combined with LoRA-based fine-tuning. Experimental results demonstrate that MedCOD significantly improves translation quality across all models. For example, Phi-4 with MedCOD and fine-tuning achieved BLEU 44.23, chrF++ 28.91, and COMET 0.863, surpassing strong baseline models like GPT-4o and GPT-4o-mini. Ablation studies confirm that both MedCOD prompting and model adaptation independently contribute to performance gains, with their combination yielding the highest improvements. These findings highlight the potential of structured knowledge integration to enhance LLMs for medical translation tasks.

[57] Structure and Destructure: Dual Forces in the Making of Knowledge Engines

Yihong Chen

Main category: cs.CL

TL;DR: This paper bridges structured (knowledge graphs) and unstructured (large language models) NLP paradigms through complementary structure and destructure forces, proposing a new approach for transparent and adaptable knowledge engines.

DetailsMotivation: To establish conceptual connections between the seemingly distinct structured and unstructured paradigms in NLP knowledge engine development, addressing the divergence between symbolic knowledge graphs and data-driven transformer models.

Method: Identifies two complementary forces - structure (organizing symbolic interactions) and destructure (periodic embedding resets for improved plasticity and generalization). Forms a new recipe combining these approaches.

Result: Develops a framework that bridges structured and unstructured paradigms, enabling knowledge engines to support transparent, controllable, and adaptable intelligent systems.

Conclusion: The thesis successfully establishes conceptual connections between structured and unstructured NLP paradigms through structure and destructure forces, providing a new foundation for developing general knowledge engines with improved transparency and adaptability.

Abstract: The making of knowledge engines in natural language processing has been shaped by two seemingly distinct paradigms: one grounded in structure, the other driven by massively available unstructured data. The structured paradigm leverages predefined symbolic interactions, such as knowledge graphs, as priors and designs models to capture them. In contrast, the unstructured paradigm centers on scaling transformer architectures with increasingly vast data and model sizes, as seen in modern large language models. Despite their divergence, this thesis seeks to establish conceptual connections bridging these paradigms. Two complementary forces, structure and destructure, emerge across both paradigms: structure organizes seen symbolic interactions, while destructure, through periodic embedding resets, improves model plasticity and generalization to unseen scenarios. These connections form a new recipe for developing general knowledge engines that can support transparent, controllable, and adaptable intelligent systems.

[58] RPRO:Ranked Preference Reinforcement Optimization for Enhancing Medical QA and Diagnostic Reasoning

Chia-Hsuan Hsu, Jun-En Ding, Hsin-Ling Hsu, Feng Liu, Fang-Ming Hung

Main category: cs.CL

TL;DR: RPRO combines reinforcement learning with preference-driven reasoning to improve clinical chain-of-thought accuracy in medical QA, outperforming larger models with a compact 1.1B parameter architecture.

DetailsMotivation: Existing LLMs generate medically unreliable reasoning chains that lack factual accuracy and clinical reliability in medical question answering tasks.

Method: Ranked Preference Reinforcement Optimization (RPRO) framework using groupwise ranking optimization based on Bradley-Terry model, KL-divergence regularization, task-adaptive reasoning templates, and probabilistic evaluation to align with clinical workflows.

Result: Consistent improvements on PubMedQA and MedQA-USMLE benchmarks; 1.1B parameter model outperforms larger 7B-13B models including medical-specialized variants.

Conclusion: Combining preference optimization with quality-driven refinement provides a scalable and effective approach for building clinically reliable medical LLMs.

Abstract: Medical question answering requires advanced reasoning that integrates domain knowledge with logical inference. However, existing large language models (LLMs) often generate reasoning chains that lack factual accuracy and clinical reliability. We propose Ranked Preference Reinforcement Optimization (RPRO), a novel framework that uniquely combines reinforcement learning with preference-driven reasoning refinement to enhance clinical chain-of-thought (CoT) performance. RPRO differentiates itself from prior approaches by employing task-adaptive reasoning templates and a probabilistic evaluation mechanism that aligns outputs with established clinical workflows, while automatically identifying and correcting low-quality reasoning chains. Unlike traditional pairwise preference methods, RPRO introduces a groupwise ranking optimization based on the Bradley-Terry model and incorporates KL-divergence regularization for stable training. Experiments on PubMedQA and MedQA-USMLE show consistent improvements over strong baselines. Remarkably, our 1.1B parameter model outperforms much larger 7B-13B models, including medical-specialized variants. These findings demonstrate that combining preference optimization with quality-driven refinement offers a scalable and effective approach to building more reliable, clinically grounded medical LLMs.

[59] Performance Analysis of Supervised Machine Learning Algorithms for Text Classification

Sadia Zaman Mishu, S M Rafiuddin

Main category: cs.CL

TL;DR: This paper compares various supervised machine learning classifiers for text classification, including an Artificial Neural Network with Back Propagation, and evaluates their accuracy on different labeled datasets.

DetailsMotivation: The growing demand for text classification in web searching, data mining, recommendation systems, and other information technology fields requires effective classification methods.

Method: The study applies standard supervised machine learning classifiers on labeled text documents, including an Artificial Neural Network model using Back Propagation Network, and compares their performance using benchmark approaches.

Result: Experimental analysis on real data reveals which classification models perform best in terms of accuracy for text classification tasks.

Conclusion: The research provides insights into the effectiveness of different supervised learning techniques for text classification, helping identify optimal models for various text classification applications.

Abstract: The demand for text classification is growing significantly in web searching, data mining, web ranking, recommendation systems, and so many other fields of information and technology. This paper illustrates the text classification process on different datasets using some standard supervised machine learning techniques. Text documents can be classified through various kinds of classifiers. Labeled text documents are used to classify the text in supervised classifications. This paper applies these classifiers on different kinds of labeled documents and measures the accuracy of the classifiers. An Artificial Neural Network (ANN) model using Back Propagation Network (BPN) is used with several other models to create an independent platform for labeled and supervised text classification process. An existing benchmark approach is used to analyze the performance of classification using labeled documents. Experimental analysis on real data reveals which model works well in terms of classification accuracy.

[60] Ranking of Bangla Word Graph using Graph-based Ranking Algorithms

S M Rafiuddin

Main category: cs.CL

TL;DR: This paper presents a method for ranking Bangla words using graph-based algorithms applied to word graphs constructed from text, with evaluation using F1 measure on Indian Language POS-tag Corpora.

DetailsMotivation: There is a lack of standard Bangla word databases and ranking methods, making it important to develop effective techniques for summarizing Bangla text and information retrieval through word ranking.

Method: Construct word graphs from Bangla text where words are vertices and relationships are edges, apply preprocessing steps, then use various graph-based ranking algorithms to calculate word importance.

Result: Experimental analysis on real data shows the accuracy of different ranking algorithms measured by F1 score, providing comparative performance evaluation.

Conclusion: The research successfully demonstrates a complete procedure for Bangla word ranking using graph-based approaches, offering valuable insights for text summarization and information retrieval in Bangla language processing.

Abstract: Ranking words is an important way to summarize a text or to retrieve information. A word graph is a way to represent the words of a sentence or a text as the vertices of a graph and to show the relationship among the words. It is also useful to determine the relative importance of a word among the words in the word-graph. In this research, the ranking of Bangla words are calculated, representing Bangla words from a text in a word graph using various graph based ranking algorithms. There is a lack of a standard Bangla word database. In this research, the Indian Language POS-tag Corpora is used, which has a rich collection of Bangla words in the form of sentences with their parts of speech tags. For applying a word graph to various graph based ranking algorithms, several standard procedures are applied. The preprocessing steps are done in every word graph and then applied to graph based ranking algorithms to make a comparison among these algorithms. This paper illustrate the entire procedure of calculating the ranking of Bangla words, including the construction of the word graph from text. Experimental result analysis on real data reveals the accuracy of each ranking algorithm in terms of F1 measure.

[61] We Politely Insist: Your LLM Must Learn the Persian Art of Taarof

Nikta Gohari Sadr, Sahar Heidariasl, Karine Megerdoomian, Laleh Seyyed-Kalantari, Ali Emami

Main category: cs.CL

TL;DR: TaarofBench is the first benchmark for evaluating LLM understanding of Persian taarof (ritual politeness), revealing 40-48% accuracy gaps compared to native speakers and showing that Western politeness metrics often violate taarof norms.

DetailsMotivation: LLMs struggle with culturally specific communication norms like Persian taarof, limiting their effectiveness in global contexts where such sophisticated ritual politeness systems exist.

Method: Created TaarofBench with 450 role-play scenarios across 12 social topics, validated by native speakers. Evaluated 5 frontier LLMs, conducted human study with 33 participants, and used supervised fine-tuning and Direct Preference Optimization for improvement.

Result: LLMs showed substantial cultural competence gaps with accuracy 40-48% below native speakers. Performance varied by topic, improved with Persian prompts, and showed gender asymmetries. Fine-tuning achieved 21.8-42.3% improvement in cultural alignment.

Conclusion: This work establishes a foundation for developing culturally aware LLMs that can better navigate complex social interactions, highlighting the limitations of Western politeness frameworks for non-Western cultural norms.

Abstract: Large language models (LLMs) struggle to navigate culturally specific communication norms, limiting their effectiveness in global contexts. We focus on Persian taarof, a social norm in Iranian interactions, which is a sophisticated system of ritual politeness that emphasizes deference, modesty, and indirectness, yet remains absent from existing cultural benchmarks. We introduce TaarofBench, the first benchmark for evaluating LLM understanding of taarof, comprising 450 role-play scenarios covering 12 common social interaction topics, validated by native speakers. Our evaluation of five frontier LLMs reveals substantial gaps in cultural competence, with accuracy rates 40-48% below native speakers when taarof is culturally appropriate. Performance varies between interaction topics, improves with Persian-language prompts, and exhibits gender-based asymmetries. We also show that responses rated “polite” by standard metrics often violate taarof norms, indicating the limitations of Western politeness frameworks. Through supervised fine-tuning and Direct Preference Optimization, we achieve 21.8% and 42.3% improvement in model alignment with cultural expectations. Our human study with 33 participants (11 native Persian, 11 heritage, and 11 non-Iranian speakers) forms baselines in varying degrees of familiarity with Persian norms. This work lays the foundation for developing diverse and culturally aware LLMs, enabling applications that better navigate complex social interactions.

[62] A Dynamic Fusion Model for Consistent Crisis Response

Xiaoying Song, Anirban Saha Anik, Eduardo Blanco, Vanessa Frias-Martinez, Lingzi Hong

Main category: cs.CL

TL;DR: Proposes a novel metric and fusion-based approach for maintaining stylistic consistency in AI-generated crisis response communications, outperforming baselines in both quality and style uniformity.

DetailsMotivation: Address the critical need for stylistic consistency in automated crisis communications to maintain trust with affected populations, as current methods often overlook this important factor.

Method: Two-stage process: 1) assesses style of candidate responses, 2) optimizes and integrates them through instance-level fusion to reduce stylistic variation while maintaining quality.

Result: Experimental results across multiple datasets show the approach consistently outperforms baselines in both response quality and stylistic uniformity.

Conclusion: The proposed fusion-based generation method effectively addresses the gap in maintaining stylistic consistency for crisis communications, enabling more trustworthy automated responses.

Abstract: In response to the urgent need for effective communication with crisis-affected populations, automated responses driven by language models have been proposed to assist in crisis communications. A critical yet often overlooked factor is the consistency of response style, which could affect the trust of affected individuals in responders. Despite its importance, few studies have explored methods for maintaining stylistic consistency across generated responses. To address this gap, we propose a novel metric for evaluating style consistency and introduce a fusion-based generation approach grounded in this metric. Our method employs a two-stage process: it first assesses the style of candidate responses and then optimizes and integrates them at the instance level through a fusion process. This enables the generation of high-quality responses while significantly reducing stylistic variation between instances. Experimental results across multiple datasets demonstrate that our approach consistently outperforms baselines in both response quality and stylistic uniformity.

[63] Speaking at the Right Level: Literacy-Controlled Counterspeech Generation with RAG-RL

Xiaoying Song, Anirban Saha Anik, Dibakar Barua, Pengcheng Luo, Junhua Ding, Lingzi Hong

Main category: cs.CL

TL;DR: A framework using RAG with RL to generate health misinformation counterspeech tailored to different literacy levels, outperforming uniform response approaches.

DetailsMotivation: Health misinformation online threatens public health, and existing counterspeech methods ignore audience health literacy levels, affecting accessibility and effectiveness.

Method: Controlled-Literacy framework combining retrieval-augmented generation (RAG) with reinforcement learning (RL) to retrieve knowledge aligned with specific literacy levels and optimize counterspeech using subjective user preferences and objective readability rewards.

Result: Experiment results show Controlled-Literacy outperforms baselines by generating more accessible and user-preferred counterspeech.

Conclusion: This research contributes to more equitable public health communication by improving counterspeech accessibility and comprehension for diverse health literacy levels.

Abstract: Health misinformation spreading online poses a significant threat to public health. Researchers have explored methods for automatically generating counterspeech to health misinformation as a mitigation strategy. Existing approaches often produce uniform responses, ignoring that the health literacy level of the audience could affect the accessibility and effectiveness of counterspeech. We propose a Controlled-Literacy framework using retrieval-augmented generation (RAG) with reinforcement learning (RL) to generate tailored counterspeech adapted to different health literacy levels. In particular, we retrieve knowledge aligned with specific health literacy levels, enabling accessible and factual information to support generation. We design a reward function incorporating subjective user preferences and objective readability-based rewards to optimize counterspeech to the target health literacy level. Experiment results show that Controlled-Literacy outperforms baselines by generating more accessible and user-preferred counterspeech. This research contributes to more equitable and impactful public health communication by improving the accessibility and comprehension of counterspeech to health misinformation.

Abdessalam Bouchekif, Samer Rashwani, Heba Sbahi, Shahd Gaben, Mutez Al-Khatib, Mohammed Ghaly

Main category: cs.CL

TL;DR: Evaluation of 7 LLMs on Islamic inheritance law knowledge using 1,000 multiple-choice questions, revealing significant performance gaps between models with o3 and Gemini 2.5 achieving >90% accuracy while others scored below 50%.

DetailsMotivation: To assess the knowledge and reasoning capabilities of Large Language Models in the specialized domain of Islamic inheritance law ('ilm al-mawarith), which requires understanding complex legal contexts and computational distribution of shares.

Method: Used a benchmark of 1,000 multiple-choice questions covering diverse inheritance scenarios to test models’ ability to understand inheritance context and compute share distributions prescribed by Islamic jurisprudence. Evaluated seven LLMs including o3, Gemini 2.5, ALLaM, Fanar, LLaMA, and Mistral.

Result: Significant performance gap observed: o3 and Gemini 2.5 achieved accuracies above 90%, while ALLaM, Fanar, LLaMA, and Mistral scored below 50%. Error analysis revealed recurring failure patterns including misunderstandings of inheritance scenarios, incorrect application of legal rules, and insufficient domain knowledge.

Conclusion: The study highlights limitations in LLMs’ ability to handle structured legal reasoning in Islamic jurisprudence and suggests directions for improving performance in domain-specific legal reasoning tasks.

Abstract: This paper evaluates the knowledge and reasoning capabilities of Large Language Models in Islamic inheritance law, known as ‘ilm al-mawarith. We assess the performance of seven LLMs using a benchmark of 1,000 multiple-choice questions covering diverse inheritance scenarios, designed to test models’ ability to understand the inheritance context and compute the distribution of shares prescribed by Islamic jurisprudence. The results reveal a significant performance gap: o3 and Gemini 2.5 achieved accuracies above 90%, whereas ALLaM, Fanar, LLaMA, and Mistral scored below 50%. These disparities reflect important differences in reasoning ability and domain adaptation. We conduct a detailed error analysis to identify recurring failure patterns across models, including misunderstandings of inheritance scenarios, incorrect application of legal rules, and insufficient domain knowledge. Our findings highlight limitations in handling structured legal reasoning and suggest directions for improving performance in Islamic legal reasoning. Code: https://github.com/bouchekif/inheritance_evaluation

[65] A Paradigm Gap in Urdu

Farah Adeeba, Rajesh Bhatt

Main category: cs.CL

TL;DR: The perfective form of the -ya: kar construction in Urdu became ungrammatical due to a morphosyntactic conflict between nominative subject requirements and ergative case assignment rules in transitive perfectives.

DetailsMotivation: To investigate why the perfective form of the -ya: kar construction, which was freely used in 19th century Urdu literature, became sharply ungrammatical in modern Urdu and Hindi.

Method: Historical text analysis, large-scale corpus study to confirm absence of perfective forms, and subjective evaluation tasks with native speakers to judge grammaticality.

Result: Perfective forms are starkly absent in modern usage and native speakers judge them as highly unnatural, confirming the paradigm gap emerged from a fundamental morphosyntactic conflict.

Conclusion: The perfective form became unstable due to conflict between construction requirements (nominative subject + invariant participle) and core grammatical rules (ergative case assignment for transitive perfectives), leading to its functional replacement and entrenchment as ungrammatical in modern grammar.

Abstract: In this paper, we document a paradigm gap in the combinatorial possibilities of verbs and aspect in Urdu: the perfective form of the -ya: kar construction (e.g. ro-ya: ki: cry-Pfv do.Pfv) is sharply ungrammatical in modern Urdu and Hindi, despite being freely attested in 19th century literature. We investigate this diachronic shift through historical text analysis, a large-scale corpus study which confirms the stark absence of perfective forms and subjective evaluation tasks with native speakers, who judge perfective examples as highly unnatural. We argue that this gap arose from a fundamental morphosyntactic conflict: the construction’s requirement for a nominative subject and an invariant participle clashes with the core grammatical rule that transitive perfective assign ergative case. This conflict rendered the perfective form unstable, and its functional replacement by other constructions allowed the gap to become entrenched in the modern grammar.

[66] Privacy-Preserving Reasoning with Knowledge-Distilled Parametric Retrieval Augmented Generation

Jinwen Chen, Hainan Zhang, Liang Pang, Yongxin Tong, Haibo Zhou, Yuan Zhan, Wei Lin, Zhiming Zheng

Main category: cs.CL

TL;DR: DistilledPRAG is a privacy-preserving RAG system that encodes documents as LoRA parameters instead of uploading plaintext, achieving RAG-level performance while protecting data privacy through knowledge distillation and structural alignment.

DetailsMotivation: Current RAG systems risk private data leakage by uploading plaintext documents to the cloud. Parametric RAG (PRAG) addresses privacy but suffers from high inference latency and poor generalization on out-of-distribution data due to reliance on synthetic QA pairs.

Method: Proposes DistilledPRAG with three key components: 1) Synthesizes QA pairs from single and multi-documents for cross-document reasoning, 2) Masks plaintext documents and translates them to LoRA via parameter generator while maintaining RAG document structure, 3) Trains parameter generator using synthetic QA data to match standard RAG’s hidden states and output logits.

Result: Experiments on four QA datasets show DistilledPRAG outperforms baselines in accuracy and generalizes well on out-of-distribution data.

Conclusion: DistilledPRAG successfully achieves high-efficiency parameterization while maintaining RAG-level performance, solving the critical challenge of privacy-preserving reasoning without exposing original documents.

Abstract: The current RAG system requires uploading plaintext documents to the cloud, risking private data leakage. Parametric RAG (PRAG) addresses this by encoding documents as LoRA within LLMs, enabling reasoning without exposing raw content. However, it still faces two issues: (1) PRAG demands synthesizing QA pairs and fine-tuning LLM for each individual document to create its corresponding LoRA, leading to unacceptable inference latency. (2) The performance of PRAG relies solely on synthetic QA data, lacking internal alignment with standard RAG, resulting in poor generalization on out-of-distribution(OOD) inputs. Therefore, achieving high-efficiency parameterization while maintaining RAG-level performance remains a critical challenge for privacy-preserving reasoning. In this paper, we propose DistilledPRAG, a generalizable knowledge-distilled parametric RAG model aligned with standard RAG in document structure and parameter activation. We first synthesize QA pairs from single and multi-documents to enhance cross-document reasoning. Then, we mask the plaintext documents with a special token and translate them to LoRA via a parameter generator, maintaining the standard RAG document structure. Finally, guided by synthetic QA data, we train the parameter generator to match standard RAG’s hidden states and output logits, enabling RAG-style reasoning without original documents. Experiments on four QA datasets show that DistilledPRAG outperforms baselines in accuracy and generalizes well on OOD data.

[67] REFRAG: Rethinking RAG based Decoding

Xiaoqiang Lin, Aritra Ghosh, Bryan Kian Hsiang Low, Anshumali Shrivastava, Vijai Mohan

Main category: cs.CL

TL;DR: REFRAG is an efficient decoding framework that reduces latency in RAG applications by exploiting the sparse attention patterns in retrieved passages, achieving 30.85% speedup without performance loss.

DetailsMotivation: RAG applications suffer from high latency and memory demands due to processing long contexts with mostly irrelevant retrieved passages, creating a trade-off between knowledge enrichment and system efficiency.

Method: REFRAG compresses, senses, and expands the context by exploiting the block-diagonal attention patterns in RAG contexts, eliminating unnecessary computations during decoding.

Result: Achieves 30.85% time-to-first-token acceleration (3.75x improvement over previous work) without perplexity loss, and extends context size by 16x while maintaining accuracy across diverse tasks.

Conclusion: REFRAG effectively addresses the latency-efficiency trade-off in RAG systems by leveraging structural sparsity in attention patterns, providing substantial speedups with no accuracy degradation.

Abstract: Large Language Models (LLMs) have demonstrated remarkable capabilities in leveraging extensive external knowledge to enhance responses in multi-turn and agentic applications, such as retrieval-augmented generation (RAG). However, processing long-context inputs introduces significant system latency and demands substantial memory for the key-value cache, resulting in reduced throughput and a fundamental trade-off between knowledge enrichment and system efficiency. While minimizing latency for long-context inputs is a primary objective for LLMs, we contend that RAG require specialized consideration. In RAG, much of the LLM context consists of concatenated passages from retrieval, with only a small subset directly relevant to the query. These passages often exhibit low semantic similarity due to diversity or deduplication during re-ranking, leading to block-diagonal attention patterns that differ from those in standard LLM generation tasks. Based on this observation, we argue that most computations over the RAG context during decoding are unnecessary and can be eliminated with minimal impact on performance. To this end, we propose REFRAG, an efficient decoding framework that compresses, senses, and expands to improve latency in RAG applications. By exploiting the sparsity structure, we demonstrate a 30.85 the time-to-first-token acceleration (3.75 improvement to previous work) without loss in perplexity. In addition, our optimization framework for large context enables REFRAG to extend the context size of LLMs by 16. We provide rigorous validation of REFRAG across diverse long-context tasks, including RAG, multi-turn conversations, and long document summarization, spanning a wide range of datasets. Experimental results confirm that REFRAG delivers substantial speedup with no loss in accuracy compared to LLaMA models and other state-of-the-art baselines across various context sizes.

[68] Natural Context Drift Undermines the Natural Language Understanding of Large Language Models

Yulong Wu, Viktor Schlegel, Riza Batista-Navarro

Main category: cs.CL

TL;DR: LLM performance on question answering declines as reading passages naturally diverge from pretraining versions, even when all necessary information remains present.

DetailsMotivation: To investigate how natural evolution of context paragraphs affects question answering performance in generative Large Language Models.

Method: Proposed a framework for curating naturally evolved, human-edited variants of reading passages from QA benchmarks and analyzing LLM performance across semantic similarity scores that quantify alignment with pretraining content.

Result: LLM performance declines as passages diverge from pretraining versions - average accuracy on BoolQ drops by over 30% from highest to lowest similarity bins, with slopes exceeding 70 across several models.

Conclusion: Natural text evolution poses a significant challenge to LLMs’ language understanding capabilities, as performance deteriorates even when all required information remains available.

Abstract: How does the natural evolution of context paragraphs affect question answering in generative Large Language Models (LLMs)? To investigate this, we propose a framework for curating naturally evolved, human-edited variants of reading passages from contemporary QA benchmarks and for analyzing LLM performance across a range of semantic similarity scores, which quantify how closely each variant aligns with content seen during pretraining. Using this framework, we evaluate six QA datasets and eight LLMs with publicly available training data. Our experiments reveal that LLM performance declines as reading passages naturally diverge from the versions encountered during pretraining-even when the question and all necessary information remains present at inference time. For instance, average model accuracy on BoolQ drops by over 30% from the highest to lowest similarity bins, with slopes exceeding 70 across several LLMs. These findings suggest that natural text evolution poses a significant challenge to the language understanding capabilities of LLMs.

[69] Dream-Coder 7B: An Open Diffusion Language Model for Code

Zhihui Xie, Jiacheng Ye, Lin Zheng, Jiahui Gao, Jingwei Dong, Zirui Wu, Xueliang Zhao, Shansan Gong, Xin Jiang, Zhenguo Li, Lingpeng Kong

Main category: cs.CL

TL;DR: Dream-Coder 7B is an open-source discrete diffusion language model for code generation that adaptively chooses decoding strategies based on task complexity, achieving 21.4% pass@1 on LiveCodeBench.

DetailsMotivation: Traditional autoregressive models decode strictly left-to-right, which may not be optimal for different coding tasks. The authors aim to develop a model that can adapt its decoding strategy based on the specific coding task requirements.

Method: Adapted a pretrained AR checkpoint to discrete diffusion framework with continuous-time weighted cross-entropy objective. Used supervised fine-tuning with random truncation and padding penalty, followed by reinforcement learning with verifiable rewards on curated prompt sets.

Result: Achieved 21.4% pass@1 on LiveCodeBench (2410-2505) and demonstrated competitive performance on HumanEval, MBPP, BigCodeBench, and CRUXEval benchmarks.

Conclusion: Dream-Coder 7B exhibits emergent any-order generation capabilities and adapts decoding strategies based on coding tasks, showing strong performance across multiple code generation benchmarks while being released as open-source for reproducibility.

Abstract: We present Dream-Coder 7B, an open-source discrete diffusion language model for code generation that exhibits emergent any-order generation capabilities. Unlike traditional autoregressive (AR) models that decode strictly left-to-right, Dream-Coder 7B adaptively determines its decoding strategy based on the coding task: sketch-first generation for complex algorithms, left-to-right generation for straightforward completions, and interleaved reasoning generation for code understanding tasks. We adapt a pretrained AR checkpoint to a discrete diffusion frameworks with a continuous-time weighted cross-entropy objective. Our post-training recipe comprises (i) supervised fine-tuning, where we mitigate padding pathologies via random truncation and a padding penalty to improve sample efficiency and stabilize generation; and (ii) reinforcement learning with verifiable rewards over a curated high-quality prompt set drawn from open-source datasets, using a tailored reinforcement learning recipe for diffusion language models. The resulting Dream-Coder 7B Instruct attains 21.4% pass@1 on LiveCodeBench (2410–2505) and demonstrates competitive performance on HumanEval, MBPP, BigCodeBench, and CRUXEval. We release Dream-Coder-7B and Dream-Coder-7B-Instruct checkpoints, training recipes, preprocessing pipelines, and inference code to facilitate reproducibility and further research.

[70] Zero-shot Cross-lingual NER via Mitigating Language Difference: An Entity-aligned Translation Perspective

Zhihao Zhang, Sophia Yat Mei Lee, Dong Zhang, Shoushan Li, Guodong Zhou

Main category: cs.CL

TL;DR: Proposes Entity-Aligned Translation (EAT) approach using LLMs for cross-lingual NER from Latin to non-Latin script languages through dual-translation strategy and Wikipedia fine-tuning.

DetailsMotivation: Existing zero-shot CL-NER approaches work well for Latin script languages but degrade for non-Latin script languages (Chinese, Japanese) due to deep structural differences, creating performance gaps.

Method: Entity-Aligned Translation (EAT) approach using large language models with dual-translation strategy to align entities between non-Latin script languages and English, plus fine-tuning LLMs on multilingual Wikipedia data.

Result: The method aims to improve entity alignment from source to target languages for non-Latin script languages, though specific performance metrics are not provided in the abstract.

Conclusion: EAT approach addresses the structural challenges in cross-lingual NER for non-Latin script languages by leveraging LLMs and translation alignment strategies.

Abstract: Cross-lingual Named Entity Recognition (CL-NER) aims to transfer knowledge from high-resource languages to low-resource languages. However, existing zero-shot CL-NER (ZCL-NER) approaches primarily focus on Latin script language (LSL), where shared linguistic features facilitate effective knowledge transfer. In contrast, for non-Latin script language (NSL), such as Chinese and Japanese, performance often degrades due to deep structural differences. To address these challenges, we propose an entity-aligned translation (EAT) approach. Leveraging large language models (LLMs), EAT employs a dual-translation strategy to align entities between NSL and English. In addition, we fine-tune LLMs using multilingual Wikipedia data to enhance the entity alignment from source to target languages.

[71] Joint Information Extraction Across Classical and Modern Chinese with Tea-MOELoRA

Xuemei Tang, Chengxi Yan, Jinghang Gu, Chu-Ren Huang

Main category: cs.CL

TL;DR: Tea-MOELoRA is a parameter-efficient multi-task framework combining LoRA with Mixture-of-Experts to handle Chinese IE tasks across different temporal domains without performance interference.

DetailsMotivation: Fine-tuning a single model on heterogeneous Chinese IE tasks across Classical and Modern documents causes interference and reduced performance due to task and temporal domain differences.

Method: Combines LoRA with Mixture-of-Experts design, where multiple low-rank LoRA experts specialize in different IE tasks and eras, using a task-era-aware router to dynamically allocate expert contributions.

Result: Outperforms both single-task and joint LoRA baselines, demonstrating effective leveraging of task and temporal knowledge.

Conclusion: Tea-MOELoRA provides an effective parameter-efficient solution for multi-task Chinese IE across diverse temporal domains without performance degradation.

Abstract: Chinese information extraction (IE) involves multiple tasks across diverse temporal domains, including Classical and Modern documents. Fine-tuning a single model on heterogeneous tasks and across different eras may lead to interference and reduced performance. Therefore, in this paper, we propose Tea-MOELoRA, a parameter-efficient multi-task framework that combines LoRA with a Mixture-of-Experts (MoE) design. Multiple low-rank LoRA experts specialize in different IE tasks and eras, while a task-era-aware router mechanism dynamically allocates expert contributions. Experiments show that Tea-MOELoRA outperforms both single-task and joint LoRA baselines, demonstrating its ability to leverage task and temporal knowledge effectively.

[72] Enhancing Large Language Model for Knowledge Graph Completion via Structure-Aware Alignment-Tuning

Yu Liu, Yanan Cao, Xixun Lin, Yanmin Shang, Shi Wang, Shirui Pan

Main category: cs.CL

TL;DR: SAT framework enhances LLMs for knowledge graph completion through structure-aware alignment-tuning, addressing representation space inconsistency and task-specific instruction redundancy.

DetailsMotivation: Existing LLM-enhanced KGC methods ignore the inconsistent representation spaces between natural language and graph structures, and require separate instructions for different tasks leading to duplicate work.

Method: Proposes hierarchical knowledge alignment through multi-task contrastive learning to align graph embeddings with natural language space, and structural instruction tuning using unified graph instruction with lightweight knowledge adapter.

Result: SAT significantly outperforms state-of-the-art methods on two KGC tasks across four benchmark datasets, with 8.7% to 29.8% improvements in link prediction.

Conclusion: The SAT framework effectively bridges the gap between natural language and graph structures while providing a unified approach for multiple KGC tasks, demonstrating substantial performance gains.

Abstract: Knowledge graph completion (KGC) aims to infer new knowledge and make predictions from knowledge graphs. Recently, large language models (LLMs) have exhibited remarkable reasoning capabilities. LLM-enhanced KGC methods primarily focus on designing task-specific instructions, achieving promising advancements. However, there are still two critical challenges. First, existing methods often ignore the inconsistent representation spaces between natural language and graph structures. Second, most approaches design separate instructions for different KGC tasks, leading to duplicate works and time-consuming processes. To address these challenges, we propose SAT, a novel framework that enhances LLMs for KGC via structure-aware alignment-tuning. Specifically, we first introduce hierarchical knowledge alignment to align graph embeddings with the natural language space through multi-task contrastive learning. Then, we propose structural instruction tuning to guide LLMs in performing structure-aware reasoning over KGs, using a unified graph instruction combined with a lightweight knowledge adapter. Experimental results on two KGC tasks across four benchmark datasets demonstrate that SAT significantly outperforms state-of-the-art methods, especially in the link prediction task with improvements ranging from 8.7% to 29.8%.

[73] Modular Techniques for Synthetic Long-Context Data Generation in Language Model Training and Evaluation

Seganrasan Subramanian, Abhigya Verma

Main category: cs.CL

TL;DR: A framework for generating synthetic long-context datasets using LLMs through prompt-based interactions, supporting multiple training objectives and generation paradigms to address the lack of quality long-context training data.

DetailsMotivation: Progress in long-context LLM capabilities is constrained by the absence of high-quality, diverse, and verifiable long-context datasets suitable for both training and evaluation.

Method: Modular framework using prompt-based interaction with LLMs to generate synthetic data, supporting SFT, DPO, and GRPO objectives through four core paradigms: multi-turn dialogues, document-grounded pairs, verifiable tasks, and long-context reasoning examples.

Result: The approach enables scalable, controllable, and purpose-aligned dataset creation through templated prompting, model-agnostic architecture, and metadata-enriched outputs.

Conclusion: This framework facilitates advancing long-context capabilities in LLMs by providing a systematic way to generate high-quality synthetic training data for various long-context applications.

Abstract: The ability of large language models (LLMs) to process and reason over long textual inputs is critical for a wide range of real-world applications. However, progress in this area is significantly constrained by the absence of high-quality, diverse, and verifiable long-context datasets suitable for both training and evaluation. This work introduces a modular, extensible framework for synthetic long-context data generation via prompt-based interaction with LLMs. The framework supports multiple training and alignment objectives, including Supervised Fine-Tuning (SFT), Direct Preference Optimization (DPO), and Group Relative Policy Optimization (GRPO). It encompasses four core generation paradigms: multi-turn conversational dialogues, document-grounded input-output pairs, verifiable instruction-response tasks, and long-context reasoning examples. Through templated prompting, a model-agnostic architecture, and metadata-enriched outputs, the proposed approach facilitates scalable, controllable, and purpose-aligned dataset creation for advancing long-context capabilities in LLMs.

[74] Statutory Construction and Interpretation for Artificial Intelligence

Luxi He, Nimra Nadeem, Michel Liao, Howard Chen, Danqi Chen, Mariano-Florentino Cuéllar, Peter Henderson

Main category: cs.CL

TL;DR: The paper addresses interpretive ambiguity in AI systems governed by natural language principles, proposing a computational framework inspired by legal mechanisms to improve consistency and stability in rule interpretation and application.

DetailsMotivation: AI systems increasingly rely on natural language principles, but face underexplored challenges with interpretive ambiguity similar to legal systems. Unlike legal systems with institutional safeguards, AI alignment pipelines lack protections against inconsistent interpretations of the same rules, leading to unstable model behavior.

Method: The authors propose a computational framework mirroring two legal mechanisms: (1) a rule refinement pipeline that minimizes interpretive disagreement by revising ambiguous rules (analogous to agency rulemaking), and (2) prompt-based interpretive constraints that reduce inconsistency in rule application (analogous to legal canons guiding judicial discretion).

Result: The framework was evaluated on a 5,000-scenario subset of the WildChat dataset, showing that both interventions significantly improve judgment consistency across a panel of reasonable interpreters.

Conclusion: This approach represents a first step toward systematically managing interpretive ambiguity, which is essential for building more robust, law-following AI systems that can handle natural language principles more consistently and reliably.

Abstract: AI systems are increasingly governed by natural language principles, yet a key challenge arising from reliance on language remains underexplored: interpretive ambiguity. As in legal systems, ambiguity arises both from how these principles are written and how they are applied. But while legal systems use institutional safeguards to manage such ambiguity, such as transparent appellate review policing interpretive constraints, AI alignment pipelines offer no comparable protections. Different interpretations of the same rule can lead to inconsistent or unstable model behavior. Drawing on legal theory, we identify key gaps in current alignment pipelines by examining how legal systems constrain ambiguity at both the rule creation and rule application steps. We then propose a computational framework that mirrors two legal mechanisms: (1) a rule refinement pipeline that minimizes interpretive disagreement by revising ambiguous rules (analogous to agency rulemaking or iterative legislative action), and (2) prompt-based interpretive constraints that reduce inconsistency in rule application (analogous to legal canons that guide judicial discretion). We evaluate our framework on a 5,000-scenario subset of the WildChat dataset and show that both interventions significantly improve judgment consistency across a panel of reasonable interpreters. Our approach offers a first step toward systematically managing interpretive ambiguity, an essential step for building more robust, law-following AI systems.

[75] Efficient Large Language Models with Zero-Shot Adjustable Acceleration

Sajjad Kachuee, Mohammad Sharifkhani

Main category: cs.CL

TL;DR: Zero-Shot Adjustable Acceleration method enables dynamic hardware usage adjustment during LLM inference without additional fine-tuning, achieving up to 11x speedup.

DetailsMotivation: Balancing computational efficiency and performance in LLM applications is challenging, requiring optimization of acceleration after fine-tuning and during inference.

Method: A novel training and inference approach that dynamically adjusts hardware usage during inference without requiring additional fine-tuning.

Result: Achieves up to 11x speedup compared to baseline across multiple classification and text generation tasks, enabling wide range of acceleration in zero-shot manner.

Conclusion: The proposed Zero-Shot Adjustable Acceleration method provides an efficient architecture for LLM deployment with significant speed improvements without additional training overhead.

Abstract: Using Large Language Models (LLMs) in real-world applications presents significant challenges, particularly in balancing computational efficiency and performance. Optimizing acceleration after the fine-tuning phase and during inference is crucial for building an efficient architecture. This paper introduces Zero-Shot Adjustable Acceleration, a novel training and inference method that dynamically adjusts hardware usage during inference without requiring additional fine-tuning. The proposed approach is applied to newly developed models and evaluated across multiple classification and text generation tasks. Experimental results demonstrate that the method enables a wide range of acceleration in a zero-shot manner and achieves up to a 11x speedup compared to the baseline.

[76] Mitigating Catastrophic Forgetting in Continual Learning through Model Growth

Ege Süalp, Mina Rezaei

Main category: cs.CL

TL;DR: Growth-based pretraining (Stack LLM) shows modest improvements in mitigating catastrophic forgetting compared to standard LLMs, with better retention in reading comprehension and stable bias ratios, though both models still suffer from forgetting in reasoning tasks.

DetailsMotivation: Catastrophic forgetting is a major challenge in continual learning for LLMs, where models lose prior knowledge when fine-tuned on new tasks. Model growth strategies using smaller models to train larger ones show promise but their impact on forgetting remains under-explored.

Method: Evaluated growth-based pretraining via transformer stacking (Stack LLM) against a standard LLM baseline across sequential fine-tuning tasks involving domain knowledge, reasoning, reading comprehension, and bias evaluation.

Result: Both models improved in domain knowledge but showed degradation in reasoning and reading comprehension. Stack LLM consistently showed less degradation, especially in reading comprehension. In bias evaluation, baseline LLM became more neutral while Stack LLM maintained steady bias ratio around 60-61%.

Conclusion: Growth-based pretraining provides modest improvements in resisting catastrophic forgetting with better retention capabilities, though trade-offs remain in handling social biases and complete forgetting prevention.

Abstract: Catastrophic forgetting is a significant challenge in continual learning, in which a model loses prior knowledge when it is fine-tuned on new tasks. This problem is particularly critical for large language models (LLMs) undergoing continual learning, as retaining performance across diverse domains is important for their general utility. In this paper, we explore model growth, a promising strategy that leverages smaller models to expedite and structure the training of larger ones for mitigating the catastrophic forgetting problem. Although growth-based pretraining, particularly via transformer stacking, has shown promise in accelerating convergence, its impact on forgetting remains under-explored. Therefore, we evaluate whether growth-based models can retain previously learned capabilities more effectively across a sequence of fine-tuning tasks involving domain knowledge, reasoning, reading comprehension, and bias. Our findings show that both models – one trained with growth (Stack LLM) and one without (LLM) – exhibit improvements in domain knowledge. However, reasoning and reading comprehension degrade over time, indicating signs of catastrophic forgetting. Stack LLM consistently shows less degradation, especially in reading comprehension, suggesting enhanced retention capabilities. Interestingly, in bias evaluation, the baseline LLM becomes progressively more neutral with continued fine-tuning, while Stack LLM maintains a steady bias ratio around 60–61%. These results indicate that growth-based pretraining may deliver modest improvements in resisting catastrophic forgetting, though trade-offs remain in handling social biases.

[77] DaMoC: Efficiently Selecting the Optimal Large Language Model for Fine-tuning Domain Taks Based on Data and Model Compression

Wei Huang, Huang Wei, Yinggui Wang

Main category: cs.CL

TL;DR: DaMoC framework helps quickly select optimal LLMs for fine-tuning by compressing data through filtering and token optimization, and compressing models via layer pruning and sparse merging, achieving 20x training time savings.

DetailsMotivation: Large language models struggle with domain-specific tasks and require fine-tuning, but selecting the best model from many open-source options is challenging and time-consuming.

Method: Two-level approach: 1) Data compression through systematic filtering (distribution/quality/hybrid methods), token compression, and LLM-based text rewriting; 2) Model compression using layer similarity scoring for pruning and sparse merging to preserve capabilities.

Result: Extensive experiments on medical Q&A, financial Q&A, general Q&A, and reading comprehension datasets show the framework can select optimal LLMs while saving approximately 20-fold in training time.

Conclusion: DaMoC provides an effective framework for rapid LLM selection and fine-tuning optimization through combined data and model compression techniques, significantly reducing computational costs while maintaining performance.

Abstract: Large language models (LLMs) excel in general tasks but struggle with domain-specific ones, requiring fine-tuning with specific data. With many open-source LLMs available, selecting the best model for fine-tuning downstream tasks is challenging, primarily focusing on how to quickly identify the optimal LLM. We introduce a Data and Model Compression Framework (DaMoC) that addresses this challenge by: 1) Data Level: A systematic categorization of data filtering methodologies for LLMs is first established, classifying them into three distinct paradigms: (1) distribution-aware methods, (2) quality-aware methods, and (3) hybrid approaches considering both dimensions. Further, we enhance the density of key tokens in the text achieving token compression. Subsequently, we use an LLM to iterative rewrite the text to optimize its expression. 2) Model Level: We use layer similarity scores to assess each layer’s importance and remove those with lower importance. Then, we introduce a sparse merging paradigm to preserve as much of the original model’s capability as possible. Extensive experiments on four datasets, medical Q&A, financial Q&A, general Q&A, and reading comprehension, show that we can select the optimal LLM while saving approximately 20-fold in training time.

[78] Rethinking the Chain-of-Thought: The Roles of In-Context Learning and Pre-trained Priors

Hao Yang, Zhiyu Yang, Yunjie Zhang, Shanyi Zhu, Lin Yang

Main category: cs.CL

TL;DR: This paper investigates how Chain-of-Thought reasoning works by analyzing the dual relationship between in-context learning and pretrained priors, revealing key insights about model behavior and performance improvements.

DetailsMotivation: Despite growing interest in Chain-of-Thought reasoning, its underlying mechanisms remain unclear, prompting the need to understand how it works from the perspective of in-context learning and pretrained priors.

Method: The researchers conducted fine-grained lexical-level analysis of rationales, incrementally introduced noisy exemplars to test model balancing, and investigated prompt engineering effects on inducing slow thinking in LLMs.

Result: Three key findings: (1) Models learn reasoning structures quickly but rely heavily on pretrained priors; (2) Sufficient exemplars shift decision-making to in-context signals while misleading prompts cause instability; (3) Long Chain-of-Thought prompting improves downstream task performance.

Conclusion: Chain-of-Thought reasoning involves a complex interplay between in-context learning and pretrained knowledge, with prompt engineering capable of inducing more deliberate reasoning processes that enhance model performance.

Abstract: Chain-of-Thought reasoning has emerged as a pivotal methodology for enhancing model inference capabilities. Despite growing interest in Chain-of-Thought reasoning, its underlying mechanisms remain unclear. This paper explores the working mechanisms of Chain-of-Thought reasoning from the perspective of the dual relationship between in-context learning and pretrained priors. We first conduct a fine-grained lexical-level analysis of rationales to examine the model’s reasoning behavior. Then, by incrementally introducing noisy exemplars, we examine how the model balances pretrained priors against erroneous in-context information. Finally, we investigate whether prompt engineering can induce slow thinking in large language models. Our extensive experiments reveal three key findings: (1) The model not only quickly learns the reasoning structure at the lexical level but also grasps deeper logical reasoning patterns, yet it heavily relies on pretrained priors. (2) Providing sufficient exemplars shifts the model’s decision-making from pretrained priors to in-context signals, while misleading prompts introduce instability. (3) Long Chain-of-Thought prompting can induce the model to generate longer reasoning chains, thereby improving its performance on downstream tasks.

[79] Annotation and modeling of emotions in a textual corpus: an evaluative approach

Jonas Noblet

Main category: cs.CL

TL;DR: This paper demonstrates that language models can effectively model emotional annotation variability in text and distinguish emotional situations using evaluative criteria, despite significant human annotation disagreement.

DetailsMotivation: Emotion in textual manifestations remains an open research area, and traditional approaches need complementary perspectives. The evaluative framework for emotion annotation is underutilized but offers valuable insights.

Method: Used language models trained on manually annotated industrial corpus following evaluative emotion approach. Analyzed statistical trends in annotations despite significant disagreement among human annotators.

Result: Language models successfully modeled the labeling process and showed that annotation variability is driven by underlying linguistic features. Models demonstrated capability to distinguish emotional situations based on evaluative criteria.

Conclusion: Evaluative approach to emotion annotation provides valuable complementary perspective to traditional methods. Language models can effectively capture and model emotional annotation patterns despite human disagreement, revealing stable statistical trends in emotional expression.

Abstract: Emotion is a crucial phenomenon in the functioning of human beings in society. However, it remains a widely open subject, particularly in its textual manifestations. This paper examines an industrial corpus manually annotated following an evaluative approach to emotion. This theoretical framework, which is currently underutilized, offers a different perspective that complements traditional approaches. Noting that the annotations we collected exhibit significant disagreement, we hypothesized that they nonetheless follow stable statistical trends. Using language models trained on these annotations, we demonstrate that it is possible to model the labeling process and that variability is driven by underlying linguistic features. Conversely, our results indicate that language models seem capable of distinguishing emotional situations based on evaluative criteria.

[80] Culture is Everywhere: A Call for Intentionally Cultural Evaluation

Juhyun Oh, Inha Cha, Michael Saxon, Hyunseung Lim, Shaily Bhatt, Alice Oh

Main category: cs.CL

TL;DR: Current LLM cultural evaluation methods focus on trivia and static facts, but need to shift to examining cultural assumptions embedded in all evaluation aspects through participatory methodologies.

DetailsMotivation: Existing cultural evaluation approaches for LLMs are inadequate as they treat culture as isolated trivia through multiple-choice questions, neglecting the pluralistic and interactive nature of culture and overlooking how cultural assumptions permeate even neutral evaluation settings.

Method: Proposes intentionally cultural evaluation - systematically examining cultural assumptions in all evaluation aspects, characterizing culturally contingent considerations, emphasizing researcher positionality, and using HCI-inspired participatory methodologies involving communities in evaluation design.

Result: A framework for moving beyond current benchmarking practices, discovering unknown important applications, and fostering inclusive, culturally aligned NLP research through systematic examination of cultural embeddedness in evaluations.

Conclusion: Cultural evaluation of LLMs requires fundamental shift from trivia-centered paradigm to examining cultural assumptions throughout evaluation processes, with emphasis on researcher positionality and community participation for more inclusive and culturally aligned AI systems.

Abstract: The prevailing trivia-centered paradigm'' for evaluating the cultural alignment of large language models (LLMs) is increasingly inadequate as these models become more advanced and widely deployed. Existing approaches typically reduce culture to static facts or values, testing models via multiple-choice or short-answer questions that treat culture as isolated trivia. Such methods neglect the pluralistic and interactive realities of culture, and overlook how cultural assumptions permeate even ostensibly neutral’’ evaluation settings. In this position paper, we argue for \textbf{intentionally cultural evaluation}: an approach that systematically examines the cultural assumptions embedded in all aspects of evaluation, not just in explicitly cultural tasks. We systematically characterize the what, how, and circumstances by which culturally contingent considerations arise in evaluation, and emphasize the importance of researcher positionality for fostering inclusive, culturally aligned NLP research. Finally, we discuss implications and future directions for moving beyond current benchmarking practices, discovering important applications that we don’t know exist, and involving communities in evaluation design through HCI-inspired participatory methodologies.

[81] TableZoomer: A Collaborative Agent Framework for Large-scale Table Question Answering

Sishi Xiong, Ziyang He, Zhongjiang He, Yu Zhao, Changzai Pan, Jie Zhang, Zhenhe Wu, Shuangyong Song, Yongxiang Li

Main category: cs.CL

TL;DR: TableZoomer is an LLM-powered agent framework that addresses table QA challenges through structured schema representation, query-aware table zooming, and Program-of-Thoughts execution, achieving significant accuracy improvements over traditional methods.

DetailsMotivation: LLMs face challenges in industrial table QA applications including structural heterogeneity, target data localization difficulties, and complex reasoning bottlenecks that limit their practical deployment.

Method: Three key innovations: 1) Structured table schema instead of verbalized tables, 2) Query-aware table zooming mechanism with column selection and entity linking, 3) Program-of-Thoughts strategy converting queries to executable code, integrated with ReAct paradigm for iterative reasoning.

Result: Achieved 19.34% accuracy improvement on DataBench dataset and 25% improvement on TableBench Fact Checking task compared to conventional PoT methods when implemented with Qwen3-8B-Instruct LLM.

Conclusion: TableZoomer framework maintains usability advantages while substantially enhancing performance and scalability across tables of varying scales, making LLMs more effective for industrial table QA applications.

Abstract: While large language models (LLMs) have shown promise in the table question answering (TQA) task through prompt engineering, they face challenges in industrial applications, including structural heterogeneity, difficulties in target data localization, and bottlenecks in complex reasoning. To address these limitations, this paper presents TableZoomer, a novel LLM-powered, programming-based agent framework. It introduces three key innovations: (1) replacing the original fully verbalized table with structured table schema to bridge the semantic gap and reduce computational complexity; (2) a query-aware table zooming mechanism that dynamically generates sub-table schema through column selection and entity linking, significantly improving target localization efficiency; and (3) a Program-of-Thoughts (PoT) strategy that transforms queries into executable code to mitigate numerical hallucination. Additionally, we integrate the reasoning workflow with the ReAct paradigm to enable iterative reasoning. Extensive experiments demonstrate that our framework maintains the usability advantages while substantially enhancing performance and scalability across tables of varying scales. When implemented with the Qwen3-8B-Instruct LLM, TableZoomer achieves accuracy improvements of 19.34% and 25% over conventional PoT methods on the large-scale DataBench dataset and the small-scale Fact Checking task of TableBench dataset, respectively.

[82] Can Smaller LLMs do better? Unlocking Cross-Domain Potential through Parameter-Efficient Fine-Tuning for Text Summarization

Anum Afzal, Mehul Kumawat, Florian Matthes

Main category: cs.CL

TL;DR: Parameter-efficient fine-tuning (PEFT) techniques enable LLMs to adapt to low-resource domains without full retraining, achieving better performance than few-shot learning and even larger models.

DetailsMotivation: LLMs struggle with domain adaptation to new, low-resource domains where labeled data is scarce, and full fine-tuning is computationally expensive.

Method: Benchmarked six PEFT techniques on Llama-3-8B-Instruct using 14 datasets across Scientific, Medical, Legal, and News domains for text summarization, exploring within-domain and cross-domain adapters.

Result: Within-domain adapters outperformed few-shot learning and larger Llama-3-70B-Instruct models. Cross-domain adapters and strategic combinations leveraged linguistic similarities for better low-resource performance.

Conclusion: PEFT techniques provide efficient domain adaptation for LLMs in low-resource settings, with within-domain adapters being most effective and cross-domain approaches offering viable alternatives.

Abstract: Large Language Models (LLMs), being generic task solvers, are versatile. However, despite the vast amount of data they are trained on, there are speculations about their adaptation capabilities to a new domain. Additionally, the simple fine-tuning of the model to incorporate knowledge of a new domain is computationally expensive and time-consuming. This becomes more challenging when the domain in question is also low-resource, and labeled data is unavailable. We leverage parameter-efficient fine-tuning techniques (PEFTs) on high-resource datasets to address these challenges to improve performance on unseen low-resource domains. Throughout our experiments, we evaluate whether intrinsic linguistic commonalities between datasets can be leveraged for efficient domain adaptation. We benchmark six PEFTs with \texttt{Llama-3-8B-Instruct} on 14 training datasets from the Scientific, Medical, Legal, and News domains for a Text Summarization task. Our experiments show that for low-resource domains, inference using Within-Domain Adapters can achieve better performance than Few-Shot as well as a much larger \texttt{Llama-3-70B-Instruct}. Lastly, in the absence of Within-Domain Adapters, we explore the concept of using Cross-Domain Adapters as well as the strategic combinations of adapters to leverage intrinsic language similarities across domains, facilitating better adaptability and performance in low-resource settings.

[83] LongCat-Flash Technical Report

Meituan LongCat Team, Bayan, Bei Li, Bingye Lei, Bo Wang, Bolin Rong, Chao Wang, Chao Zhang, Chen Gao, Chen Zhang, Cheng Sun, Chengcheng Han, Chenguang Xi, Chi Zhang, Chong Peng, Chuan Qin, Chuyu Zhang, Cong Chen, Congkui Wang, Dan Ma, Daoru Pan, Defei Bu, Dengchang Zhao, Deyang Kong, Dishan Liu, Feiye Huo, Fengcun Li, Fubao Zhang, Gan Dong, Gang Liu, Gang Xu, Ge Li, Guoqiang Tan, Guoyuan Lin, Haihang Jing, Haomin Fu, Haonan Yan, Haoxing Wen, Haozhe Zhao, Hong Liu, Hongmei Shi, Hongyan Hao, Hongyin Tang, Huantian Lv, Hui Su, Jiacheng Li, Jiahao Liu, Jiahuan Li, Jiajun Yang, Jiaming Wang, Jian Yang, Jianchao Tan, Jiaqi Sun, Jiaqi Zhang, Jiawei Fu, Jiawei Yang, Jiaxi Hu, Jiayu Qin, Jingang Wang, Jiyuan He, Jun Kuang, Junhui Mei, Kai Liang, Ke He, Kefeng Zhang, Keheng Wang, Keqing He, Liang Gao, Liang Shi, Lianhui Ma, Lin Qiu, Lingbin Kong, Lingtong Si, Linkun Lyu, Linsen Guo, Liqi Yang, Lizhi Yan, Mai Xia, Man Gao, Manyuan Zhang, Meng Zhou, Mengxia Shen, Mingxiang Tuo, Mingyang Zhu, Peiguang Li, Peng Pei, Peng Zhao, Pengcheng Jia, Pingwei Sun, Qi Gu, Qianyun Li, Qingyuan Li, Qiong Huang, Qiyuan Duan, Ran Meng, Rongxiang Weng, Ruichen Shao, Rumei Li, Shizhe Wu, Shuai Liang, Shuo Wang, Suogui Dang, Tao Fang, Tao Li, Tefeng Chen, Tianhao Bai, Tianhao Zhou, Tingwen Xie, Wei He, Wei Huang, Wei Liu, Wei Shi, Wei Wang, Wei Wu, Weikang Zhao, Wen Zan, Wenjie Shi, Xi Nan, Xi Su, Xiang Li, Xiang Mei, Xiangyang Ji, Xiangyu Xi, Xiangzhou Huang, Xianpeng Li, Xiao Fu, Xiao Liu, Xiao Wei, Xiaodong Cai, Xiaolong Chen, Xiaoqing Liu, Xiaotong Li, Xiaowei Shi, Xiaoyu Li, Xili Wang, Xin Chen, Xing Hu, Xingyu Miao, Xinyan He, Xuemiao Zhang, Xueyuan Hao, Xuezhi Cao, Xunliang Cai, Xurui Yang, Yan Feng, Yang Bai, Yang Chen, Yang Yang, Yaqi Huo, Yerui Sun, Yifan Lu, Yifan Zhang, Yipeng Zang, Yitao Zhai, Yiyang Li, Yongjing Yin, Yongkang Lv, Yongwei Zhou, Yu Yang, Yuchen Xie, Yueqing Sun, Yuewen Zheng, Yuhua Wei, Yulei Qian, Yunfan Liang, Yunfang Tai, Yunke Zhao, Zeyang Yu, Zhao Zhang, Zhaohua Yang, Zhenchao Zhang, Zhikang Xia, Zhiye Zou, Zhizhao Zeng, Zhongda Su, Zhuofan Chen, Zijian Zhang, Ziwen Wang, Zixu Jiang, Zizhe Zhao, Zongyu Wang, Zunhai Su

Main category: cs.CL

TL;DR: LongCat-Flash is a 560B parameter MoE language model with novel Zero-computation Experts and Shortcut-connected MoE designs for computational efficiency, achieving 100+ TPS inference at $0.70 per million tokens, with strong performance in agentic tasks.

DetailsMotivation: Address the need for scalable efficiency in large language models by optimizing computational resource allocation and improving inference throughput while maintaining competitive performance.

Method: Uses Mixture-of-Experts architecture with two novel designs: Zero-computation Experts for dynamic budget allocation and Shortcut-connected MoE for computation-communication overlap. Includes comprehensive scaling framework with hyperparameter transfer, model-growth initialization, and multi-pronged stability suite.

Result: Trained on 20+ trillion tokens in 30 days, achieves 100+ tokens per second inference at $0.70 per million output tokens. Demonstrates competitive performance among leading models with exceptional strengths in agentic tasks.

Conclusion: LongCat-Flash successfully combines scalable architectural design with infrastructure optimizations to deliver an efficient, high-performance language model with strong agentic capabilities, made available as open-source for community research.

Abstract: We introduce LongCat-Flash, a 560-billion-parameter Mixture-of-Experts (MoE) language model designed for both computational efficiency and advanced agentic capabilities. Stemming from the need for scalable efficiency, LongCat-Flash adopts two novel designs: (a) Zero-computation Experts, which enables dynamic computational budget allocation and activates 18.6B-31.3B (27B on average) per token depending on contextual demands, optimizing resource usage. (b) Shortcut-connected MoE, which enlarges the computation-communication overlap window, demonstrating notable gains in inference efficiency and throughput compared to models of a comparable scale. We develop a comprehensive scaling framework for large models that combines hyperparameter transfer, model-growth initialization, a multi-pronged stability suite, and deterministic computation to achieve stable and reproducible training. Notably, leveraging the synergy among scalable architectural design and infrastructure efforts, we complete model training on more than 20 trillion tokens within 30 days, while achieving over 100 tokens per second (TPS) for inference at a cost of $0.70 per million output tokens. To cultivate LongCat-Flash towards agentic intelligence, we conduct a large-scale pre-training on optimized mixtures, followed by targeted mid- and post-training on reasoning, code, and instructions, with further augmentation from synthetic data and tool use tasks. Comprehensive evaluations demonstrate that, as a non-thinking foundation model, LongCat-Flash delivers highly competitive performance among other leading models, with exceptional strengths in agentic tasks. The model checkpoint of LongCat-Flash is open-sourced to foster community research. LongCat Chat: https://longcat.ai Hugging Face: https://huggingface.co/meituan-longcat GitHub: https://github.com/meituan-longcat

Jihyung Lee, Daehui Kim, Seonjeong Hwang, Hyounghun Kim, Gary Lee

Main category: cs.CL

TL;DR: KoBLEX is a Korean legal QA benchmark for evaluating provision-grounded multi-hop reasoning, with ParSeR method achieving significant performance improvements over baselines.

DetailsMotivation: Existing legal benchmarks fail to evaluate open-ended and provision-grounded question answering, creating a need for better evaluation of LLMs' legal reasoning capabilities.

Method: Created KoBLEX benchmark with 226 scenario-based QA instances using LLM-human pipeline. Proposed ParSeR method with parametric provisions and three-stage retrieval for legal reasoning.

Result: ParSeR outperforms strong baselines, achieving +37.91 higher F1 and +30.81 higher LF-Eval compared to standard retrieval with GPT-4o.

Conclusion: The proposed KoBLEX benchmark and ParSeR method effectively address legal reasoning evaluation and demonstrate superior performance in provision-grounded legal QA tasks.

Abstract: Large Language Models (LLM) have achieved remarkable performances in general domains and are now extending into the expert domain of law. Several benchmarks have been proposed to evaluate LLMs’ legal capabilities. However, these benchmarks fail to evaluate open-ended and provision-grounded Question Answering (QA). To address this, we introduce a Korean Benchmark for Legal EXplainable QA (KoBLEX), designed to evaluate provision-grounded, multi-hop legal reasoning. KoBLEX includes 226 scenario-based QA instances and their supporting provisions, created using a hybrid LLM-human expert pipeline. We also propose a method called Parametric provision-guided Selection Retrieval (ParSeR), which uses LLM-generated parametric provisions to guide legally grounded and reliable answers. ParSeR facilitates multi-hop reasoning on complex legal questions by generating parametric provisions and employing a three-stage sequential retrieval process. Furthermore, to better evaluate the legal fidelity of the generated answers, we propose Legal Fidelity Evaluation (LF-Eval). LF-Eval is an automatic metric that jointly considers the question, answer, and supporting provisions and shows a high correlation with human judgments. Experimental results show that ParSeR consistently outperforms strong baselines, achieving the best results across multiple LLMs. Notably, compared to standard retrieval with GPT-4o, ParSeR achieves +37.91 higher F1 and +30.81 higher LF-Eval. Further analyses reveal that ParSeR efficiently delivers consistent performance across reasoning depths, with ablations confirming the effectiveness of ParSeR.

[85] Can Large Language Models Master Complex Card Games?

Wei Wang, Fuqing Bie, Junzhe Chen, Dan Zhang, Shiyu Huang, Evgeny Kharlamov, Jie Tang

Main category: cs.CL

TL;DR: LLMs can master complex card games through fine-tuning on quality gameplay data, achieving performance comparable to strong game AIs while maintaining general capabilities when balanced with instruction data.

DetailsMotivation: To explore whether large language models can achieve similar success in complex games as specialized AI systems like AlphaGo, given LLMs' remarkable capabilities across various tasks.

Method: Systematic assessment of LLM learning capabilities across eight diverse card games, evaluating fine-tuning impact on gameplay data, and examining model retention of general capabilities.

Result: LLMs approach strong game AI performance through supervised fine-tuning, master multiple complex card games simultaneously (with performance benefits for similar games), but experience general capability decline that can be mitigated with general instruction data.

Conclusion: LLMs demonstrate strong learning ability and versatility in complex games, showing potential to master multiple games while maintaining general capabilities through balanced training approaches.

Abstract: Complex games have long been an important benchmark for testing the progress of artificial intelligence algorithms. AlphaGo, AlphaZero, and MuZero have defeated top human players in Go and Chess, garnering widespread societal attention towards artificial intelligence. Concurrently, large language models (LLMs) have exhibited remarkable capabilities across various tasks, raising the question of whether LLMs can achieve similar success in complex games. In this paper, we explore the potential of LLMs in mastering complex card games. We systematically assess the learning capabilities of LLMs across eight diverse card games, evaluating the impact of fine-tuning on high-quality gameplay data, and examining the models’ ability to retain general capabilities while mastering these games. Our findings indicate that: (1) LLMs can approach the performance of strong game AIs through supervised fine-tuning on high-quality data, (2) LLMs can master multiple complex card games simultaneously, with performance augmentation for games with similar rules and conflicts for dissimilar ones, and (3) LLMs experience a decline in general capabilities when mastering complex games, but this decline can be mitigated by integrating a certain amount of general instruction data. The evaluation results demonstrate strong learning ability and versatility of LLMs.

[86] Reasoning Vectors: Transferring Chain-of-Thought Capabilities via Task Arithmetic

Mohammad Zbeeb, Hasan Abed Al Kader Hammoud, Bernard Ghanem

Main category: cs.CL

TL;DR: Reasoning capabilities can be extracted as task vectors from RL-trained models and transferred to other models through simple arithmetic operations, significantly improving reasoning performance across multiple benchmarks.

DetailsMotivation: To reduce the costly optimization required for large language models to master complex reasoning tasks by extracting and reusing reasoning capabilities from existing models rather than training from scratch.

Method: Extract a reasoning vector as the parameter difference between two identically initialized models (one fine-tuned with SFT, the other with GRPO on the same dataset): v_reason = θ_GRPO - θ_SFT, then add this vector to compatible instruction-tuned models.

Result: Consistent performance improvements across diverse reasoning benchmarks: GSM8K (+4.9%), HumanEval (+4.3%), SciQ (+1.7%), and BigBenchHard (+12.3% for 1.5B model). Performance improvements persist under adversarial conditions, while subtracting the vector causes significant degradation (-11.8% on GSM8K).

Conclusion: Reasoning capabilities developed through expensive training can be extracted from open-source models and reused through simple tensor arithmetic, offering a practical way to enhance models by recycling prior computational investments.

Abstract: Large language models often require costly optimization, such as reinforcement learning, to master complex reasoning tasks. This work demonstrates that reasoning ability, once learned, can be extracted and transferred between models as a compact task vector. We source two publicly available, identically initialized Qwen2.5 models, one fine-tuned with supervised fine-tuning (SFT) and the other with group relative policy optimization (GRPO) on the same dataset. From these, we extract a reasoning vector: $v_{\text{reason}} = \theta_{\text{GRPO}} - \theta_{\text{SFT}}$. We hypothesize that this vector captures the reasoning capability instilled by reinforcement learning while factoring out shared knowledge from the SFT process. When added to compatible instruction-tuned models through simple arithmetic, this vector consistently improves performance across diverse reasoning benchmarks: GSM8K (+4.9%), HumanEval (+4.3%), SciQ (+1.7%), and BigBenchHard (+12.3% for the 1.5B model). The performance improvements persist under adversarial conditions. Conversely, subtracting the vector causes significant performance degradation (-11.8% on GSM8K), demonstrating the vector’s strong contribution to the model’s reasoning abilities. This work shows how reasoning capabilities, typically developed through expensive training, can be extracted from existing open-source models and reused through simple tensor arithmetic, offering a practical way to enhance models by recycling prior computational investments.

[87] WATCHED: A Web AI Agent Tool for Combating Hate Speech by Expanding Data

Paloma Piot, Diego Sánchez, Javier Parapar

Main category: cs.CL

TL;DR: WATCHED is an AI-powered chatbot that helps content moderators detect and explain hate speech decisions using LLMs, BERT classifiers, slang lookup, and policy checks.

DetailsMotivation: Online harms like hate speech threaten user safety and trust in social media platforms, requiring tools that combine automated speed with human judgment and provide clear explanations.

Method: Built as an AI agent system using Large Language Models with specialized tools: compares posts with hate speech examples, uses BERT classifier, looks up slang via Urban Dictionary, generates chain-of-thought reasoning, and checks platform guidelines.

Result: Achieves state-of-the-art performance with macro F1 score of 0.91, surpassing existing methods in hate speech detection.

Conclusion: The tool effectively supports collaboration between AI and human moderators by providing explainable hate speech detection grounded in precedent and policy, helping reduce online harms.

Abstract: Online harms are a growing problem in digital spaces, putting user safety at risk and reducing trust in social media platforms. One of the most persistent forms of harm is hate speech. To address this, we need tools that combine the speed and scale of automated systems with the judgment and insight of human moderators. These tools should not only find harmful content but also explain their decisions clearly, helping to build trust and understanding. In this paper, we present WATCHED, a chatbot designed to support content moderators in tackling hate speech. The chatbot is built as an Artificial Intelligence Agent system that uses Large Language Models along with several specialised tools. It compares new posts with real examples of hate speech and neutral content, uses a BERT-based classifier to help flag harmful messages, looks up slang and informal language using sources like Urban Dictionary, generates chain-of-thought reasoning, and checks platform guidelines to explain and support its decisions. This combination allows the chatbot not only to detect hate speech but to explain why content is considered harmful, grounded in both precedent and policy. Experimental results show that our proposed method surpasses existing state-of-the-art methods, reaching a macro F1 score of 0.91. Designed for moderators, safety teams, and researchers, the tool helps reduce online harms by supporting collaboration between AI and human oversight.

Serwar Basch, Ilia Kuznetsov, Tom Hope, Iryna Gurevych

Main category: cs.CL

TL;DR: A framework for creating cross-document link datasets using semi-synthetic data and human evaluation, combining retrieval models with LLMs to achieve 78% human approval rate.

DetailsMotivation: Lack of efficient methods to create training and evaluation datasets for cross-document links limits automated assistance in understanding fine-grained document relations.

Method: Domain-agnostic framework that generates semi-synthetic datasets, performs automatic evaluation to shortlist best linking approaches, then conducts extensive human evaluation on natural text pairs using retrieval models combined with LLMs.

Result: Combining retrieval models with LLMs achieves 78% link approval from human raters, more than doubling the precision of strong retrievers alone. Applied successfully in peer review and news domains.

Conclusion: The framework enables systematic study of cross-document understanding across applications, with resulting datasets supporting tasks like media framing and peer review analysis.

Abstract: Understanding fine-grained relations between documents is crucial for many application domains. However, the study of automated assistance is limited by the lack of efficient methods to create training and evaluation datasets of cross-document links. To address this, we introduce a new domain-agnostic framework for selecting a best-performing approach and annotating cross-document links in a new domain from scratch. We first generate and validate semi-synthetic datasets of interconnected documents. This data is used to perform automatic evaluation, producing a shortlist of best-performing linking approaches. These approaches are then used in an extensive human evaluation study, yielding performance estimates on natural text pairs. We apply our framework in two distinct domains – peer review and news – and show that combining retrieval models with LLMs achieves 78% link approval from human raters, more than doubling the precision of strong retrievers alone. Our framework enables systematic study of cross-document understanding across application scenarios, and the resulting novel datasets lay foundation for numerous cross-document tasks like media framing and peer review. We make the code, data, and annotation protocols openly available.

[89] LLMs cannot spot math errors, even when allowed to peek into the solution

KV Aditya Srivatsa, Kaushal Kumar Maurya, Ekaterina Kochmar

Main category: cs.CL

TL;DR: LLMs struggle to locate first error steps in student math solutions despite strong math performance. Proposed method generates intermediate corrected solutions to improve error detection.

DetailsMotivation: Large language models perform well on math problems but fail at meta-reasoning tasks like identifying errors in student stepwise solutions, which is crucial for educational applications.

Method: Propose generating intermediate corrected student solutions that align more closely with the original student’s solution to help improve error localization performance.

Result: Experiments on VtG and PRM800K datasets show state-of-the-art LLMs struggle with error step localization even when given reference solutions.

Conclusion: The proposed approach of generating intermediate corrected solutions shows promise for improving LLM performance on error detection in stepwise math solutions.

Abstract: Large language models (LLMs) demonstrate remarkable performance on math word problems, yet they have been shown to struggle with meta-reasoning tasks such as identifying errors in student solutions. In this work, we investigate the challenge of locating the first error step in stepwise solutions using two error reasoning datasets: VtG and PRM800K. Our experiments show that state-of-the-art LLMs struggle to locate the first error step in student solutions even when given access to the reference solution. To that end, we propose an approach that generates an intermediate corrected student solution, aligning more closely with the original student’s solution, which helps improve performance.

[90] Vis-CoT: A Human-in-the-Loop Framework for Interactive Visualization and Intervention in LLM Chain-of-Thought Reasoning

Kaviraj Pather, Elena Hadjigeorgiou, Arben Krasniqi, Claire Schmit, Irina Rusu, Marc Pons, Kabir Khan

Main category: cs.CL

TL;DR: Vis-CoT is an interactive framework that converts chain-of-thought reasoning into visual graphs, allowing human intervention to prune incorrect paths and improve LLM reasoning accuracy by up to 24%.

DetailsMotivation: Chain-of-thought prompting in LLMs is opaque and difficult to verify, debug, and control in high-stakes settings, making human oversight necessary for reliable reasoning.

Method: Converts linear CoT text into interactive reasoning graphs, enabling users to visualize logical flow, identify flawed steps, and intervene by pruning incorrect paths or grafting new premises.

Result: Improves final-answer accuracy by up to 24 percentage points over non-interactive baselines on GSM8K and StrategyQA datasets, with significant gains in perceived usability and trust.

Conclusion: Vis-CoT provides a practical approach for more reliable and understandable reasoning by combining LLMs with targeted human oversight through interactive visualization and intervention.

Abstract: Large language models (LLMs) show strong reasoning via chain-of-thought (CoT) prompting, but the process is opaque, which makes verification, debugging, and control difficult in high-stakes settings. We present Vis-CoT, a human-in-the-loop framework that converts linear CoT text into an interactive reasoning graph. Users can visualize the logical flow, identify flawed steps, and intervene by pruning incorrect paths and grafting new, user-defined premises. This shifts interaction from passive observation to active collaboration, steering models toward more accurate and trustworthy conclusions. Across GSM8K and StrategyQA, Vis-CoT improves final-answer accuracy by up to 24 percentage points over non-interactive baselines. A user study also shows large gains in perceived usability and trust. Vis-CoT points to a practical path for more reliable, understandable, and collaborative reasoning by combining LLMs with targeted human oversight.

[91] On the Alignment of Large Language Models with Global Human Opinion

Yang Liu, Masahiro Kaneko, Chenhui Chu

Main category: cs.CL

TL;DR: This study investigates how LLMs align with human opinions across different countries, languages, and historical periods using World Values Survey data, finding significant alignment gaps and demonstrating that prompt language matching can effectively steer LLM opinions.

DetailsMotivation: Existing research on LLM opinion alignment focuses mainly on US demographics and lacks global country samples, historical period studies, and investigation of how prompt language influences opinion alignment. The study aims to address these gaps.

Method: Created an evaluation framework based on World Values Survey (WVS) to systematically assess LLM opinion alignment across different countries, languages, and historical periods worldwide.

Result: LLMs appropriately or over-align with only a few countries while under-aligning with most countries. Changing prompt language to match questionnaire language effectively steers LLMs to better align with corresponding country opinions. LLMs show better alignment with contemporary populations.

Conclusion: This is the first comprehensive investigation of opinion alignment in LLMs across global, language, and temporal dimensions, demonstrating that language-based steering is an effective method for improving cross-cultural opinion alignment.

Abstract: Today’s large language models (LLMs) are capable of supporting multilingual scenarios, allowing users to interact with LLMs in their native languages. When LLMs respond to subjective questions posed by users, they are expected to align with the views of specific demographic groups or historical periods, shaped by the language in which the user interacts with the model. Existing studies mainly focus on researching the opinions represented by LLMs among demographic groups in the United States or a few countries, lacking worldwide country samples and studies on human opinions in different historical periods, as well as lacking discussion on using language to steer LLMs. Moreover, they also overlook the potential influence of prompt language on the alignment of LLMs’ opinions. In this study, our goal is to fill these gaps. To this end, we create an evaluation framework based on the World Values Survey (WVS) to systematically assess the alignment of LLMs with human opinions across different countries, languages, and historical periods around the world. We find that LLMs appropriately or over-align the opinions with only a few countries while under-aligning the opinions with most countries. Furthermore, changing the language of the prompt to match the language used in the questionnaire can effectively steer LLMs to align with the opinions of the corresponding country more effectively than existing steering methods. At the same time, LLMs are more aligned with the opinions of the contemporary population. To our knowledge, our study is the first comprehensive investigation of the topic of opinion alignment in LLMs across global, language, and temporal dimensions. Our code and data are publicly available at https://github.com/nlply/global-opinion-alignment.

[92] Trusted Uncertainty in Large Language Models: A Unified Framework for Confidence Calibration and Risk-Controlled Refusal

Markus Oehri, Giulia Conti, Kaviraj Pather, Alexandre Rossi, Laia Serra, Adrian Parody, Rogvi Johannesen, Aviaja Petersen, Arben Krasniqi

Main category: cs.CL

TL;DR: UniCR is a unified framework that converts various uncertainty evidence into calibrated correctness probabilities and enforces user-specified error budgets through principled refusal mechanisms.

DetailsMotivation: Language models need to know when not to answer to avoid incorrect responses, requiring a systematic approach to uncertainty quantification and refusal decisions.

Method: Learns lightweight calibration head with temperature scaling and proper scoring, uses black-box features for API-only models, employs conformal risk control for distribution-free guarantees, and aligns confidence with semantic fidelity using atomic factuality scores.

Result: Consistent improvements in calibration metrics, lower area under risk-coverage curve, higher coverage at fixed risk compared to baseline methods across short-form QA, code generation, and retrieval-augmented long-form QA tasks.

Conclusion: UniCR provides a portable evidence-to-probability-to-decision framework that improves trustworthiness without fine-tuning base models and remains valid under distribution shift, with evidence contradiction, semantic dispersion, and tool inconsistency being key refusal drivers.

Abstract: Deployed language models must decide not only what to answer but also when not to answer. We present UniCR, a unified framework that turns heterogeneous uncertainty evidence including sequence likelihoods, self-consistency dispersion, retrieval compatibility, and tool or verifier feedback into a calibrated probability of correctness and then enforces a user-specified error budget via principled refusal. UniCR learns a lightweight calibration head with temperature scaling and proper scoring, supports API-only models through black-box features, and offers distribution-free guarantees using conformal risk control. For long-form generation, we align confidence with semantic fidelity by supervising on atomic factuality scores derived from retrieved evidence, reducing confident hallucinations while preserving coverage. Experiments on short-form QA, code generation with execution tests, and retrieval-augmented long-form QA show consistent improvements in calibration metrics, lower area under the risk-coverage curve, and higher coverage at fixed risk compared to entropy or logit thresholds, post-hoc calibrators, and end-to-end selective baselines. Analyses reveal that evidence contradiction, semantic dispersion, and tool inconsistency are the dominant drivers of abstention, yielding informative user-facing refusal messages. The result is a portable recipe of evidence fusion to calibrated probability to risk-controlled decision that improves trustworthiness without fine-tuning the base model and remains valid under distribution shift.

[93] Robust Knowledge Editing via Explicit Reasoning Chains for Distractor-Resilient Multi-Hop QA

Yuchen Wu, Liang Ding, Li Shen, Dacheng Tao

Main category: cs.CL

TL;DR: Reason-KE is a reasoning-chain-based knowledge editing framework that improves LLM’s ability to integrate new facts while filtering distractors in a single pass, achieving 90.2% multi-hop QA accuracy.

DetailsMotivation: Large language models struggle with timely integration of emerging facts due to static nature after training. Existing knowledge-editing techniques either rely on superficial cues or have complex pipelines that fail under noisy, multi-hop conditions.

Method: End-to-end reasoning-chain framework with four structured stages: fact acknowledgment, relevance determination, selective application, and final reasoning. Trained on MQuAKE-CF dataset with up to four irrelevant facts.

Result: Achieves 90.2% multi-hop QA accuracy on Qwen2.5-7B, suffers only 6.3% drop under heavy distraction and <1% when answers are leaked. Demonstrates resilience and efficiency.

Conclusion: Reason-KE establishes a new state-of-the-art for reliable LLM knowledge updates, providing robust performance in noisy, multi-hop conditions through structured reasoning chains.

Abstract: Large language models (LLMs) encode vast amounts of world knowledge but remain static once trained, making the timely integration of emerging facts prohibitively expensive via full retraining. Knowledge-editing techniques have thus emerged to inject or overwrite specific facts into LLMs, yet they either over-rely on superficial cues or incur complex, iterative pipelines that collapse under noisy, multi-hop conditions. We introduce Reason-KE, an end-to-end reasoning-chain-based editing framework that steers a pretrained LLM through four structured stages-fact acknowledgment, relevance determination, selective application, and final reasoning-to filter distractors in a single pass. Trained on MQuAKE-CF with up to four irrelevant facts, Reason-KE elevates Qwen2.5-7B’s multi-hop QA accuracy to 90.2% while suffering merely a 6.3% drop under heavy distraction and <1% when answers are leaked. Our quantitative analysis confirms Reason-KE’s resilience and efficiency, establishing a new state-of-the-art for reliable LLM knowledge updates.

[94] Do Retrieval Augmented Language Models Know When They Don’t Know?

Youchao Zhou, Heyan Huang, Yicheng Liu, Rui Dai, Xinglin Wang, Xingchen Zhang, Shumin Shi, Yang Deng

Main category: cs.CL

TL;DR: This paper investigates whether Retrieval Augmented Language Models (RALMs) can properly refuse to answer when they lack knowledge, finding they exhibit significant over-refusal behavior and examining how different training methods affect this issue.

DetailsMotivation: Existing LLMs generate hallucinations, and while RALMs and refusal post-training are used to mitigate this, there's limited evaluation of RALMs' refusal capabilities. The researchers want to understand if RALMs know when they don't know.

Method: The study examines RALM calibration across different knowledge states, investigates refusal post-training methods (Refusal-aware Instruction Tuning and In-Context Fine-tuning), and develops a simple refusal method for post-trained models.

Result: RALMs show significant over-refusal behavior. In-context fine-tuning mitigates over-refusal while R-tuning magnifies it, but refusal ability may conflict with answer quality. The researchers developed an effective refusal method to improve overall answer quality.

Conclusion: The study provides comprehensive understanding of factors influencing RALM systems, revealing over-refusal issues and developing methods to balance refusal capability with answer quality.

Abstract: Existing Large Language Models (LLMs) occasionally generate plausible yet factually incorrect responses, known as hallucinations. Researchers are primarily using two approaches to mitigate hallucinations, namely Retrieval Augmented Language Models (RALMs) and refusal post-training. However, current research predominantly emphasizes their individual effectiveness while overlooking the evaluation of the refusal capability of RALMs. In this study, we ask the fundamental question: Do RALMs know when they don’t know? Specifically, we ask three questions. First, are RALMs well-calibrated regarding different internal and external knowledge states? We examine the influence of various factors. Contrary to expectations, we find that LLMs exhibit significant \textbf{over-refusal} behavior. Then, how does refusal post-training affect the over-refusal issue? We investigate the Refusal-aware Instruction Tuning and In-Context Fine-tuning methods. Our results show that the over-refusal problem is mitigated by In-context fine-tuning. but magnified by R-tuning. However, we also find that the refusal ability may conflict with the quality of the answer. Finally, we develop a simple yet effective refusal method for refusal post-trained models to improve their overall answer quality in terms of refusal and correct answers. Our study provides a more comprehensive understanding of the influence of important factors on RALM systems.

[95] MeVe: A Modular System for Memory Verification and Effective Context Control in Language Models

Andreas Ottem

Main category: cs.CL

TL;DR: MeVe introduces a modular 5-phase RAG architecture for memory verification and smart context composition, achieving 57-75% context reduction while improving accuracy.

DetailsMotivation: Standard RAG systems suffer from irrelevant/redundant information due to simple top-k semantic search, degrading performance and efficiency.

Method: Five-phase modular design: initial retrieval, relevance verification, fallback retrieval, context prioritization, and token budgeting for fine-grained control.

Result: 57% context reduction on Wikipedia dataset and 75% reduction on HotpotQA dataset compared to standard RAG, with improved accuracy.

Conclusion: MeVe provides a scalable framework for reliable LLM applications through refined context distillation and better factual grounding.

Abstract: Retrieval-Augmented Generation (RAG) systems typically face constraints because of their inherent mechanism: a simple top-k semantic search [1]. The approach often leads to the incorporation of irrelevant or redundant information in the context, degrading performance and efficiency [10][11]. This paper presents MeVe, a novel modular architecture intended for Memory Verification and smart context composition. MeVe rethinks the RAG paradigm by proposing a five-phase modular design that distinctly breaks down the retrieval and context composition process into distinct, auditable, and independently tunable phases: initial retrieval, relevance verification, fallback retrieval, context prioritization, and token budgeting. This architecture enables fine-grained control of what knowledge is made available to an LLM, enabling task-dependent filtering and adaptation. We release a reference implementation of MeVe as a proof of concept and evaluate its performance on knowledge-heavy QA tasks over a subset of English Wikipedia [22]. Our results demonstrate that by actively verifying information before composition, MeVe significantly improves context efficiency, achieving a 57% reduction on the Wikipedia dataset and a 75% reduction on the more complex HotpotQA dataset compared to standard RAG implementations [25]. This work provides a framework for more scalable and reliable LLM applications. By refining and distilling contextual information, MeVe offers a path toward better grounding and more accurate factual support [16].

[96] Service, Solidarity, and Self-Help: A Comparative Topic Modeling Analysis of Community Unionism in the Boot and Shoe Union and Unite Community

Thomas Compton

Main category: cs.CL

TL;DR: Comparative analysis of community unionism in 1920s National Boot & Shoe Union vs 2010s-2020s Unite Community using NLP techniques, revealing divergent engagement models despite both addressing community themes.

DetailsMotivation: To examine how community unionism discourse differs across historical periods and organizational contexts, and to test whether modern NLP methods can effectively analyze historical labor archives.

Method: Used BERTopic for thematic modeling and cTF-IDF weighting alongside word frequency analysis to compare union discourses on key CU features like coalition-building and grassroots engagement.

Result: Significant differences found: Unite Community showed stronger alignment with outward-facing social justice themes, while B&S focused on internal administration and traditional servicing model. Modern NLP techniques proved effective for historical analysis.

Conclusion: Both unions engage with community themes but have fundamentally different engagement models, challenging assumptions about continuity and universality of community unionism across time and sectors.

Abstract: This paper presents a comparative analysis of community unionism (CU) in two distinct historical and organizational contexts: the National Boot and Shoe Union (B&S) in the 1920s and Unite Community in the 2010s–2020s. Using BERTopic for thematic modeling and cTF-IDF weighting, alongside word frequency analysis, the study examines the extent to which each union’s discourse aligns with key features of CU – such as coalition-building, grassroots engagement, and action beyond the workplace. The results reveal significant differences in thematic focus and discursive coherence. While Unite Community demonstrates stronger alignment with outward-facing, social justice-oriented themes, the B&S corpus emphasizes internal administration, industrial relations, and member services – reflecting a more traditional, servicing-oriented union model. The analysis also highlights methodological insights, demonstrating how modern NLP techniques can enhance the study of historical labor archives. Ultimately, the findings suggest that while both unions engage with community-related themes, their underlying models of engagement diverge significantly, challenging assumptions about the continuity and universality of community unionism across time and sector.

[97] CAT: Causal Attention Tuning For Injecting Fine-grained Causal Knowledge into Large Language Models

Kairong Han, Wenshuo Zhao, Ziyu Zhao, JunJian Ye, Lujia Pan, Kun Kuang

Main category: cs.CL

TL;DR: CAT (Causal Attention Tuning) is a novel method that injects fine-grained causal knowledge into LLM attention mechanisms to improve prediction robustness and reduce reliance on spurious correlations, especially in out-of-distribution scenarios.

DetailsMotivation: LLMs often capture spurious correlations rather than true causal relationships from large-scale training data, leading to suboptimal performance in out-of-distribution scenarios where these correlations break down.

Method: Proposes Causal Attention Tuning (CAT) with an automated pipeline that leverages human priors to generate token-level causal signals and introduces Re-Attention mechanism to guide training, helping models focus on causal structures while mitigating attention noise and biases.

Result: Experimental results on the proposed Spurious Token Game benchmark and multiple downstream tasks show that CAT effectively leverages causal knowledge for prediction and remains robust in OOD scenarios.

Conclusion: CAT successfully addresses LLMs’ tendency to rely on spurious correlations by injecting causal knowledge into attention mechanisms, demonstrating improved robustness and performance in out-of-distribution settings.

Abstract: Large Language Models (LLMs) have achieved remarkable success across various domains. However, a fundamental question remains: Can LLMs effectively utilize causal knowledge for prediction and generation? Through empirical studies, we find that LLMs trained directly on large-scale data often capture spurious correlations rather than true causal relationships, leading to suboptimal performance, especially in out-of-distribution (OOD) scenarios. To address this challenge, we propose Causal Attention Tuning (CAT), a novel approach that injects fine-grained causal knowledge into the attention mechanism. We propose an automated pipeline that leverages human priors to automatically generate token-level causal signals and introduce the Re-Attention mechanism to guide training, helping the model focus on causal structures while mitigating noise and biases in attention scores. Experimental results on our proposed Spurious Token Game (STG) benchmark and multiple downstream tasks demonstrate that our approach effectively leverages causal knowledge for prediction and remains robust in OOD scenarios. Implementation details can be found at https://github.com/Kairong-Han/CAT.

[98] In-N-Out: A Parameter-Level API Graph Dataset for Tool Agents

Seungkyu Lee, Nalim Kim, Yohan Jo

Main category: cs.CL

TL;DR: In-N-Out dataset converts API documentation into structured API graphs to help tool agents better handle complex multi-tool queries requiring compositional API calls, nearly doubling performance compared to using documentation alone.

DetailsMotivation: Tool agents struggle with complex tasks requiring proper API identification and sequencing. Current approaches using raw API documentation are insufficient for handling compositional API calls in multi-tool queries.

Method: Convert API documentation into structured API graphs capturing API dependencies, create In-N-Out expert-annotated dataset from real-world API benchmarks, and use this for tool retrieval and multi-tool query generation.

Result: Significant performance improvement on both tool retrieval and multi-tool query generation, nearly doubling LLM performance compared to using documentation alone. Fine-tuned models close 90% of the performance gap.

Conclusion: Explicit API graphs show promise for tool agents, and In-N-Out serves as a valuable resource for helping models learn API documentation comprehension and parameter relationships.

Abstract: Tool agents – LLM-based systems that interact with external APIs – offer a way to execute real-world tasks. However, as tasks become increasingly complex, these agents struggle to identify and call the correct APIs in the proper order. To tackle this problem, we investigate converting API documentation into a structured API graph that captures API dependencies and leveraging it for multi-tool queries that require compositional API calls. To support this, we introduce In-N-Out, the first expert-annotated dataset of API graphs built from two real-world API benchmarks and their documentation. Using In-N-Out significantly improves performance on both tool retrieval and multi-tool query generation, nearly doubling that of LLMs using documentation alone. Moreover, graphs generated by models fine-tuned on In-N-Out close 90% of this gap, showing that our dataset helps models learn to comprehend API documentation and parameter relationships. Our findings highlight the promise of using explicit API graphs for tool agents and the utility of In-N-Out as a valuable resource. We will release the dataset and code publicly.

[99] Enhancing Uncertainty Estimation in LLMs with Expectation of Aggregated Internal Belief

Zeguan Xiao, Diyang Dou, Boya Xiong, Yun Chen, Guanhua Chen

Main category: cs.CL

TL;DR: EAGLE is a novel self-evaluation calibration method that uses internal hidden states from multiple intermediate layers to produce more accurate confidence scores for LLMs, addressing overconfidence issues.

DetailsMotivation: LLMs often exhibit overconfidence and generate plausible but incorrect answers, especially after RLHF training, which poses challenges for reliable uncertainty estimation and safe deployment.

Method: Extracts internal beliefs from multiple intermediate layers during self-evaluation, aggregates these layer-wise beliefs, and calculates expectation over the confidence score distribution to produce refined confidence scores.

Result: Extensive experiments show EAGLE significantly improves calibration performance over existing baselines across diverse datasets and LLMs.

Conclusion: EAGLE effectively leverages internal model states to provide more faithful confidence estimation, enhancing reliability and safety of LLM deployments.

Abstract: Large Language Models (LLMs) have achieved remarkable success across a wide range of natural language tasks, but often exhibit overconfidence and generate plausible yet incorrect answers. This overconfidence, especially in models undergone Reinforcement Learning from Human Feedback (RLHF), poses significant challenges for reliable uncertainty estimation and safe deployment. In this paper, we propose EAGLE (Expectation of AGgregated internaL bEief), a novel self-evaluation-based calibration method that leverages the internal hidden states of LLMs to derive more accurate confidence scores. Instead of relying on the model’s final output, our approach extracts internal beliefs from multiple intermediate layers during self-evaluation. By aggregating these layer-wise beliefs and calculating the expectation over the resulting confidence score distribution, EAGLE produces a refined confidence score that more faithfully reflects the model’s internal certainty. Extensive experiments on diverse datasets and LLMs demonstrate that EAGLE significantly improves calibration performance over existing baselines. We also provide an in-depth analysis of EAGLE, including a layer-wise examination of uncertainty patterns, a study of the impact of self-evaluation prompts, and an analysis of the effect of self-evaluation score range.

[100] Testing the assumptions about the geometry of sentence embedding spaces: the cosine measure need not apply

Vivi Nastase, Paola Merlo

Main category: cs.CL

TL;DR: Sentence embedding geometry (cosine similarity) does not predict performance on linguistic tasks - linguistic information is encoded in weighted dimension combinations rather than geometric proximity.

DetailsMotivation: To investigate whether the geometric proximity in sentence embedding space correlates with performance on linguistic tasks, similar to how word embeddings work.

Method: Computed sentence embeddings using three approaches: averaged token embeddings, [CLS] token embeddings, and random token embeddings. Analyzed correlation between embedding distances and task performance.

Result: Cosine similarity captures only shallow commonalities/differences and is not predictive of performance on specific tasks.

Conclusion: Linguistic information is encoded in weighted combinations of dimensions rather than reflected in the geometry of the sentence embedding space.

Abstract: Transformer models learn to encode and decode an input text, and produce contextual token embeddings as a side-effect. The mapping from language into the embedding space maps words expressing similar concepts onto points that are close in the space. In practice, the reverse implication is also assumed: words corresponding to close points in this space are similar or related, those that are further are not. Does closeness in the embedding space extend to shared properties for sentence embeddings? We present an investigation of sentence embeddings and show that the geometry of their embedding space is not predictive of their relative performances on a variety of tasks. We compute sentence embeddings in three ways: as averaged token embeddings, as the embedding of the special [CLS] token, and as the embedding of a random token from the sentence. We explore whether there is a correlation between the distance between sentence embedding variations and their performance on linguistic tasks, and whether despite their distances, they do encode the same information in the same manner. The results show that the cosine similarity – which treats dimensions shallowly – captures (shallow) commonalities or differences between sentence embeddings, which are not predictive of their performance on specific tasks. Linguistic information is rather encoded in weighted combinations of different dimensions, which are not reflected in the geometry of the sentence embedding space.

[101] Benchmarking the Detection of LLMs-Generated Modern Chinese Poetry

Shanshan Wang, Junchao Wu, Fengying Ye, Jingming Yao, Lidia S. Chao, Derek F. Wong

Main category: cs.CL

TL;DR: A new benchmark for detecting AI-generated modern Chinese poetry shows current detectors fail to reliably identify LLM-generated poems, with style being the most difficult feature to detect.

DetailsMotivation: The proliferation of AI-generated modern Chinese poetry is disrupting the poetry ecosystem, but existing detection methods haven't addressed the unique characteristics of modern Chinese poetry, making it difficult to distinguish human-written from AI-generated poems.

Method: Constructed a high-quality dataset with 800 human-written poems by professional poets and 41,600 poems generated by four mainstream LLMs, then systematically evaluated six detectors on this dataset.

Result: Experimental results show current detectors cannot reliably detect modern Chinese poems generated by LLMs, with intrinsic qualities (especially style) being the most difficult features to detect.

Conclusion: The proposed benchmark is effective and necessary, laying a foundation for future detection of AI-generated poetry, as current detection tools are inadequate for modern Chinese poetry.

Abstract: The rapid development of advanced large language models (LLMs) has made AI-generated text indistinguishable from human-written text. Previous work on detecting AI-generated text has made effective progress, but has not involved modern Chinese poetry. Due to the distinctive characteristics of modern Chinese poetry, it is difficult to identify whether a poem originated from humans or AI. The proliferation of AI-generated modern Chinese poetry has significantly disrupted the poetry ecosystem. Based on the urgency of identifying AI-generated poetry in the real Chinese world, this paper proposes a novel benchmark for detecting LLMs-generated modern Chinese poetry. We first construct a high-quality dataset, which includes both 800 poems written by six professional poets and 41,600 poems generated by four mainstream LLMs. Subsequently, we conduct systematic performance assessments of six detectors on this dataset. Experimental results demonstrate that current detectors cannot be used as reliable tools to detect modern Chinese poems generated by LLMs. The most difficult poetic features to detect are intrinsic qualities, especially style. The detection results verify the effectiveness and necessity of our proposed benchmark. Our work lays a foundation for future detection of AI-generated poetry.

[102] TransGAT: Transformer-Based Graph Neural Networks for Multi-Dimensional Automated Essay Scoring

Hind Aljuaid, Areej Alhothali, Ohoud Al-Zamzami, Hussein Assalahi

Main category: cs.CL

TL;DR: TransGAT: Transformer + GNN model for automated essay scoring that outperforms baselines with 0.854 QWK score on analytic scoring dimensions

DetailsMotivation: Manual essay scoring is labor-intensive and inconsistent. Current AES approaches use static word embeddings that fail to capture contextual meaning and mostly rely on holistic scoring, overlooking specific writing aspects.

Method: TransGAT integrates fine-tuned Transformer models (BERT, RoBERTa, DeBERTaV3) with Graph Attention Networks (GAT) for two-stream predictions. First stream generates essay-level predictions, second stream applies GAT to Transformer token embeddings with syntactic dependency edges. Predictions are fused for final analytic scores.

Result: Experiments on ELLIPSE dataset show TransGAT outperforms baseline models, achieving average Quadratic Weighted Kappa (QWK) of 0.854 across all analytic scoring dimensions.

Conclusion: TransGAT demonstrates the potential of combining contextual Transformers with relational GNNs to advance automated essay scoring systems through improved analytic scoring capabilities.

Abstract: Essay writing is a critical component of student assessment, yet manual scoring is labor-intensive and inconsistent. Automated Essay Scoring (AES) offers a promising alternative, but current approaches face limitations. Recent studies have incorporated Graph Neural Networks (GNNs) into AES using static word embeddings that fail to capture contextual meaning, especially for polysemous words. Additionally, many methods rely on holistic scoring, overlooking specific writing aspects such as grammar, vocabulary, and cohesion. To address these challenges, this study proposes TransGAT, a novel approach that integrates fine-tuned Transformer models with GNNs for analytic scoring. TransGAT combines the contextual understanding of Transformers with the relational modeling strength of Graph Attention Networks (GAT). It performs two-stream predictions by pairing each fine-tuned Transformer (BERT, RoBERTa, and DeBERTaV3) with a separate GAT. In each pair, the first stream generates essay-level predictions, while the second applies GAT to Transformer token embeddings, with edges constructed from syntactic dependencies. The model then fuses predictions from both streams to produce the final analytic score. Experiments on the ELLIPSE dataset show that TransGAT outperforms baseline models, achieving an average Quadratic Weighted Kappa (QWK) of 0.854 across all analytic scoring dimensions. These findings highlight the potential of TransGAT to advance AES systems.

[103] Parallel Needleman-Wunsch on CUDA to measure word similarity based on phonetic transcriptions

Dominic Plein

Main category: cs.CL

TL;DR: A method for calculating word similarity using phonetic transcriptions with Needleman-Wunsch algorithm, implemented in Rust with CPU/GPU parallelization for large datasets.

DetailsMotivation: To efficiently analyze phonetic similarity between words in large datasets and identify groups of phonetically similar words through graph analysis.

Method: Uses Needleman-Wunsch algorithm on phonetic transcriptions, implemented in Rust with parallel CPU/GPU processing using CUDA and cudarc library. Constructs fully-connected graph with similarity-weighted edges and applies clustering algorithms.

Result: Demonstrated feasibility and effectiveness in analyzing phonetic language structure. GPU implementation achieved significant performance improvements for large datasets.

Conclusion: The proposed method is effective for phonetic similarity analysis and can be easily expanded to other languages beyond the tested implementation.

Abstract: We present a method to calculate the similarity between words based on their phonetic transcription (their pronunciation) using the Needleman-Wunsch algorithm. We implement this algorithm in Rust and parallelize it on both CPU and GPU to handle large datasets efficiently. The GPU implementation leverages CUDA and the cudarc Rust library to achieve significant performance improvements. We validate our approach by constructing a fully-connected graph where nodes represent words and edges have weights according to the similarity between the words. This graph is then analyzed using clustering algorithms to identify groups of phonetically similar words. Our results demonstrate the feasibility and effectiveness of the proposed method in analyzing the phonetic structure of languages. It might be easily expanded to other languages.

[104] Bridging Thoughts and Words: Graph-Based Intent-Semantic Joint Learning for Fake News Detection

Zhengjia Wang, Qiang Sheng, Danding Wang, Beizhe Hu, Juan Cao

Main category: cs.CL

TL;DR: Proposes InSide, a graph-based intent-semantic joint modeling approach for fake news detection that combines semantic patterns with news intent understanding to overcome limitations of surface-level detection.

DetailsMotivation: Existing fake news detectors focus on semantic clues like writing patterns and emotional words, but these surface patterns can shift rapidly in dynamic environments, leading to limited performance. The paper aims to incorporate news intent to deeply understand deception thoughts rather than just surface word patterns.

Method: Graph-based Intent-Semantic Joint Modeling (InSide) that models deception clues from both semantic and intent signals via graph-based joint learning. It reformulates signals into heterogeneous graph structures, uses entity guidance for long-range context interaction, and employs coarse-to-fine intent modeling. Includes dynamic pathway-based graph alignment strategy for better semantics-intent alignment.

Result: Extensive experiments on four benchmark datasets demonstrate the superiority of the proposed InSide compared to state-of-the-art methods.

Conclusion: Incorporating news intent alongside semantic clues through graph-based joint modeling provides a more robust approach to fake news detection that can better handle evolving news landscapes and avoid surface pattern limitations.

Abstract: Fake news detection is an important and challenging task for defending online information integrity. Existing state-of-the-art approaches typically extract news semantic clues, such as writing patterns that include emotional words, stylistic features, etc. However, detectors tuned solely to such semantic clues can easily fall into surface detection patterns, which can shift rapidly in dynamic environments, leading to limited performance in the evolving news landscape. To address this issue, this paper investigates a novel perspective by incorporating news intent into fake news detection, bridging intents and semantics together. The core insight is that by considering news intents, one can deeply understand the inherent thoughts behind news deception, rather than the surface patterns within words alone. To achieve this goal, we propose Graph-based Intent-Semantic Joint Modeling (InSide) for fake news detection, which models deception clues from both semantic and intent signals via graph-based joint learning. Specifically, InSide reformulates news semantic and intent signals into heterogeneous graph structures, enabling long-range context interaction through entity guidance and capturing both holistic and implementation-level intent via coarse-to-fine intent modeling. To achieve better alignment between semantics and intents, we further develop a dynamic pathway-based graph alignment strategy for effective message passing and aggregation across these signals by establishing a common space. Extensive experiments on four benchmark datasets demonstrate the superiority of the proposed InSide compared to state-of-the-art methods.

[105] chDzDT: Word-level morphology-aware language model for Algerian social media text

Abdelkrime Aries

Main category: cs.CL

TL;DR: chDzDT is a character-level pre-trained language model designed specifically for the Algerian dialect to handle its complex morphology, code-switching, and multiple scripts through isolated word training rather than token sequences.

DetailsMotivation: The Algerian dialect is under-represented in NLP with few dedicated models, and its complex morphology, frequent code-switching, multiple scripts, and lexical influences from other languages make conventional tokenization approaches ineffective.

Method: Developed a character-level PLM trained on isolated words from diverse sources including YouTube comments, French/English/Berber Wikipedia, and Tatoeba project, covering multiple scripts and linguistic varieties.

Result: The model robustly encodes morphological patterns without depending on token boundaries or standardized orthography, demonstrating effectiveness for morphologically rich, low-resource dialects.

Conclusion: Character-level modeling shows strong potential for handling complex dialects like Algerian, providing a foundation for more inclusive and adaptable NLP systems that can better serve under-represented languages.

Abstract: Pre-trained language models (PLMs) have substantially advanced natural language processing by providing context-sensitive text representations. However, the Algerian dialect remains under-represented, with few dedicated models available. Processing this dialect is challenging due to its complex morphology, frequent code-switching, multiple scripts, and strong lexical influences from other languages. These characteristics complicate tokenization and reduce the effectiveness of conventional word- or subword-level approaches. To address this gap, we introduce chDzDT, a character-level pre-trained language model tailored for Algerian morphology. Unlike conventional PLMs that rely on token sequences, chDzDT is trained on isolated words. This design allows the model to encode morphological patterns robustly, without depending on token boundaries or standardized orthography. The training corpus draws from diverse sources, including YouTube comments, French, English, and Berber Wikipedia, as well as the Tatoeba project. It covers multiple scripts and linguistic varieties, resulting in a substantial pre-training workload. Our contributions are threefold: (i) a detailed morphological analysis of Algerian dialect using YouTube comments; (ii) the construction of a multilingual Algerian lexicon dataset; and (iii) the development and extensive evaluation of a character-level PLM as a morphology-focused encoder for downstream tasks. The proposed approach demonstrates the potential of character-level modeling for morphologically rich, low-resource dialects and lays a foundation for more inclusive and adaptable NLP systems.

[106] Flaw or Artifact? Rethinking Prompt Sensitivity in Evaluating LLMs

Andong Hua, Kenan Tang, Chenhe Gu, Jindong Gu, Eric Wong, Yao Qin

Main category: cs.CL

TL;DR: Prompt sensitivity in LLMs may be more due to evaluation artifacts than inherent model weaknesses, as LLM-as-a-Judge evaluation shows reduced variance and better correlation across prompts.

DetailsMotivation: To determine whether the widely reported high prompt sensitivity is truly an inherent weakness of LLMs or largely an artifact of evaluation processes.

Method: Systematically evaluated 7 LLMs across 6 benchmarks using 12 diverse prompt templates, comparing heuristic evaluation methods with LLM-as-a-Judge evaluations.

Result: Found that heuristic methods (log-likelihood scoring, rigid answer matching) overlook semantically correct responses, while LLM-as-a-Judge shows substantial reduction in performance variance and higher correlation in model rankings.

Conclusion: Modern LLMs are more robust to prompt templates than previously believed, and prompt sensitivity may be more an artifact of evaluation than a flaw in the models.

Abstract: Prompt sensitivity, referring to the phenomenon where paraphrasing (i.e., repeating something written or spoken using different words) leads to significant changes in large language model (LLM) performance, has been widely accepted as a core limitation of LLMs. In this work, we revisit this issue and ask: Is the widely reported high prompt sensitivity truly an inherent weakness of LLMs, or is it largely an artifact of evaluation processes? To answer this question, we systematically evaluate 7 LLMs (e.g., GPT and Gemini family) across 6 benchmarks, including both multiple-choice and open-ended tasks on 12 diverse prompt templates. We find that much of the prompt sensitivity stems from heuristic evaluation methods, including log-likelihood scoring and rigid answer matching, which often overlook semantically correct responses expressed through alternative phrasings, such as synonyms or paraphrases. When we adopt LLM-as-a-Judge evaluations, we observe a substantial reduction in performance variance and a consistently higher correlation in model rankings across prompts. Our findings suggest that modern LLMs are more robust to prompt templates than previously believed, and that prompt sensitivity may be more an artifact of evaluation than a flaw in the models.

[107] Mic Drop or Data Flop? Evaluating the Fitness for Purpose of AI Voice Interviewers for Data Collection within Quantitative & Qualitative Research Contexts

Shreyas Tirumala, Nishant Jain, Danny D. Leybzon, Trent D. Buskirk

Main category: cs.CL

TL;DR: AI interviewers using LLMs show promise for both quantitative and qualitative data collection, outperforming traditional IVR systems but still facing challenges in transcription accuracy, emotion detection, and follow-up quality.

DetailsMotivation: To evaluate the fitness of AI interviewing systems based on LLMs for quantitative and qualitative research data collection compared to traditional Interactive Voice Response systems.

Method: Position paper reviewing emerging evidence and evaluating AI interviewers across two dimensions: input/output performance (speech recognition, recording, emotion handling) and verbal reasoning (probing, clarifying, branching logic).

Result: AI interviewers already exceed IVR capabilities for both quantitative and qualitative data collection, but face challenges including real-time transcription error rates, limited emotion detection abilities, and uneven follow-up quality.

Conclusion: The utility and adoption of current AI interviewer technology may be context-dependent for qualitative data collection efforts, indicating that while promising, these systems still have limitations that need addressing.

Abstract: Transformer-based Large Language Models (LLMs) have paved the way for “AI interviewers” that can administer voice-based surveys with respondents in real-time. This position paper reviews emerging evidence to understand when such AI interviewing systems are fit for purpose for collecting data within quantitative and qualitative research contexts. We evaluate the capabilities of AI interviewers as well as current Interactive Voice Response (IVR) systems across two dimensions: input/output performance (i.e., speech recognition, answer recording, emotion handling) and verbal reasoning (i.e., ability to probe, clarify, and handle branching logic). Field studies suggest that AI interviewers already exceed IVR capabilities for both quantitative and qualitative data collection, but real-time transcription error rates, limited emotion detection abilities, and uneven follow-up quality indicate that the utility, use and adoption of current AI interviewer technology may be context-dependent for qualitative data collection efforts.

[108] Extracting OPQRST in Electronic Health Records using Large Language Models with Reasoning

Zhimeng Luo, Abhibha Gupta, Adam Frisch, Daqing He

Main category: cs.CL

TL;DR: Novel LLM-based approach for extracting OPQRST assessment from EHRs by reframing from sequence labeling to text generation with reasoning steps, using semantic similarity metrics for evaluation.

DetailsMotivation: Traditional ML approaches fail to efficiently capture critical patient information from complex, unstructured EHR data, making it difficult for clinicians to utilize these tools effectively in patient care.

Method: Reframe OPQRST extraction from sequence labeling to text generation using LLMs, enabling models to provide reasoning steps mimicking physician cognition. Propose modified NER metrics with BERT Score integration for semantic similarity evaluation.

Result: Significant advancement in AI healthcare applications, offering scalable solution that improves accuracy and usability of EHR information extraction.

Conclusion: The approach enhances interpretability, adapts to limited labeled data in healthcare, and aids clinicians in making more informed decisions to improve patient care outcomes.

Abstract: The extraction of critical patient information from Electronic Health Records (EHRs) poses significant challenges due to the complexity and unstructured nature of the data. Traditional machine learning approaches often fail to capture pertinent details efficiently, making it difficult for clinicians to utilize these tools effectively in patient care. This paper introduces a novel approach to extracting the OPQRST assessment from EHRs by leveraging the capabilities of Large Language Models (LLMs). We propose to reframe the task from sequence labeling to text generation, enabling the models to provide reasoning steps that mimic a physician’s cognitive processes. This approach enhances interpretability and adapts to the limited availability of labeled data in healthcare settings. Furthermore, we address the challenge of evaluating the accuracy of machine-generated text in clinical contexts by proposing a modification to traditional Named Entity Recognition (NER) metrics. This includes the integration of semantic similarity measures, such as the BERT Score, to assess the alignment between generated text and the clinical intent of the original records. Our contributions demonstrate a significant advancement in the use of AI in healthcare, offering a scalable solution that improves the accuracy and usability of information extraction from EHRs, thereby aiding clinicians in making more informed decisions and enhancing patient care outcomes.

[109] Weakly Supervised Medical Entity Extraction and Linking for Chief Complaints

Zhimeng Luo, Zhendong Wang, Rui Meng, Diyang Xue, Adam Frisch, Daqing He

Main category: cs.CL

TL;DR: Weakly supervised method for extracting and linking medical entities from chief complaints without human annotation, using split-and-match algorithm and BERT model.

DetailsMotivation: Chief complaint records have varied entering methods and medical notations, making standardization difficult across medical institutions for record keeping and text mining.

Method: Adopt split-and-match algorithm to produce weak annotations on 1.2M chief complaint records, then train BERT-based model to locate entity mentions and link to pre-defined ontology.

Result: Superior performance over previous methods without any human annotation.

Conclusion: The proposed weakly supervised method effectively extracts and links entities from chief complaints, addressing standardization challenges in medical text mining.

Abstract: A Chief complaint (CC) is the reason for the medical visit as stated in the patient’s own words. It helps medical professionals to quickly understand a patient’s situation, and also serves as a short summary for medical text mining. However, chief complaint records often take a variety of entering methods, resulting in a wide variation of medical notations, which makes it difficult to standardize across different medical institutions for record keeping or text mining. In this study, we propose a weakly supervised method to automatically extract and link entities in chief complaints in the absence of human annotation. We first adopt a split-and-match algorithm to produce weak annotations, including entity mention spans and class labels, on 1.2 million real-world de-identified and IRB approved chief complaint records. Then we train a BERT-based model with generated weak labels to locate entity mentions in chief complaint text and link them to a pre-defined ontology. We conducted extensive experiments, and the results showed that our Weakly Supervised Entity Extraction and Linking (\ours) method produced superior performance over previous methods without any human annotation.

[110] DRAssist: Dispute Resolution Assistance using Large Language Models

Sachin Pawar, Manoj Apte, Girish K. Palshikar, Basit Ali, Nitin Ramrakhiyani

Main category: cs.CL

TL;DR: DRAssist system uses LLMs to assist human judges in resolving disputes by structuring case information and providing multi-level resolution outputs across insurance and domain name disputes.

DetailsMotivation: Disputes occur across many domains and are typically resolved by human judges. This paper explores using LLMs to assist judges by automating the analysis and structuring of dispute information.

Method: DRAssist system identifies key structural elements (facts, disagreement aspects, arguments) from unstructured dispute descriptions, creates structured summaries, and uses multiple prompting strategies with LLMs to provide resolution outputs at three levels: identifying stronger party, evaluating specific demands, and assessing argument strength.

Result: The system was evaluated on automobile insurance and domain name disputes, comparing LLM performance against baselines using appropriate metrics (though specific results are not detailed in the abstract).

Conclusion: LLMs show promise as assistants for human judges in dispute resolution by providing structured analysis and multi-level evaluation capabilities across different dispute domains.

Abstract: Disputes between two parties occur in almost all domains such as taxation, insurance, banking, healthcare, etc. The disputes are generally resolved in a specific forum (e.g., consumer court) where facts are presented, points of disagreement are discussed, arguments as well as specific demands of the parties are heard, and finally a human judge resolves the dispute by often favouring one of the two parties. In this paper, we explore the use of large language models (LLMs) as assistants for the human judge to resolve such disputes, as part of our DRAssist system. We focus on disputes from two specific domains – automobile insurance and domain name disputes. DRAssist identifies certain key structural elements (e.g., facts, aspects or disagreement, arguments) of the disputes and summarizes the unstructured dispute descriptions to produce a structured summary for each dispute. We then explore multiple prompting strategies with multiple LLMs for their ability to assist in resolving the disputes in these domains. In DRAssist, these LLMs are prompted to produce the resolution output at three different levels – (i) identifying an overall stronger party in a dispute, (ii) decide whether each specific demand of each contesting party can be accepted or not, (iii) evaluate whether each argument by each contesting party is strong or weak. We evaluate the performance of LLMs on all these tasks by comparing them with relevant baselines using suitable evaluation metrics.

[111] StructCoh: Structured Contrastive Learning for Context-Aware Text Semantic Matching

Chao Xue, Ziyuan Gao

Main category: cs.CL

TL;DR: StructCoh is a graph-enhanced contrastive learning framework that combines structural reasoning with representation optimization for text semantic matching, achieving state-of-the-art results on legal and academic document matching tasks.

DetailsMotivation: Pre-trained language models excel at token-level interactions but often overlook hierarchical structural patterns and struggle with subtle semantic discrimination in text matching tasks.

Method: Uses dual-graph encoder with dependency parsing and topic modeling, graph isomorphism networks for structural feature propagation, and hierarchical contrastive learning with node-level and graph-aware objectives.

Result: Achieves 86.7% F1-score (+6.2% absolute gain) on legal statute matching and demonstrates significant improvements on three legal document matching benchmarks and academic plagiarism detection datasets.

Conclusion: StructCoh effectively combines structural reasoning with representation optimization, outperforming state-of-the-art methods by better capturing hierarchical structural patterns and semantic distinctions.

Abstract: Text semantic matching requires nuanced understanding of both structural relationships and fine-grained semantic distinctions. While pre-trained language models excel at capturing token-level interactions, they often overlook hierarchical structural patterns and struggle with subtle semantic discrimination. In this paper, we proposed StructCoh, a graph-enhanced contrastive learning framework that synergistically combines structural reasoning with representation space optimization. Our approach features two key innovations: (1) A dual-graph encoder constructs semantic graphs via dependency parsing and topic modeling, then employs graph isomorphism networks to propagate structural features across syntactic dependencies and cross-document concept nodes. (2) A hierarchical contrastive objective enforces consistency at multiple granularities: node-level contrastive regularization preserves core semantic units, while graph-aware contrastive learning aligns inter-document structural semantics through both explicit and implicit negative sampling strategies. Experiments on three legal document matching benchmarks and academic plagiarism detection datasets demonstrate significant improvements over state-of-the-art methods. Notably, StructCoh achieves 86.7% F1-score (+6.2% absolute gain) on legal statute matching by effectively identifying argument structure similarities.

[112] DeepSeek performs better than other Large Language Models in Dental Cases

Hexian Zhang, Xinyu Yan, Yanqi Yang, Lijian Jin, Ping Yang, Junwen Wang

Main category: cs.CL

TL;DR: DeepSeek V3 outperformed GPT-4o, Gemini 2.0 Flash, and Copilot in analyzing longitudinal dental cases, showing superior faithfulness and expert ratings without compromising readability.

DetailsMotivation: To evaluate LLMs' ability to interpret longitudinal patient narratives in healthcare, using dentistry as a test case with rich structured clinical data.

Method: Evaluated 4 LLMs on 34 standardized longitudinal periodontal cases (258 QA pairs) using automated metrics and blinded evaluations by licensed dentists.

Result: DeepSeek achieved highest faithfulness (median 0.528 vs 0.367-0.457) and expert ratings (median 4.5/5 vs 4.0/5) while maintaining readability.

Conclusion: DeepSeek is the leading LLM for dental case analysis, suitable for integration as an adjunct tool in medical education and research, with potential as a domain-specific agent.

Abstract: Large language models (LLMs) hold transformative potential in healthcare, yet their capacity to interpret longitudinal patient narratives remains inadequately explored. Dentistry, with its rich repository of structured clinical data, presents a unique opportunity to rigorously assess LLMs' reasoning abilities. While several commercial LLMs already exist, DeepSeek, a model that gained significant attention earlier this year, has also joined the competition. This study evaluated four state-of-the-art LLMs (GPT-4o, Gemini 2.0 Flash, Copilot, and DeepSeek V3) on their ability to analyze longitudinal dental case vignettes through open-ended clinical tasks. Using 34 standardized longitudinal periodontal cases (comprising 258 question-answer pairs), we assessed model performance via automated metrics and blinded evaluations by licensed dentists. DeepSeek emerged as the top performer, demonstrating superior faithfulness (median score = 0.528 vs. 0.367-0.457) and higher expert ratings (median = 4.5/5 vs. 4.0/5), without significantly compromising readability. Our study positions DeepSeek as the leading LLM for case analysis, endorses its integration as an adjunct tool in both medical education and research, and highlights its potential as a domain-specific agent.

[113] Attributes as Textual Genes: Leveraging LLMs as Genetic Algorithm Simulators for Conditional Synthetic Data Generation

Guangzeng Han, Weisi Liu, Xiaolei Huang

Main category: cs.CL

TL;DR: Genetic Prompt combines genetic algorithms with LLMs to enhance synthetic data generation quality and diversity by treating text attributes as genes and using LLMs for crossover/mutation operations.

DetailsMotivation: LLMs are good at generating synthetic data but struggle with ensuring quality and diversity, requiring better methods to produce synthetic distributions closer to real-world data.

Method: A framework that treats semantic text attributes as gene sequences, uses LLMs to simulate genetic operations (crossover and mutation), and integrates active learning for optimal parent selection to expand search space.

Result: Significantly outperforms state-of-the-art baselines, shows robust performance across various model sizes, and fusing synthetic data with original training boosts downstream performance, especially in class-imbalanced scenarios.

Conclusion: Genetic Prompt is an effective method for producing high-quality synthetic data that works well across various NLP applications and model scales.

Abstract: Large Language Models (LLMs) excel at generating synthetic data, but ensuring its quality and diversity remains challenging. We propose Genetic Prompt, a novel framework that combines genetic algorithms with LLMs to augment synthetic data generation. Our approach treats semantic text attributes as gene sequences and leverages the LLM to simulate crossover and mutation operations. This genetic process enhances data quality and diversity by creating novel attribute combinations, yielding synthetic distributions closer to real-world data. To optimize parent selection, we also integrate an active learning scheme that expands the offspring search space. Our experiments on multiple NLP tasks reveal several key findings: Genetic Prompt not only significantly outperforms state-of-the-art baselines but also shows robust performance across various generator model sizes and scales. Moreover, we demonstrate that fusing our synthetic data with the original training set significantly boosts downstream model performance, particularly for class-imbalanced scenarios. Our findings validate that Genetic Prompt is an effective method for producing high-quality synthetic data for a wide range of NLP applications.

[114] How Instruction-Tuning Imparts Length Control: A Cross-Lingual Mechanistic Analysis

Elisabetta Rocchetti, Alfio Ferrara

Main category: cs.CL

TL;DR: Instruction-tuning improves LLMs’ length control in text generation by specializing deeper layers, with attention heads showing stronger contributions in English and MLPs compensating in Italian.

DetailsMotivation: Large Language Models struggle with precise length constraints in text generation, and this study aims to understand how foundation models differ from instruction-tuned models in handling length-controlled generation across English and Italian.

Method: The researchers analyzed performance and internal component contributions using Cumulative Weighted Attribution, a metric derived from Direct Logit Attribution, comparing foundation models and instruction-tuned models.

Result: Instruction-tuning substantially improves length control by specializing components in deeper model layers. Attention heads in later layers show increasingly positive contributions in English, while in Italian, final-layer MLPs exhibit stronger positive role as a compensatory mechanism.

Conclusion: Instruction-tuning reconfigures later layers for better task adherence, with component-level strategies adapting to linguistic context, demonstrating that model architecture responds differently to length constraints across languages.

Abstract: Adhering to explicit length constraints, such as generating text with a precise word count, remains a significant challenge for Large Language Models (LLMs). This study aims at investigating the differences between foundation models and their instruction-tuned counterparts, on length-controlled text generation in English and Italian. We analyze both performance and internal component contributions using Cumulative Weighted Attribution, a metric derived from Direct Logit Attribution. Our findings reveal that instruction-tuning substantially improves length control, primarily by specializing components in deeper model layers. Specifically, attention heads in later layers of IT models show increasingly positive contributions, particularly in English. In Italian, while attention contributions are more attenuated, final-layer MLPs exhibit a stronger positive role, suggesting a compensatory mechanism. These results indicate that instruction-tuning reconfigures later layers for task adherence, with component-level strategies potentially adapting to linguistic context.

[115] Better by Comparison: Retrieval-Augmented Contrastive Reasoning for Automatic Prompt Optimization

Juhyeon Lee, Wonduk Seo, Hyunjin An, Seunghyun Lee, Yi Bu

Main category: cs.CL

TL;DR: CRPO is a novel prompt optimization framework that uses contrastive reasoning with retrieved reference prompts to improve LLM performance by learning from both good and bad examples.

DetailsMotivation: Most prior work focuses on direct prompt refinement or model fine-tuning, overlooking LLMs' inherent reasoning capability to learn from contrasting examples of prompt quality.

Method: Retrieves top k reference prompts from HelpSteer2 dataset and uses two paradigms: tiered contrastive reasoning (comparing high/medium/low quality prompts) and multi-metric contrastive reasoning (analyzing best prompts along each evaluation dimension).

Result: Experimental results on HelpSteer2 benchmark demonstrate that CRPO significantly outperforms baselines.

Conclusion: Contrastive, retrieval-augmented reasoning shows promise for advancing automatic prompt optimization by enabling more robust and interpretable optimization.

Abstract: Automatic prompt optimization has recently emerged as a strategy for improving the quality of prompts used in Large Language Models (LLMs), with the goal of generating more accurate and useful responses. However, most prior work focuses on direct prompt refinement or model fine-tuning, overlooking the potential of leveraging LLMs’ inherent reasoning capability to learn from contrasting examples. In this paper, we present Contrastive Reasoning Prompt Optimization (CRPO), a novel framework that formulates prompt optimization as a retrieval augmented reasoning process. Our approach retrieves top k reference prompts from the HelpSteer2 dataset, an open-source collection annotated for helpfulness, correctness, coherence, complexity, and verbosity, and constructs two complementary optimization paradigms: (1) tiered contrastive reasoning, where the LLM compares high, medium, and low quality prompts to refine its own generation through reflective reasoning, and (2) multi-metric contrastive reasoning, where the LLM analyzes the best prompts along each evaluation dimension and integrates their strengths into an optimized prompt. By explicitly contrasting high and low quality exemplars, CRPO enables the model to deduce why certain prompts succeed while others fail, thereby achieving more robust and interpretable optimization. Experimental results on the HelpSteer2 benchmark demonstrate that CRPO significantly outperforms baselines. Our findings highlight the promise of contrastive, retrieval-augmented reasoning for advancing automatic prompt optimization.

[116] JudgeAgent: Dynamically Evaluate LLMs with Agent-as-Interviewer

Zhichao Shi, Xuhui Jiang, Chengjin Xu, Cangli Yao, Zhenxin Huang, Shengjie Ma, Yinghan Shen, Yuanzhuo Wang

Main category: cs.CL

TL;DR: JudgeAgent is a dynamic evaluation framework that uses interviewer-style interaction to overcome limitations of current static LLM evaluation methods, providing more accurate capability assessment through adaptive difficulty adjustment and knowledge-driven testing.

DetailsMotivation: Current LLM evaluation methods based on predefined question sets have limitations including limited interaction, insufficient difficulty control, and difficulty verifying result validity, making it hard to precisely determine model capabilities.

Method: JudgeAgent employs a knowledge-target adaptive dynamic evaluation framework with benchmark grading, interactive extension, and evaluation feedback. It uses knowledge-driven data synthesis and target-adaptive difficulty adjustment for extended testing.

Result: Extensive experiments demonstrate the effectiveness of JudgeAgent and its dynamic evaluation paradigm in providing accurate and effective evaluation results.

Conclusion: The proposed interviewer-style evaluation paradigm with JudgeAgent addresses key limitations of current static evaluation methods, enabling more precise determination of LLM knowledge and capability boundaries through adaptive, interactive testing.

Abstract: Evaluating the capabilities of large language models (LLMs) is an essential step to ensure the successful application of LLMs across various domains. The current evaluation of LLMs is based on a paradigm that involves querying them with predefined question sets and assessing their outputs. This paradigm offers controllable processes and simplicity, but faces challenges such as limited interaction with targets, insufficient difficulty control, and difficulties in verifying the validity of evaluation results, making it hard to precisely determine the knowledge and capability boundaries of target models. To address these challenges, we propose JudgeAgent, a knowledge-target adaptive dynamic evaluation framework based on a new interviewer-style evaluation paradigm. JudgeAgent employs a comprehensive evaluation approach consisting of benchmark grading, interactive extension, and evaluation feedback. It utilizes knowledge-driven data synthesis and target-adaptive difficulty adjustment methods to conduct extended testing, providing accurate and effective evaluation results. We also introduce a novel insight into validating evaluation methods, demonstrating the effectiveness of JudgeAgent and its dynamic evaluation paradigm through extensive experiments.

[117] CMRAG: Co-modality-based document retrieval and visual question answering

Wang Chen, Guanqiang Qi, Weikang Li, Yang Li

Main category: cs.CL

TL;DR: CMRAG: A co-modality RAG framework that combines text and image information for better document question answering, outperforming vision-only approaches.

DetailsMotivation: Existing RAG methods struggle with multimodal documents - text-only methods miss visual content, while vision-only methods ignore text semantics, leading to suboptimal performance.

Method: Structured document parsing to get co-modality representations, separate text/image channel retrieval, cross-modal result aggregation, and VLM prompting for final response generation.

Result: Significantly outperforms pure-vision-based RAG in visual document question answering tasks.

Conclusion: Integrating co-modality information in a unified RAG framework effectively improves complex document VQA system performance.

Abstract: Retrieval-Augmented Generation (RAG) has become a core paradigm in document question answering tasks. However, existing methods have limitations when dealing with multimodal documents: one category of methods relies on layout analysis and text extraction, which can only utilize explicit text information and struggle to capture images or unstructured content; the other category treats document segmentation as visual input and directly passes it to visual language models (VLMs) for processing, yet it ignores the semantic advantages of text, leading to suboptimal generation results. This paper proposes co-modality-based RAG (CMRAG), which can simultaneously leverage text and images for efficient retrieval and generation. Specifically, we first perform structured parsing on documents to obtain co-modality representations of text segments and image regions. Subsequently, in response to user queries, we retrieve candidate evidence from text and image channels, respectively, and aggregate the results at the cross-modal retrieval level. Finally, we prompt the VLM to generate the final response based on the co-modality retrieval results. Experiments demonstrate that our method significantly outperforms pure-vision-based RAG in visual document question answering tasks. The findings of this paper show that integrating co-modality information into the RAG framework in a unified manner is an effective approach to improving the performance of complex document visual question-answering (VQA) systems.

[118] AMBEDKAR-A Multi-level Bias Elimination through a Decoding Approach with Knowledge Augmentation for Robust Constitutional Alignment of Language Models

Snehasis Mukhopadhyay, Aryan Kasat, Shivam Dubey, Rahul Karthikeyan, Dhruv Sood, Vinija Jain, Aman Chadha, Amitava Das

Main category: cs.CL

TL;DR: AMBEDKAR framework uses speculative decoding with constitutional guidance to reduce caste and religious biases in LLMs for Indian context, achieving up to 26.41% bias reduction without model retraining.

DetailsMotivation: LLMs reflect societal biases from training data, particularly caste and religion biases in Indian context. Existing mitigation strategies are Western-centric and fail to address local nuances.

Method: Constitution-Aware Decoding Layer applied at inference time using speculative decoding. Small Language Model acts as biased generator, while constitutionally guided LLM verifies and enforces bias-robust trajectories without parameter updates.

Result: Achieves absolute bias reduction up to 26.41% compared to baseline models, addressing casteist and communal biases effectively.

Conclusion: The framework successfully reduces harmful biases in LLM outputs for Indian context using constitutional guidance and speculative decoding, providing a cost-effective solution without model retraining.

Abstract: Large Language Models (LLMs) can inadvertently reflect societal biases present in their training data, leading to harmful or prejudiced outputs. In the Indian context, our empirical evaluations across a suite of models reveal that biases around caste and religion are particularly salient. Yet, most existing mitigation strategies are Western-centric and fail to address these local nuances. We propose AMBEDKAR, a framework inspired by the egalitarian vision of Dr B. R. Ambedkar, architect of the Indian Constitution, to guide LLM outputs toward fairness, neutrality, and inclusion in line with Articles 14 to 17. Our approach introduces a Constitution-Aware Decoding Layer, guided by the AI Constitution of India and applied only at inference time, without any parameter updates to the base model. We incorporate a speculative decoding algorithm that proactively reduces casteist and communal bias during generation. This mitigation layer operates directly within the decoding process, avoiding changes to model internals and lowering the computational and infrastructural costs associated with retraining. We reinterpret speculative decoding not merely as an efficiency tool but as a mechanism for fairness. In this framework, a Small Language Model (SLM) acts as a potentially biased generator, while a constitutionally guided Large Language Model (LLM) serves as the verifier. Rather than accelerating generation, the LLM enforces bias-robust trajectories in the SLM outputs. This inversion of roles gives rise to a fairness-by-speculation paradigm. Our approach yields an absolute reduction of bias up to 26.41 percent compared to baseline. Our source code, datasets, and results are available at https://anonymous.4open.science/r/AMBEDKAR-983B/

[119] Meta-Pretraining for Zero-Shot Cross-Lingual Named Entity Recognition in Low-Resource Philippine Languages

David Demitri Africa, Suchir Salhan, Yuval Weiss, Paula Buttery, Richard Diehl Martinez

Main category: cs.CL

TL;DR: Small decoder LMs pretrained with MAML show improved zero-shot NER performance in low-resource languages, with 2-6pp F1 gains and faster convergence.

DetailsMotivation: Address NER in low-resource languages where large multilingual LMs are infeasible due to memory/latency constraints, by enabling small LMs to adapt quickly and transfer zero-shot to unseen languages.

Method: Replace part of autoregressive objective with first-order model-agnostic meta-learning (MAML) during pretraining, tested across four model sizes (11M-570M parameters) on Tagalog and Cebuano languages.

Result: MAML improved zero-shot micro-F1 by 2-6pp with head-only tuning and 1-3pp with full tuning, while reducing convergence time by up to 8%. Largest gains for single-token person entities co-occurring with Tagalog case particles.

Conclusion: Meta-learning enables small LMs to effectively transfer to unseen low-resource languages, with surface anchors like case particles playing crucial role in zero-shot NER performance.

Abstract: Named-entity recognition (NER) in low-resource languages is usually tackled by finetuning very large multilingual LMs, an option that is often infeasible in memory- or latency-constrained settings. We ask whether small decoder LMs can be pretrained so that they adapt quickly and transfer zero-shot to languages unseen during pretraining. To this end we replace part of the autoregressive objective with first-order model-agnostic meta-learning (MAML). Tagalog and Cebuano are typologically similar yet structurally different in their actor/non-actor voice systems, and hence serve as a challenging test-bed. Across four model sizes (11 M - 570 M) MAML lifts zero-shot micro-F1 by 2-6 pp under head-only tuning and 1-3 pp after full tuning, while cutting convergence time by up to 8%. Gains are largest for single-token person entities that co-occur with Tagalog case particles si/ni, highlighting the importance of surface anchors.

[120] Avoidance Decoding for Diverse Multi-Branch Story Generation

Kyeongman Park, Nakyeong Yang, Kyomin Jung

Main category: cs.CL

TL;DR: Avoidance Decoding is a novel LLM decoding strategy that penalizes similarity to previous outputs to generate more diverse stories, achieving 2.6x higher diversity and 30% less repetition.

DetailsMotivation: LLMs often produce repetitive and monotonous outputs, especially in story generation tasks, due to limited creative diversity when given the same input prompt.

Method: A decoding strategy that modifies token logits by penalizing similarity to previously generated outputs using two adaptive similarity measures: Concept-level Similarity Penalty (early stages) and Narrative-level Similarity Penalty (later stages).

Result: Achieves up to 2.6 times higher output diversity, reduces repetition by an average of 30% compared to strong baselines, and effectively mitigates text degeneration while activating a broader range of neurons.

Conclusion: The method successfully enhances creative diversity in LLM outputs by leveraging the model’s intrinsic creativity through adaptive similarity penalties at different generation stages.

Abstract: Large Language Models (LLMs) often generate repetitive and monotonous outputs, especially in tasks like story generation, due to limited creative diversity when given the same input prompt. To address this challenge, we propose a novel decoding strategy, Avoidance Decoding, that modifies token logits by penalizing similarity to previously generated outputs, thereby encouraging more diverse multi-branch stories. This penalty adaptively balances two similarity measures: (1) Concept-level Similarity Penalty, which is prioritized in early stages to diversify initial story concepts, and (2) Narrative-level Similarity Penalty, which is increasingly emphasized later to ensure natural yet diverse plot development. Notably, our method achieves up to 2.6 times higher output diversity and reduces repetition by an average of 30% compared to strong baselines, while effectively mitigating text degeneration. Furthermore, we reveal that our method activates a broader range of neurons, demonstrating that it leverages the model’s intrinsic creativity.

[121] FActBench: A Benchmark for Fine-grained Automatic Evaluation of LLM-Generated Text in the Medical Domain

Anum Afzal, Juraj Vladika, Florian Matthes

Main category: cs.CL

TL;DR: FActBench is a comprehensive fact-checking benchmark for medical domain LLMs, using CoT prompting and NLI techniques with unanimous voting showing best correlation with expert evaluation.

DetailsMotivation: LLMs struggle with specialized domains like medicine where factuality is critical, and there's a need for reliable fact-checking tools and data sources for hallucination mitigation.

Method: Created FActBench covering 4 generation tasks and 6 state-of-the-art LLMs for medical domain, using Chain-of-Thought Prompting and Natural Language Inference techniques with unanimous voting approach.

Result: Experiments showed that fact-checking scores obtained through unanimous voting of both CoT and NLI techniques correlate best with domain expert evaluation.

Conclusion: The unanimous voting approach combining CoT and NLI provides the most reliable fact-checking method for medical domain LLMs, addressing critical factuality issues in specialized domains.

Abstract: Large Language Models tend to struggle when dealing with specialized domains. While all aspects of evaluation hold importance, factuality is the most critical one. Similarly, reliable fact-checking tools and data sources are essential for hallucination mitigation. We address these issues by providing a comprehensive Fact-checking Benchmark FActBench covering four generation tasks and six state-of-the-art Large Language Models (LLMs) for the Medical domain. We use two state-of-the-art Fact-checking techniques: Chain-of-Thought (CoT) Prompting and Natural Language Inference (NLI). Our experiments show that the fact-checking scores acquired through the Unanimous Voting of both techniques correlate best with Domain Expert Evaluation.

[122] Towards Fundamental Language Models: Does Linguistic Competence Scale with Model Size?

Jaime Collado-Montañez, L. Alfonso Ureña-López, Arturo Montejo-Ráez

Main category: cs.CL

TL;DR: Smaller language models with external fact retrieval tools outperform monolithic models by separating linguistic competence from factual memorization, leading to more efficient and sustainable NLP solutions.

DetailsMotivation: Address limitations of large language models including hallucinations, biases, privacy concerns, and high computational costs caused by combining linguistic competence with factual memorization in single models.

Method: Proposed Fundamental Language Model (FLM) paradigm using smaller linguistically competent models that offload factual retrieval to external tools. Evaluated models from 135M to 32B parameters across linguistic competence, external factual knowledge, and internal factual knowledge.

Result: Found that while both linguistic competence and factual knowledge improve with scale, internal factual knowledge grows significantly faster, showing model size is more tied to memorization than core language ability.

Conclusion: Supports modular approach where compact linguistically proficient models serve as foundation for tool-augmented systems, offering path to more efficient, interpretable, and sustainable NLP solutions.

Abstract: Large Language Models offer impressive language capabilities but suffer from well-known limitations, including hallucinations, biases, privacy concerns, and high computational costs. These issues are largely driven by the combination of linguistic competence and factual memorization within a single monolithic model. This paper introduces and empirically supports the Fundamental Language Model (FLM) paradigm, which advocates for smaller, linguistically competent models that offload factual retrieval to external tools. We evaluate models ranging from 135M to 32B parameters across three dimensions: linguistic competence, external factual knowledge, and internal factual knowledge. Our findings reveal that while both linguistic competence and factual knowledge improve with scale, internal factual knowledge grows significantly faster, suggesting that model size is more closely tied to memorization than to core language ability. These results support a modular approach to language modeling, where compact, linguistically proficient models serve as the foundation for tool-augmented systems. The FLM paradigm offers a path toward more efficient, interpretable, and sustainable NLP solutions.

[123] LLMs and their Limited Theory of Mind: Evaluating Mental State Annotations in Situated Dialogue

Katharine Kowalyshyn, Matthias Scheutz

Main category: cs.CL

TL;DR: LLM-based framework for detecting team mental model discrepancies through dialogue annotation and comparison with human annotations

DetailsMotivation: To leverage LLMs for identifying blind spots and discrepancies in team members' shared understanding during collaborative tasks

Method: Two-step framework: 1) LLM generates annotations of shared mental models from team dialogues, 2) Secondary LLM compares LLM vs human annotations against gold-standard labels to detect divergences

Result: LLMs show coherence on straightforward annotation tasks but systematically fail in scenarios requiring spatial reasoning or prosodic cue disambiguation

Conclusion: While LLMs can be used for team mental model analysis, they have limitations in complex reasoning scenarios requiring spatial understanding and nuanced interpretation

Abstract: What if large language models could not only infer human mindsets but also expose every blind spot in team dialogue such as discrepancies in the team members’ joint understanding? We present a novel, two-step framework that leverages large language models (LLMs) both as human-style annotators of team dialogues to track the team’s shared mental models (SMMs) and as automated discrepancy detectors among individuals’ mental states. In the first step, an LLM generates annotations by identifying SMM elements within task-oriented dialogues from the Cooperative Remote Search Task (CReST) corpus. Then, a secondary LLM compares these LLM-derived annotations and human annotations against gold-standard labels to detect and characterize divergences. We define an SMM coherence evaluation framework for this use case and apply it to six CReST dialogues, ultimately producing: (1) a dataset of human and LLM annotations; (2) a reproducible evaluation framework for SMM coherence; and (3) an empirical assessment of LLM-based discrepancy detection. Our results reveal that, although LLMs exhibit apparent coherence on straightforward natural-language annotation tasks, they systematically err in scenarios requiring spatial reasoning or disambiguation of prosodic cues.

[124] DCPO: Dynamic Clipping Policy Optimization

Shihui Yang, Chengfeng Dou, Peidong Guo, Kai Lu, Qiang Ju, Fei Deng, Rihui Xin

Main category: cs.CL

TL;DR: DCPO introduces dynamic clipping and smooth advantage standardization to overcome zero gradient issues in RLVR, achieving SOTA performance on multiple benchmarks with improved training efficiency.

DetailsMotivation: Existing RLVR approaches like GRPO suffer from zero gradients due to fixed clipping bounds and reward standardization, leading to ineffective gradient updates and underutilization of generated responses.

Method: Dynamic Clipping Policy Optimization (DCPO) with adaptive clipping bounds based on token-specific prior probabilities for better exploration, and smooth advantage standardization across cumulative training steps for improved response utilization.

Result: Achieved state-of-the-art performance on four benchmarks with four different models. On AIME24 with Qwen2.5-Math-7B: 46.7/38.8 (greedy/sampling) vs DAPO (36.7/31.6) and GRPO (36.7/32.1). On AIME25 with Qwen2.5-14B: 23.3/19.0 vs GRPO (13.3/10.5) and DAPO (20.0/15.3). 28% improvement in nonzero advantage over GRPO, doubled training efficiency over DAPO, and order-of-magnitude reduction in token clipping ratio.

Conclusion: DCPO effectively leverages generated data more efficiently for reinforcement learning in LLMs by addressing gradient issues through dynamic clipping and advantage standardization, demonstrating superior performance and training efficiency.

Abstract: Reinforcement Learning from Verifiable Rewards (RLVR) has emerged as a promising framework for enhancing the reasoning capabilities of large language models. However, existing approaches such as GRPO often suffer from zero gradients. This problem arises primarily due to fixed clipping bounds for token-level probability ratios and the standardization of identical rewards, which can lead to ineffective gradient updates and underutilization of generated responses. In this work, we propose Dynamic Clipping Policy Optimization (DCPO), which introduces a dynamic clipping strategy that adaptively adjusts the clipping bounds based on token-specific prior probabilities to enhance token-level exploration, and a smooth advantage standardization technique that standardizes rewards across cumulative training steps to improve the response-level effective utilization of generated responses. DCPO achieved state-of-the-art performance on four benchmarks based on four different models. In particular, DCPO achieved an Avg@1 of 46.7 under greedy decoding and an Avg@32 of 38.8 under 32 times sampling on the AIME24 benchmark, surpassing both DAPO (36.7/31.6) and GRPO (36.7/32.1) on the Qwen2.5-Math-7B model. On the AIME25 benchmark based on Qwen2.5-14B, DCPO achieves a performance of (23.3/19.0), surpassing GRPO (13.3/10.5) and DAPO (20.0/15.3). Furthermore, DCPO achieved an average 28% improvement in the nonzero advantage over GRPO in four models, doubled the training efficiency over DAPO, and significantly reduced the token clipping ratio by an order of magnitude compared to both GRPO and DAPO, while achieving superior performance. These results highlight DCPO’s effectiveness in leveraging generated data more efficiently for reinforcement learning in large language models.

[125] Implicit Reasoning in Large Language Models: A Comprehensive Survey

Jindong Li, Yali Fu, Li Fan, Jiahong Liu, Yao Shu, Chengwei Qin, Menglin Yang, Irwin King, Rex Ying

Main category: cs.CL

TL;DR: Survey paper on implicit reasoning in LLMs, focusing on internal computational mechanisms rather than explicit chain-of-thought prompting, with taxonomy of three execution paradigms.

DetailsMotivation: Existing surveys discuss latent representations but lack dedicated mechanism-level examination of how reasoning unfolds internally within LLMs without emitting intermediate textual steps.

Method: Introduces taxonomy centered on execution paradigms: latent optimization, signal-guided control, and layer-recurrent execution. Reviews structural, behavioral and representation-based evidence supporting implicit reasoning.

Result: Organizes existing methods into three computational paradigms and provides structured overview of evaluation metrics and benchmarks for assessing implicit reasoning effectiveness.

Conclusion: Fills the gap in understanding internal reasoning mechanisms in LLMs, providing a systematic framework for analyzing implicit reasoning approaches and their computational strategies.

Abstract: Large Language Models (LLMs) have demonstrated strong generalization across a wide range of tasks. Reasoning with LLMs is central to solving multi-step problems and complex decision-making. To support efficient reasoning, recent studies have shifted attention from explicit chain-of-thought prompting toward implicit reasoning, where reasoning occurs silently via latent structures without emitting intermediate textual steps. Implicit reasoning brings advantages such as lower generation cost, faster inference, and better alignment with internal computation. Although prior surveys have discussed latent representations in the context of reasoning, a dedicated and mechanism-level examination of how reasoning unfolds internally within LLMs remains absent. This survey fills that gap by introducing a taxonomy centered on execution paradigms, shifting the focus from representational forms to computational strategies. We organize existing methods into three execution paradigms based on \textbf{\textit{how and where internal computation unfolds}}: latent optimization, signal-guided control, and layer-recurrent execution. We also review structural, behavioral and representation-based evidence that supports the presence of implicit reasoning in LLMs. We further provide a structured overview of the evaluation metrics and benchmarks used in existing works to assess the effectiveness and reliability of implicit reasoning.We maintain a continuously updated project at: https://github.com/digailab/awesome-llm-implicit-reasoning.

[126] Towards Temporal Knowledge-Base Creation for Fine-Grained Opinion Analysis with Language Models

Gaurav Negi, Atul Kr. Ojha, Omnia Zayed, Paul Buitelaar

Main category: cs.CL

TL;DR: A scalable method using LLMs as automated annotators to build temporal opinion knowledge bases with structured opinion extraction without manual prompt engineering.

DetailsMotivation: Existing methodologies underutilize time-series opinion analysis due to lack of temporally grounded fine-grained annotations, creating a gap for downstream applications like forecasting and trend analysis.

Method: Integrates established opinion mining formulations into declarative LLM annotation pipeline, defines three data models based on sentiment/opinion mining literature, uses two separate LLMs for annotations with inter-annotator agreement evaluation.

Result: Rigorous quantitative evaluation using human-annotated test samples shows successful construction of time-aligned structured opinion knowledge base.

Conclusion: The resulting knowledge base enables applications in Retrieval-Augmented Generation, temporal question answering, and timeline summarization through automated LLM-based annotation.

Abstract: We propose a scalable method for constructing a temporal opinion knowledge base with large language models (LLMs) as automated annotators. Despite the demonstrated utility of time-series opinion analysis of text for downstream applications such as forecasting and trend analysis, existing methodologies underexploit this potential due to the absence of temporally grounded fine-grained annotations. Our approach addresses this gap by integrating well-established opinion mining formulations into a declarative LLM annotation pipeline, enabling structured opinion extraction without manual prompt engineering. We define three data models grounded in sentiment and opinion mining literature, serving as schemas for structured representation. We perform rigorous quantitative evaluation of our pipeline using human-annotated test samples. We carry out the final annotations using two separate LLMs, and inter-annotator agreement is computed label-wise across the fine-grained opinion dimensions, analogous to human annotation protocols. The resulting knowledge base encapsulates time-aligned, structured opinions and is compatible with applications in Retrieval-Augmented Generation (RAG), temporal question answering, and timeline summarisation.

[127] An Ensemble Classification Approach in A Multi-Layered Large Language Model Framework for Disease Prediction

Ali Hamdi, Malak Mohamed, Rokaia Emad, Khaled Shaban

Main category: cs.CL

TL;DR: This paper presents a novel approach for Arabic disease classification using LLM-based preprocessing (summarization, refinement, NER) with fine-tuned Arabic transformer models and ensemble learning, achieving 80.56% accuracy on social telehealth data.

DetailsMotivation: Social telehealth platforms generate massive Arabic medical text data that can be leveraged for disease classification, but existing methods need improvement in handling Arabic medical text complexity and achieving higher accuracy.

Method: Three Arabic text preprocessing methods using LLMs (summarization, refinement, NER) followed by fine-tuning Arabic transformer models (CAMeLBERT, AraBERT, AsafayaBERT) and majority voting ensemble of original and preprocessed text representations.

Result: Achieved 80.56% classification accuracy, demonstrating effectiveness of combining various text representations and model predictions for Arabic medical text understanding.

Conclusion: The integration of LLM-based preprocessing with fine-tuned Arabic transformers and ensemble learning provides an effective framework for disease classification in Arabic social telehealth data, representing the first work of its kind in this domain.

Abstract: Social telehealth has made remarkable progress in healthcare by allowing patients to post symptoms and participate in medical consultations remotely. Users frequently post symptoms on social media and online health platforms, creating a huge repository of medical data that can be leveraged for disease classification. Large language models (LLMs) such as LLAMA3 and GPT-3.5, along with transformer-based models like BERT, have demonstrated strong capabilities in processing complex medical text. In this study, we evaluate three Arabic medical text preprocessing methods such as summarization, refinement, and Named Entity Recognition (NER) before applying fine-tuned Arabic transformer models (CAMeLBERT, AraBERT, and AsafayaBERT). To enhance robustness, we adopt a majority voting ensemble that combines predictions from original and preprocessed text representations. This approach achieved the best classification accuracy of 80.56%, thus showing its effectiveness in leveraging various text representations and model predictions to improve the understanding of medical texts. To the best of our knowledge, this is the first work that integrates LLM-based preprocessing with fine-tuned Arabic transformer models and ensemble learning for disease classification in Arabic social telehealth data.

[128] EmoPerso: Enhancing Personality Detection with Self-Supervised Emotion-Aware Modelling

Lingzhi Shen, Xiaohao Cai, Yunfei Long, Imran Razzak, Guanming Chen, Shoaib Jameel

Main category: cs.CL

TL;DR: EmoPerso is a self-supervised framework that improves personality detection from text by integrating emotion-aware modeling, synthetic data augmentation, and multi-task learning with cross-attention mechanisms.

DetailsMotivation: Existing personality detection methods rely heavily on large annotated datasets and treat emotion and personality as independent variables, overlooking their interactions, making it challenging to obtain high-quality personality labels.

Method: Proposes EmoPerso framework that uses generative mechanisms for synthetic data augmentation, extracts pseudo-labeled emotion features, employs multi-task learning with personality prediction, and uses cross-attention modules to capture fine-grained interactions between personality traits and emotional representations.

Result: Extensive experiments on two benchmark datasets demonstrate that EmoPerso surpasses state-of-the-art models in personality detection performance.

Conclusion: The proposed EmoPerso framework effectively addresses the limitations of existing methods by leveraging emotion-aware modeling and self-supervised learning, achieving superior performance in personality detection from text.

Abstract: Personality detection from text is commonly performed by analysing users' social media posts. However, existing methods heavily rely on large-scale annotated datasets, making it challenging to obtain high-quality personality labels. Moreover, most studies treat emotion and personality as independent variables, overlooking their interactions. In this paper, we propose a novel self-supervised framework, EmoPerso, which improves personality detection through emotion-aware modelling. EmoPerso first leverages generative mechanisms for synthetic data augmentation and rich representation learning. It then extracts pseudo-labeled emotion features and jointly optimizes them with personality prediction via multi-task learning. A cross-attention module is employed to capture fine-grained interactions between personality traits and the inferred emotional representations. To further refine relational reasoning, EmoPerso adopts a self-taught strategy to enhance the model’s reasoning capabilities iteratively. Extensive experiments on two benchmark datasets demonstrate that EmoPerso surpasses state-of-the-art models. The source code is available at https://github.com/slz0925/EmoPerso.

[129] Do LLMs Adhere to Label Definitions? Examining Their Receptivity to External Label Definitions

Seyedali Mohammadi, Bhaskara Hanuma Vedula, Hemank Lamba, Edward Raff, Ponnurangam Kumaraguru, Francis Ferraro, Manas Gaur

Main category: cs.CL

TL;DR: LLMs don’t consistently integrate external label definitions, often defaulting to internal parametric knowledge instead, with domain-specific tasks benefiting more from explicit definitions than general tasks.

DetailsMotivation: To determine whether LLMs genuinely incorporate external definitions or primarily rely on their parametric knowledge when solving tasks.

Method: Conducted controlled experiments across multiple explanation benchmark datasets (general and domain-specific) with various label definition conditions including expert-curated, LLM-generated, perturbed, and swapped definitions.

Result: Explicit label definitions can enhance accuracy and explainability, but their integration is neither guaranteed nor consistent. Models often default to internal representations, especially in general tasks, while domain-specific tasks benefit more from explicit definitions.

Conclusion: LLMs’ processing of external knowledge alongside pre-existing capabilities requires deeper understanding, as they don’t reliably integrate external definitions and frequently rely on internalized representations.

Abstract: Do LLMs genuinely incorporate external definitions, or do they primarily rely on their parametric knowledge? To address these questions, we conduct controlled experiments across multiple explanation benchmark datasets (general and domain-specific) and label definition conditions, including expert-curated, LLM-generated, perturbed, and swapped definitions. Our results reveal that while explicit label definitions can enhance accuracy and explainability, their integration into an LLM’s task-solving processes is neither guaranteed nor consistent, suggesting reliance on internalized representations in many cases. Models often default to their internal representations, particularly in general tasks, whereas domain-specific tasks benefit more from explicit definitions. These findings underscore the need for a deeper understanding of how LLMs process external knowledge alongside their pre-existing capabilities.

[130] SpecEval: Evaluating Model Adherence to Behavior Specifications

Ahmed Ahmed, Kevin Klyman, Yi Zeng, Sanmi Koyejo, Percy Liang

Main category: cs.CL

TL;DR: Automated framework audits foundation models against provider specifications, finding systematic inconsistencies and compliance gaps up to 20% across major AI companies.

DetailsMotivation: To systematically verify if foundation models actually adhere to the behavioral guidelines and safety constraints that their developers publicly pledge to follow, as there has been no comprehensive audit of this adherence despite detailed published specifications.

Method: Developed an automated framework that parses behavioral statements from provider specifications, generates targeted prompts to test adherence, and uses the providers’ own models as judges to evaluate consistency between specifications, model outputs, and model judgments.

Result: Applied to 16 models from 6 developers across 100+ behavioral statements, revealing systematic inconsistencies including compliance gaps of up to 20% across different providers.

Conclusion: The framework establishes a necessary baseline for model consistency and demonstrates that current foundation models show significant gaps in adhering to their own developers’ behavioral specifications when judged by the developers’ own evaluation models.

Abstract: Companies that develop foundation models publish behavioral guidelines they pledge their models will follow, but it remains unclear if models actually do so. While providers such as OpenAI, Anthropic, and Google have published detailed specifications describing both desired safety constraints and qualitative traits for their models, there has been no systematic audit of adherence to these guidelines. We introduce an automated framework that audits models against their providers specifications by parsing behavioral statements, generating targeted prompts, and using models to judge adherence. Our central focus is on three way consistency between a provider specification, its model outputs, and its own models as judges; an extension of prior two way generator validator consistency. This establishes a necessary baseline: at minimum, a foundation model should consistently satisfy the developer behavioral specifications when judged by the developer evaluator models. We apply our framework to 16 models from six developers across more than 100 behavioral statements, finding systematic inconsistencies including compliance gaps of up to 20 percent across providers.

[131] GRAM-R$^2$: Self-Training Generative Foundation Reward Models for Reward Reasoning

Chenglong Wang, Yongyu Mu, Hang Zhou, Yifu Huo, Ziming Zhu, Jiali Zeng, Murun Yang, Bei Li, Tong Xiao, Xiaoyang Hao, Chunliang Zhang, Fandong Meng, Jingbo Zhu

Main category: cs.CL

TL;DR: GRAM-R² is a generative reward model that produces both preference labels and reward rationales through self-training on unlabeled data, outperforming discriminative and generative baselines across multiple tasks.

DetailsMotivation: Current reward models heavily rely on large-scale labeled preference data, and existing pre-training approaches fail to instill explicit reasoning capabilities into reward models.

Method: Proposed self-training approach leveraging unlabeled data to elicit reward reasoning, developing GRAM-R² - a generative reward model that produces preference labels and accompanying rationales.

Result: GRAM-R² consistently delivers strong performance in response ranking, task adaptation, and reinforcement learning from human feedback, outperforming several strong discriminative and generative baselines.

Conclusion: GRAM-R² serves as an effective foundation model for reward reasoning that can be applied to wide range of tasks with minimal fine-tuning, supporting downstream applications like response ranking and task-specific reward tuning.

Abstract: Significant progress in reward modeling over recent years has been driven by a paradigm shift from task-specific designs towards generalist reward models. Despite this trend, developing effective reward models remains a fundamental challenge: the heavy reliance on large-scale labeled preference data. Pre-training on abundant unlabeled data offers a promising direction, but existing approaches fall short of instilling explicit reasoning into reward models. To bridge this gap, we propose a self-training approach that leverages unlabeled data to elicit reward reasoning in reward models. Based on this approach, we develop GRAM-R$^2$, a generative reward model trained to produce not only preference labels but also accompanying reward rationales. GRAM-R$^2$ can serve as a foundation model for reward reasoning and can be applied to a wide range of tasks with minimal or no additional fine-tuning. It can support downstream applications such as response ranking and task-specific reward tuning. Experiments on response ranking, task adaptation, and reinforcement learning from human feedback demonstrate that GRAM-R$^2$ consistently delivers strong performance, outperforming several strong discriminative and generative baselines.

[132] MoSEs: Uncertainty-Aware AI-Generated Text Detection via Mixture of Stylistics Experts with Conditional Thresholds

Junxi Wu, Jinpeng Wang, Zheng Liu, Bin Chen, Dongjian Hu, Hao Wu, Shu-Tao Xiu

Main category: cs.CL

TL;DR: MoSEs framework improves AI-generated text detection by 11.34% on average through stylistic modeling and dynamic threshold estimation, with 39.15% improvement in low-resource scenarios.

DetailsMotivation: Address public concerns about LLM misuse by building trustworthy detection systems, overcoming limitations of existing methods that neglect stylistic modeling and rely on static thresholds.

Method: Mixture of Stylistic Experts (MoSEs) framework with three components: Stylistics Reference Repository (SRR) for reference data activation, Stylistics-Aware Router (SAR), and Conditional Threshold Estimator (CTE) that jointly models linguistic statistical properties and semantic features for dynamic threshold determination.

Result: Achieves 11.34% average improvement in detection performance compared to baselines, with 39.15% improvement in low-resource cases.

Conclusion: MoSEs framework effectively addresses stylistic modeling limitations in AI-generated text detection and demonstrates significant performance gains, particularly in resource-constrained environments.

Abstract: The rapid advancement of large language models has intensified public concerns about the potential misuse. Therefore, it is important to build trustworthy AI-generated text detection systems. Existing methods neglect stylistic modeling and mostly rely on static thresholds, which greatly limits the detection performance. In this paper, we propose the Mixture of Stylistic Experts (MoSEs) framework that enables stylistics-aware uncertainty quantification through conditional threshold estimation. MoSEs contain three core components, namely, the Stylistics Reference Repository (SRR), the Stylistics-Aware Router (SAR), and the Conditional Threshold Estimator (CTE). For input text, SRR can activate the appropriate reference data in SRR and provide them to CTE. Subsequently, CTE jointly models the linguistic statistical properties and semantic features to dynamically determine the optimal threshold. With a discrimination score, MoSEs yields prediction labels with the corresponding confidence level. Our framework achieves an average improvement 11.34% in detection performance compared to baselines. More inspiringly, MoSEs shows a more evident improvement 39.15% in the low-resource case. Our code is available at https://github.com/creator-xi/MoSEs.

[133] L3Cube-IndicHeadline-ID: A Dataset for Headline Identification and Semantic Evaluation in Low-Resource Indian Languages

Nishant Tanksale, Tanmay Kokate, Darshan Gohad, Sarvadnyaa Barate, Raviraj Joshi

Main category: cs.CL

TL;DR: L3Cube-IndicHeadline-ID dataset for evaluating semantic understanding in 10 low-resource Indic languages with 20K news articles per language and four headline variants to test fine-grained similarity.

DetailsMotivation: Address the lack of high-quality benchmarks for semantic evaluation in low-resource Indic languages, particularly for sentence transformers which are underexplored in these settings.

Method: Created curated dataset with news articles paired with four headline variants (original, semantically similar, lexically similar, unrelated) and benchmarked multilingual and language-specific sentence transformers using cosine similarity for headline identification task.

Result: Multilingual models consistently perform well across languages, while language-specific models show varying effectiveness. The dataset serves as valuable resource for evaluating semantic understanding in RAG pipelines and other NLP applications.

Conclusion: The dataset bridges the benchmark gap for Indic languages, provides versatile evaluation resource for multiple tasks, and demonstrates the importance of multilingual models for low-resource language semantic understanding.

Abstract: Semantic evaluation in low-resource languages remains a major challenge in NLP. While sentence transformers have shown strong performance in high-resource settings, their effectiveness in Indic languages is underexplored due to a lack of high-quality benchmarks. To bridge this gap, we introduce L3Cube-IndicHeadline-ID, a curated headline identification dataset spanning ten low-resource Indic languages: Marathi, Hindi, Tamil, Gujarati, Odia, Kannada, Malayalam, Punjabi, Telugu, Bengali and English. Each language includes 20,000 news articles paired with four headline variants: the original, a semantically similar version, a lexically similar version, and an unrelated one, designed to test fine-grained semantic understanding. The task requires selecting the correct headline from the options using article-headline similarity. We benchmark several sentence transformers, including multilingual and language-specific models, using cosine similarity. Results show that multilingual models consistently perform well, while language-specific models vary in effectiveness. Given the rising use of similarity models in Retrieval-Augmented Generation (RAG) pipelines, this dataset also serves as a valuable resource for evaluating and improving semantic understanding in such applications. Additionally, the dataset can be repurposed for multiple-choice question answering, headline classification, or other task-specific evaluations of LLMs, making it a versatile benchmark for Indic NLP. The dataset is shared publicly at https://github.com/l3cube-pune/indic-nlp

[134] The Forgotten Code: Validating a Century-Old Translation System with AI

Jean-Marie Le Ray

Main category: cs.CL

TL;DR: AI successfully revives and validates Federico Pucci’s 1931 mechanical translation system by reproducing his translations with minimal differences, demonstrating its historical significance and potential for modern applications.

DetailsMotivation: To breathe new life into Federico Pucci's pioneering 1931 mechanical translation system, validate its methodology using modern AI, and establish Pucci's historical importance as a precursor to modern machine translation.

Method: Using AI to retranslate the exact same text excerpts that Pucci translated in 1931 (Dante’s La Vita Nuova from Italian to French and Voltaire’s Zadig from French to Italian) following Pucci’s original method of international keys and ideograms.

Result: The AI translations showed low average difference from Pucci’s 1931 translations with only minor variations. The method was then successfully extended to English, Spanish, and German translations, and applied to modern technical texts.

Conclusion: Pucci’s 1931 mechanical translation system is validated and proven effective, establishing him as a significant historical figure in machine translation alongside pioneers like Troyanskij, Booth, and Weaver, with implications for rewriting the history of the field.

Abstract: A pioneering rule-based mechanical translation system (precursor of modern RBMTs) was first presented in December 1929 by its inventor, Federico Pucci, who later published the full method in a book titled “Il traduttore meccanico ed il metodo per corrispondersi fra Europei conoscendo ciascuno solo la propria lingua: Parte I”, in Salerno (Italy), in 1931. This study illustrates how AI breathes new life into the system of international keys and ideograms devised by Pucci to translate from/into any Romance language (at least as a first step). The methodology involves having the AIs retranslate, following Pucci’s method, the two text excerpts originally translated in 1931 and clearly documented in his publication: a passage from Dante’s La Vita Nuova, translated from Italian into French, and a passage from Voltaire’s Zadig, translated from French into Italian. The result is notable: the two texts, translated 94 years apart using the same method–by Pucci in 1931 and by AIs in 2025–show a low average difference, with only minor variations observed. With Pucci’s system thus validated, it became feasible to have the AIs reproduce the excerpts in English, Spanish, and German according to his method. The results were consistent, and Pucci–via Artificial Intelligence–was tasked with translating more modern and technical texts, thereby reviving, nearly a century later, an invention that had remained almost entirely unknown and never applied beyond its creator, now brought to wider attention and opened to possible experimentation. Such a demonstration would not only affirm Pucci’s historical status but also place him among the precursors and intellectual contributors to machine translation, whose work merits examination alongside figures such as Troyanskij, Booth, and Weaver, with possible consequences for how the history of the field is understood.

[135] Top-H Decoding: Adapting the Creativity and Coherence with Bounded Entropy in Text Generation

Erfan Baghaei Potraghloo, Seyedarmin Azizi, Souvik Kundu, Massoud Pedram

Main category: cs.CL

TL;DR: Top-H decoding is a new sampling method that better balances creativity and coherence in LLM text generation by effectively incorporating model confidence information, outperforming min-p sampling by up to 25.63% on creative writing tasks.

DetailsMotivation: Existing truncated sampling techniques (temperature scaling, top-p, min-p) struggle to effectively incorporate model confidence information, often relying on limited heuristics that underutilize the full probability distribution.

Method: The authors formulate an entropy-constrained minimum divergence problem, prove it’s equivalent to an NP-hard entropy-constrained mass maximization problem, and develop top-H decoding as a computationally efficient greedy algorithm to solve it.

Result: Top-H outperforms min-p sampling by up to 25.63% on creative writing benchmarks while maintaining robustness on question-answering datasets (GPQA, GSM8K, MT-Bench). LLM-as-judge evaluation confirms it produces coherent outputs even at higher temperatures.

Conclusion: Top-H advances state-of-the-art in open-ended text generation and can be easily integrated into creative writing applications, providing better balance between diversity/creativity and logical coherence.

Abstract: Large language models (LLMs), despite their impressive performance across a wide range of tasks, often struggle to balance two competing objectives in open-ended text generation: fostering diversity and creativity while preserving logical coherence. Existing truncated sampling techniques, including temperature scaling, top-$p$ (nucleus) sampling, and min-$p$ sampling, aim to manage this trade-off. However, they exhibit limitations, particularly in the effective incorporation of the confidence of the model into the corresponding sampling strategy. For example, min-$p$ sampling relies on a single top token as a heuristic for confidence, eventually underutilizing the information of the probability distribution. Toward effective incorporation of the confidence of the model, in this paper, we present top-H decoding. We first establish the theoretical foundation of the interplay between creativity and coherence in truncated sampling by formulating an entropy-constrained minimum divergence problem. We then prove this minimization problem to be equivalent to an entropy-constrained mass maximization (ECMM) problem, which is NP-hard. Finally, we present top-H decoding, a computationally efficient greedy algorithm to solve the ECMM problem. Extensive empirical evaluations demonstrate that top-H outperforms the state-of-the-art (SoTA) alternative of min-$p$ sampling by up to 25.63% on creative writing benchmarks, while maintaining robustness on question-answering datasets such as GPQA, GSM8K, and MT-Bench. Additionally, an LLM-as-judge evaluation confirms that top-H indeed produces coherent outputs even at higher temperatures, where creativity is especially critical. In summary, top-H advances SoTA in open-ended text generation and can be easily integrated into creative writing applications. The code is available at https://github.com/ErfanBaghaei/Top-H-Decoding.

[136] Comparative Study of Pre-Trained BERT and Large Language Models for Code-Mixed Named Entity Recognition

Mayur Shirke, Amey Shembade, Pavan Thorat, Madhushri Wagh, Raviraj Joshi

Main category: cs.CL

TL;DR: Comparative evaluation shows code-mixed fine-tuned models (HingRoBERTa, HingBERT) outperform multilingual models and zero-shot LLMs like Google Gemini for Hindi-English NER tasks, demonstrating the value of domain-specific pretraining.

DetailsMotivation: NER in code-mixed text like Hinglish presents unique challenges due to informal structure, transliteration, and language switching, requiring specialized approaches beyond standard multilingual models.

Method: Comparative evaluation of code-mixed fine-tuned models (HingBERT, HingMBERT, HingRoBERTa) vs non-code-mixed multilingual models (BERT Base Cased, IndicBERT, RoBERTa, MuRIL) and zero-shot Google Gemini on benchmark Hinglish NER dataset using Precision, Recall, and F1-score metrics.

Result: Code-mixed models, particularly HingRoBERTa and HingBERT-based models, outperformed all others including Google Gemini. Non-code-mixed models showed limited adaptability. Google Gemini demonstrated competitive zero-shot performance despite not being specifically trained on code-mixed data.

Conclusion: Specialized code-mixed models are most effective for Hinglish NER tasks due to domain-specific pretraining, though modern LLMs show strong generalization capabilities in zero-shot settings, highlighting the trade-off between specialization and generalization.

Abstract: Named Entity Recognition (NER) in code-mixed text, particularly Hindi-English (Hinglish), presents unique challenges due to informal structure, transliteration, and frequent language switching. This study conducts a comparative evaluation of code-mixed fine-tuned models and non-code-mixed multilingual models, along with zero-shot generative large language models (LLMs). Specifically, we evaluate HingBERT, HingMBERT, and HingRoBERTa (trained on code-mixed data), and BERT Base Cased, IndicBERT, RoBERTa and MuRIL (trained on non-code-mixed multilingual data). We also assess the performance of Google Gemini in a zero-shot setting using a modified version of the dataset with NER tags removed. All models are tested on a benchmark Hinglish NER dataset using Precision, Recall, and F1-score. Results show that code-mixed models, particularly HingRoBERTa and HingBERT-based fine-tuned models, outperform others - including closed-source LLMs like Google Gemini - due to domain-specific pretraining. Non-code-mixed models perform reasonably but show limited adaptability. Notably, Google Gemini exhibits competitive zero-shot performance, underlining the generalization strength of modern LLMs. This study provides key insights into the effectiveness of specialized versus generalized models for code-mixed NER tasks.

[137] Implicit Actor Critic Coupling via a Supervised Learning Framework for RLVR

Jiaming Li, Longze Chen, Ze Gong, Yukun Chen, Lu Wang, Wanwei He, Run Luo, Min Yang

Main category: cs.CL

TL;DR: PACS is a novel RLVR framework that reformulates reinforcement learning with verifiable rewards as a supervised learning task, achieving implicit actor-critic coupling through cross-entropy loss optimization for more stable training.

DetailsMotivation: Existing RLVR methods suffer from sparse reward signals and unstable policy gradient updates, particularly in RL-based approaches, limiting their effectiveness in guiding LLMs for reasoning tasks.

Method: PACS treats outcome rewards as predictable labels and reformulates RLVR as a supervised learning task over a score function parameterized by the policy model, optimized using cross-entropy loss to achieve implicit actor-critic coupling.

Result: PACS outperforms strong RLVR baselines (PPO and GRPO) on mathematical reasoning tasks, achieving 59.78% at pass@256 on AIME 2025 with improvements of 13.32 and 14.36 points over PPO and GRPO respectively.

Conclusion: The supervised learning formulation provides a simple yet powerful framework for LLM post-training with verifiable rewards, offering more stable and efficient training while recovering classical policy gradient updates implicitly.

Abstract: Recent advances in Reinforcement Learning with Verifiable Rewards (RLVR) have empowered large language models (LLMs) to tackle challenging reasoning tasks such as mathematics and programming. RLVR leverages verifiable outcome rewards to guide policy optimization, enabling LLMs to progressively improve output quality in a grounded and reliable manner. Despite its promise, the RLVR paradigm poses significant challenges, as existing methods often suffer from sparse reward signals and unstable policy gradient updates, particularly in RL-based approaches. To address the challenges, we propose $\textbf{PACS}$, a novel RLVR framework that achieves im$\textbf{P}$licit $\textbf{A}$ctor $\textbf{C}$ritic coupling via a $\textbf{S}$upervised learning framework. By treating the outcome reward as a predictable label, we reformulate the RLVR problem into a supervised learning task over a score function parameterized by the policy model and optimized using cross-entropy loss. A detailed gradient analysis shows that this supervised formulation inherently recovers the classical policy gradient update while implicitly coupling actor and critic roles, yielding more stable and efficient training. Benchmarking on challenging mathematical reasoning tasks, PACS outperforms strong RLVR baselines, such as PPO and GRPO, achieving superior reasoning performance. For instance, PACS achieves 59.78% at pass@256 on AIME 2025, representing improvements of 13.32 and 14.36 points over PPO and GRPO. This simple yet powerful framework offers a promising avenue for LLMs post-training with verifiable rewards. Our code and data are available as open source at https://github.com/ritzz-ai/PACS.

[138] Jointly Reinforcing Diversity and Quality in Language Model Generations

Tianjian Li, Yiming Zhang, Ping Yu, Swarnadeep Saha, Daniel Khashabi, Jason Weston, Jack Lanchantin, Tianlu Wang

Main category: cs.CL

TL;DR: DARLING is a diversity-aware RL framework that optimizes for both response quality and semantic diversity in LLMs, outperforming quality-only approaches across creative and verifiable tasks.

DetailsMotivation: Post-training of LLMs prioritizes accuracy and helpfulness at the expense of diversity, limiting their usefulness in creative and exploratory tasks like brainstorming and problem solving.

Method: DARLING introduces a learned partition function to measure semantic diversity beyond lexical variations, combining this diversity signal with quality rewards during online reinforcement learning.

Result: DARLING outperforms quality-only RL baselines on five benchmarks for non-verifiable tasks (higher quality and novelty) and achieves higher pass@1 and pass@k for verifiable tasks like math problems.

Conclusion: Explicitly optimizing for diversity catalyzes exploration in online RL, leading to higher-quality responses while maintaining semantic diversity across various model families and sizes.

Abstract: Post-training of Large Language Models (LMs) often prioritizes accuracy and helpfulness at the expense of diversity. This creates a tension: while post-training improves response quality, it also sharpens output distributions and reduces the range of ideas, limiting the usefulness of LMs in creative and exploratory tasks such as brainstorming, storytelling, or problem solving. We address this challenge with Diversity-Aware Reinforcement Learning (DARLING), a framework that jointly optimizes for response quality and semantic diversity. At its core, DARLING introduces a learned partition function to measure diversity beyond surface-level lexical variations. This diversity signal is then combined with a quality reward during online reinforcement learning, encouraging models to generate outputs that are both high-quality and distinct. Experiments across multiple model families and sizes show that DARLING generalizes to two regimes: non-verifiable tasks (instruction following and creative writing) and verifiable tasks (competition math). On five benchmarks in the first setting, DARLING consistently outperforms quality-only RL baselines, producing outputs that are simultaneously of higher quality and novelty. In the second setting, DARLING achieves higher pass@1 (solution quality) and pass@k (solution variety). Most strikingly, explicitly optimizing for diversity catalyzes exploration in online RL, which manifests itself as higher-quality responses.

[139] PalmX 2025: The First Shared Task on Benchmarking LLMs on Arabic and Islamic Culture

Fakhraddin Alwajih, Abdellah El Mekki, Hamdy Mubarak, Majd Hawasly, Abubakr Mohamed, Muhammad Abdul-Mageed

Main category: cs.CL

TL;DR: PalmX 2025 is the first shared task benchmarking LLMs’ cultural competence in Arabic and Islamic domains, showing fine-tuning significantly improves performance on cultural and religious knowledge.

DetailsMotivation: LLMs reflect skewed web data distributions favoring Western cultures, leading to diminished understanding of Arabic and Islamic communities, especially for underrepresented topics.

Method: Two subtasks with multiple-choice questions in Modern Standard Arabic covering General Arabic Culture and General Islamic Culture across 22 Arab countries, with 26 and 19 teams participating respectively.

Result: Task-specific fine-tuning substantially boosted performance over baselines, with top systems achieving 72.15% accuracy on cultural questions and 84.22% on Islamic knowledge. Parameter-efficient fine-tuning was most effective.

Conclusion: The benchmark reveals LLMs’ cultural competence gaps and demonstrates that fine-tuning, particularly parameter-efficient methods, can significantly improve performance on underrepresented cultural domains.

Abstract: Large Language Models (LLMs) inherently reflect the vast data distributions they encounter during their pre-training phase. As this data is predominantly sourced from the web, there is a high chance it will be skewed towards high-resourced languages and cultures, such as those of the West. Consequently, LLMs often exhibit a diminished understanding of certain communities, a gap that is particularly evident in their knowledge of Arabic and Islamic cultures. This issue becomes even more pronounced with increasingly under-represented topics. To address this critical challenge, we introduce PalmX 2025, the first shared task designed to benchmark the cultural competence of LLMs in these specific domains. The task is composed of two subtasks featuring multiple-choice questions (MCQs) in Modern Standard Arabic (MSA): General Arabic Culture and General Islamic Culture. These subtasks cover a wide range of topics, including traditions, food, history, religious practices, and language expressions from across 22 Arab countries. The initiative drew considerable interest, with 26 teams registering for Subtask 1 and 19 for Subtask 2, culminating in nine and six valid submissions, respectively. Our findings reveal that task-specific fine-tuning substantially boosts performance over baseline models. The top-performing systems achieved an accuracy of 72.15% on cultural questions and 84.22% on Islamic knowledge. Parameter-efficient fine-tuning emerged as the predominant and most effective approach among participants, while the utility of data augmentation was found to be domain-dependent.

[140] Similarity between Units of Natural Language: The Transition from Coarse to Fine Estimation

Wenchuan Mu

Main category: cs.CL

TL;DR: Develops a progressively refined similarity computation framework that combines attack testing with adversarial training to improve similarity regression models by catching loopholes and providing explanations.

DetailsMotivation: Existing similarity computation methods often rely on continuous fitting to human judgments, have vague definitions, and lack interpretability. There's a need for more precise similarity measures, especially in critical domains like legal and medical affairs where small language unit differences can have significant real-world consequences.

Method: Proposes a progressively refined similarity computation framework that combines attack testing with adversarial training. The algorithm constantly improves the model by catching different loopholes in similarity calculations and provides reasonable explanations for each refinement.

Result: The regression model achieves state-of-the-art performance in handling edge cases, demonstrating improved precision in similarity computation.

Conclusion: The proposed framework successfully addresses the shortcomings of existing similarity computation methods by providing both improved performance through adversarial training and better interpretability through explanatory refinements, making it particularly valuable for critical applications requiring precision.

Abstract: Capturing the similarities between human language units is crucial for explaining how humans associate different objects, and therefore its computation has received extensive attention, research, and applications. With the ever-increasing amount of information around us, calculating similarity becomes increasingly complex, especially in many cases, such as legal or medical affairs, measuring similarity requires extra care and precision, as small acts within a language unit can have significant real-world effects. My research goal in this thesis is to develop regression models that account for similarities between language units in a more refined way. Computation of similarity has come a long way, but approaches to debugging the measures are often based on continually fitting human judgment values. To this end, my goal is to develop an algorithm that precisely catches loopholes in a similarity calculation. Furthermore, most methods have vague definitions of the similarities they compute and are often difficult to interpret. The proposed framework addresses both shortcomings. It constantly improves the model through catching different loopholes. In addition, every refinement of the model provides a reasonable explanation. The regression model introduced in this thesis is called progressively refined similarity computation, which combines attack testing with adversarial training. The similarity regression model of this thesis achieves state-of-the-art performance in handling edge cases.

[141] Rule-Guided Joint Embedding Learning over Knowledge Graphs

Qisong Li, Ji Lin, Sijia Wei, Neng Liu

Main category: cs.CL

TL;DR: A novel knowledge graph embedding model that integrates contextual and textual information using graph convolutional networks with confidence and relatedness metrics for improved weighting.

DetailsMotivation: Most existing knowledge graph embedding models focus only on structural information, but knowledge graphs contain rich contextual and textual information that could enhance embedding effectiveness.

Method: Proposes a model that integrates contextual and textual signals through graph convolutional networks, introducing two metrics (confidence via rule-based method and relatedness from textual representations) for precise weighting of contextual information.

Result: Extensive experiments on two benchmark datasets show consistent improvements over strong baselines, demonstrating the effectiveness of the approach.

Conclusion: The integration of contextual and textual information with proper weighting metrics significantly enhances knowledge graph embedding performance compared to models that only use structural information.

Abstract: Recent studies on knowledge graph embedding focus on mapping entities and relations into low-dimensional vector spaces. While most existing models primarily exploit structural information, knowledge graphs also contain rich contextual and textual information that can enhance embedding effectiveness. In this work, we propose a novel model that integrates both contextual and textual signals into entity and relation embeddings through a graph convolutional network. To better utilize context, we introduce two metrics: confidence, computed via a rule-based method, and relatedness, derived from textual representations. These metrics enable more precise weighting of contextual information during embedding learning. Extensive experiments on two widely used benchmark datasets demonstrate the effectiveness of our approach, showing consistent improvements over strong baselines.

[142] Semantic Parsing for Question Answering over Knowledge Graphs

Sijia Wei, Wenwen Zhang, Qisong Li, Jiang Zhao

Main category: cs.CL

TL;DR: Novel graph-to-segment mapping method for knowledge graph question answering that combines rule-based and neural approaches to handle implicit entities, relations, and complex constraints through semantic parsing.

DetailsMotivation: To improve natural language question understanding over knowledge graphs by addressing challenges with implicit entities/relations and complex constraints like temporal conditions and aggregation.

Method: Integrates rule-based and neural methods to parse questions into semantic segment sequences, formulates parsing as sequence generation using encoder-decoder network, and employs graph neural network to leverage KG context for better implicit entity/relation identification.

Result: Experimental evaluations on two benchmark datasets demonstrate the model’s effectiveness and superior performance in semantic parsing for knowledge graph question answering.

Conclusion: The proposed framework successfully handles complex question parsing challenges and shows strong performance in knowledge graph question answering through integrated rule-based and neural approaches with graph context enrichment.

Abstract: In this paper, we propose a novel method for question answering over knowledge graphs based on graph-to-segment mapping, designed to improve the understanding of natural language questions. Our approach is grounded in semantic parsing, a key technique for interpreting question utterances. The main challenges arise from handling implicit entities and relations, as well as complex constraints such as temporal conditions, ordinality, and aggregation within the context of a knowledge graph. To address these issues, our framework integrates both rule-based and neural methods to parse and construct accurate, comprehensive semantic segment sequences. These sequences are then assembled into semantic query graphs, providing precise representations of question utterances. We formulate question semantic parsing as a sequence generation task, employing an encoder-decoder neural network to map natural language questions into semantic segments. Furthermore, to enhance the identification of implicit entities and relations, we incorporate a graph neural network that leverages knowledge graph context to enrich question representations. Experimental evaluations on two benchmark datasets demonstrate the effectiveness and superior performance of our model in semantic parsing for knowledge graph question answering.

[143] Into the crossfire: evaluating the use of a language model to crowdsource gun violence reports

Adriano Belisario, Scott A. Hale, Luc Rocher

Main category: cs.CL

TL;DR: Fine-tuned BERT model helps Brazilian human rights organization monitor gun violence from social media data, improving analyst efficiency and engagement.

DetailsMotivation: Gun violence is a critical human rights issue requiring reliable data, but traditional data collection methods are limited. Social media provides potential data source but manual monitoring is inefficient.

Method: Developed fine-tuned BERT-based model trained on Twitter texts to distinguish gun violence reports from ordinary Portuguese texts. Integrated into web application and tested in live intervention with Brazilian analysts.

Result: Qualitative: All analysts used time more efficiently and expanded search capacities. Quantitative: Model use associated with increased interactions with users reporting gun violence.

Conclusion: Human-centered interventions using language models can effectively support human rights organizations’ work in monitoring real-world firearm events.

Abstract: Gun violence is a pressing human rights issue that affects nearly every dimension of the social fabric, from healthcare and education to psychology and the economy. Reliable data on firearm events is paramount to developing more effective public policy and emergency responses. However, the lack of comprehensive databases and the risks of in-person surveys prevent human rights organizations from collecting needed data in most countries. Here, we partner with a Brazilian human rights organization to conduct a systematic evaluation of language models to assist with monitoring real-world firearm events from social media data. We propose a fine-tuned BERT-based model trained on Twitter (now X) texts to distinguish gun violence reports from ordinary Portuguese texts. We then incorporate our model into a web application and test it in a live intervention. We study and interview Brazilian analysts who continuously check social media texts to identify new gun violence events. Qualitative assessments show that our solution helped all analysts use their time more efficiently and expanded their search capacities. Quantitative assessments show that the use of our model was associated with analysts having further interactions with online users reporting gun violence. Our findings suggest that human-centered interventions using language models can help support the work of human rights organizations.

[144] Whose LLM is it Anyway? Linguistic Comparison and LLM Attribution for GPT-3.5, GPT-4 and Bard

Ariel Rosenfeld, Teddy Lazebnik

Main category: cs.CL

TL;DR: LLMs exhibit distinctive linguistic styles that can be identified with 88% accuracy using simple classification, revealing significant variations in vocabulary, POS, dependency, and sentiment across different models.

DetailsMotivation: To determine whether Large Language Models develop unique linguistic signatures similar to human authors, despite their capability to generate human-quality text.

Method: Comprehensive linguistic analysis comparing vocabulary, Part-Of-Speech distribution, dependency distribution, and sentiment of texts generated by GPT-3.5, GPT-4, and Bard against diverse inputs.

Result: Significant linguistic variations were found that enable attribution of text to its LLM origin with 88% accuracy using off-the-shelf classification models.

Conclusion: LLMs do exhibit distinctive linguistic styles, which has important theoretical and practical implications for AI-generated text detection and model identification.

Abstract: Large Language Models (LLMs) are capable of generating text that is similar to or surpasses human quality. However, it is unclear whether LLMs tend to exhibit distinctive linguistic styles akin to how human authors do. Through a comprehensive linguistic analysis, we compare the vocabulary, Part-Of-Speech (POS) distribution, dependency distribution, and sentiment of texts generated by three of the most popular LLMS today (GPT-3.5, GPT-4, and Bard) to diverse inputs. The results point to significant linguistic variations which, in turn, enable us to attribute a given text to its LLM origin with a favorable 88% accuracy using a simple off-the-shelf classification model. Theoretical and practical implications of this intriguing finding are discussed.

[145] Guiding Large Language Models to Post-Edit Machine Translation with Error Annotations

Dayeon Ki, Marine Carpuat

Main category: cs.CL

TL;DR: LLMs guided by MQM feedback to post-edit machine translations show improved quality metrics, with fine-tuning further enhancing performance.

DetailsMotivation: Machine Translation remains one area where LLMs haven't surpassed dedicated supervised systems, so this work explores combining LLMs' strengths with supervised MT through quality-guided post-editing.

Method: Using LLaMA-2 models, researchers employed prompting strategies with MQM quality feedback and fine-tuned the LLM to better utilize this guidance for MT post-editing across multiple language pairs.

Result: Prompting LLMs to post-edit MT improved TER, BLEU and COMET scores. Fine-tuning helped integrate fine-grained feedback more effectively and further improved translation quality in both automatic and human evaluations.

Conclusion: Combining LLMs with quality feedback from supervised MT systems through prompting and fine-tuning can effectively improve machine translation quality, though the benefits of fine-grained feedback require fine-tuning to be fully realized.

Abstract: Machine Translation (MT) remains one of the last NLP tasks where large language models (LLMs) have not yet replaced dedicated supervised systems. This work exploits the complementary strengths of LLMs and supervised MT by guiding LLMs to automatically post-edit MT with external feedback on its quality, derived from Multidimensional Quality Metric (MQM) annotations. Working with LLaMA-2 models, we consider prompting strategies varying the nature of feedback provided and then fine-tune the LLM to improve its ability to exploit the provided guidance. Through experiments on Chinese-English, English-German, and English-Russian MQM data, we demonstrate that prompting LLMs to post-edit MT improves TER, BLEU and COMET scores, although the benefits of fine-grained feedback are not clear. Fine-tuning helps integrate fine-grained feedback more effectively and further improves translation quality based on both automatic and human evaluation.

[146] Why Not Transform Chat Large Language Models to Non-English?

Xiang Geng, Ming Zhu, Jiahuan Li, Zhejian Lai, Wei Zou, Shuaijie She, Jiaxin Guo, Xiaofeng Zhao, Yinglu Li, Yuang Li, Chang Su, Yanqing Zhao, Xinglin Lyu, Min Zhang, Jiajun Chen, Hao Yang, Shujian Huang

Main category: cs.CL

TL;DR: TransLLM is a framework that transforms English chat LLMs to non-English languages by using translation chain-of-thought and recovery knowledge distillation, achieving better performance than ChatGPT and GPT-4 on Thai language benchmarks.

DetailsMotivation: The scarcity of non-English data limits non-English LLM development. While transforming English-centric LLMs is effective, chat LLMs present unique challenges: transferring advanced abilities without supervised data and preventing catastrophic forgetting of original knowledge.

Method: TransLLM uses translation chain-of-thought to divide transfer into sub-tasks, enhanced with public data. It employs low-rank adaptation to maintain original parameters and recovery KD using data generated by the chat LLM itself to preserve original knowledge.

Result: Transforming LLaMA-2-chat-7B to Thai, the method outperformed strong baselines and ChatGPT on multi-turn MT-bench using only single-turn data. It also rejected more harmful queries on AdvBench safety benchmark than both ChatGPT and GPT-4 without safety data.

Conclusion: TransLLM effectively transfers advanced abilities from English chat LLMs to non-English languages while preventing catastrophic forgetting, demonstrating superior performance in both helpfulness and safety compared to existing models.

Abstract: The scarcity of non-English data limits the development of non-English large language models (LLMs). Transforming English-centric LLMs to non-English has been identified as an effective and resource-efficient method. Previous works start from base LLMs and perform knowledge distillation (KD) with data generated by stronger LLMs, e.g. GPT-4. Compared to base LLMs, chat LLMs are further optimized for advanced abilities, e.g. multi-turn conversation and human preference alignment, and thus more powerful in both helpfulness and safety. However, transforming a chat LLM involves two critical issues: (1) How can we effectively transfer advanced abilities without their supervised data? (2) How can we prevent the original knowledge from catastrophic forgetting during transformation? We target these issues by introducing a simple framework called TransLLM. For the first issue, TransLLM divides the transfer problem into some common sub-tasks with the translation chain-of-thought, which uses the translation as the bridge between English and non-English step-by-step. We further enhance the performance of sub-tasks with publicly available data. For the second issue, we propose a method comprising two synergistic components: low-rank adaptation for training to maintain the original LLM parameters, and recovery KD, which utilizes data generated by the chat LLM itself to recover the original knowledge from the frozen parameters. In the experiments, we transform the LLaMA-2-chat-7B to the Thai language. Our method, using only single-turn data, outperforms strong baselines and ChatGPT on multi-turn benchmark MT-bench. Furthermore, our method, without safety data, rejects more harmful queries of safety benchmark AdvBench than both ChatGPT and GPT-4. Code is available at https://github.com/hy5468/TransLLM.

[147] Intrinsic Test of Unlearning Using Parametric Knowledge Traces

Yihuai Hong, Lei Yu, Haiqin Yang, Shauli Ravfogel, Mor Geva

Main category: cs.CL

TL;DR: Current unlearning evaluation methods rely on behavioral tests but fail to detect residual parametric knowledge that can be exploited. The paper proposes parameter-based evaluation using concept vectors and shows existing methods only suppress concepts rather than truly removing them.

DetailsMotivation: Behavioral tests for unlearning evaluation don't monitor residual knowledge in model parameters, which can be adversarially exploited to recover erased information. There's a need for internal evaluation of parametric knowledge traces.

Method: Proposes vocabulary projections to inspect concepts encoded in parameters, localizes concept vectors, and creates ConceptVectors benchmark with hundreds of concepts and their parametric traces in open-source LLMs.

Result: Existing unlearning methods minimally impact concept vectors and mostly suppress them during inference. Direct ablation of concept vectors removes associated knowledge and reduces susceptibility to adversarial manipulation.

Conclusion: Behavioral-based unlearning evaluations have limitations. Future work should include parameter-based evaluations. The authors release code and benchmark to support this approach.

Abstract: The task of “unlearning” certain concepts in large language models (LLMs) has attracted immense attention recently, due to its importance in mitigating undesirable model behaviours, such as the generation of harmful, private, or incorrect information. Current protocols to evaluate unlearning methods largely rely on behavioral tests, without monitoring the presence of unlearned knowledge within the model’s parameters. This residual knowledge can be adversarially exploited to recover the erased information post-unlearning. We argue that unlearning should also be evaluated internally, by considering changes in the parametric knowledge traces of the unlearned concepts. To this end, we propose a general evaluation methodology that leverages vocabulary projections to inspect concepts encoded in model parameters. We use this approach to localize “concept vectors” - parameter vectors that encode concrete concepts - and construct ConceptVectors, a benchmark dataset containing hundreds of common concepts and their parametric knowledge traces within two open-source LLMs. Evaluation on ConceptVectors shows that existing unlearning methods minimally impact concept vectors and mostly suppress them during inference, while directly ablating these vectors demonstrably removes the associated knowledge and significantly reduces the model’s susceptibility to adversarial manipulation. Our results highlight limitations in behavioral-based unlearning evaluations and call for future work to include parameter-based evaluations. To support this, we release our code and benchmark at https://github.com/yihuaihong/ConceptVectors.

[148] MEGen: Generative Backdoor into Large Language Models via Model Editing

Jiyang Qiu, Xinbei Ma, Zhuosheng Zhang, Hai Zhao, Yun Li, Qianren Wang

Main category: cs.CL

TL;DR: MEGen is an editing-based generative backdoor method that enables LLMs to output dangerous information when triggered, expanding backdoor attacks from discriminative to generative tasks.

DetailsMotivation: Traditional backdoor injection methods are limited to yes-or-no discriminative tasks, causing users to underestimate the safety risks of backdoored LLMs. Given LLMs' generative nature, there's a need to reveal the true safety risks through generative backdoors.

Method: Proposed MEGen, an editing-based generative backdoor that expands backdoors to generative tasks in a unified any-text-to-any-text format. It adjusts only a small set of local parameters with few-shot samples to achieve natural generations with specific intentions.

Result: MEGen achieves high attack success rate. The backdoored model can freely output pre-set dangerous information while completing downstream tasks when triggered.

Conclusion: MEGen enables backdoors in LLMs to exhibit generative capabilities, causing potential safety risks by altering generative style, highlighting the expanded threat landscape of backdoored LLMs.

Abstract: Large language models (LLMs) have exhibited remarkable versatility and adaptability, while their widespread adoption across various applications also raises critical safety concerns. This paper focuses on the impact of backdoored LLMs. Traditional backdoor injection methods are primarily limited to yes-or-no discriminative tasks, leading users to underestimate the potential risks of backdoored LLMs. Given the inherently generative nature of LLMs, this paper reveals that a generative backdoor injected into LLMs can expose the true safety risks in their applications. We propose an editing-based generative backdoor, named MEGen, aiming to expand the backdoor to generative tasks in a unified format of any text-to any text, leading to natural generations with a specific intention. Experiments show that MEGen achieves a high attack success rate by adjusting only a small set of local parameters with few-shot samples. Notably, we show that the backdoored model, when triggered, can freely output pre-set dangerous information while completing downstream tasks. Our work highlights that MEGen enables backdoors in LLMs to exhibit generative capabilities, causing potential safety risks by altering the generative style. The code is available at https://github.com/MonoQ-hub/MEGen.

[149] On the Diagram of Thought

Yifan Zhang, Yang Yuan, Andrew Chi-Chih Yao

Main category: cs.CL

TL;DR: DoT is a framework that enables LLMs to build dynamic mental diagrams for complex reasoning, using category theory to ensure logical consistency and produce auditable reasoning traces.

DetailsMotivation: LLMs struggle with complex multi-step reasoning problems that require structured thinking, often producing linear and error-prone reasoning processes.

Method: The Diagram of Thought framework allows LLMs to construct dynamic diagrams of ideas, enabling them to propose multiple lines of thought, self-critique steps, and synthesize insights. Grounded in category theory for logical consistency.

Result: Creates a more powerful and transparent reasoning process that produces fully auditable, step-by-step traces of the LLM’s thinking without requiring external controllers.

Conclusion: DoT bridges the gap between fluent language generation and formal reasoning, providing a self-contained, efficient framework for complex problem-solving with guaranteed logical consistency.

Abstract: Large Language Models (LLMs) excel at many tasks but often falter on complex problems that require structured, multi-step reasoning. We introduce the Diagram of Thought (DoT), a new framework that enables a single LLM to build and navigate a mental map of its reasoning. Instead of thinking in a straight line, the model constructs a dynamic diagram of ideas, where it can propose different lines of thought, critique its own steps, and synthesize validated insights into a final conclusion. This entire process is self-contained within the model, making it highly efficient by avoiding the complex external controllers or search algorithms required by other methods. To ensure the reliability of this process, we ground DoT in a rigorous mathematical framework from category theory. This foundation guarantees that the way the model combines information is logical, consistent, and robust, regardless of the order in which ideas were explored. The result is a more powerful and transparent reasoning process that produces a fully auditable, step-by-step trace of the LLM’s thinking, bridging the gap between fluent language and formal reasoning.

[150] Mitigating Semantic Leakage in Cross-lingual Embeddings via Orthogonality Constraint

Dayeon Ki, Cheonbok Park, Hyunjoong Kim

Main category: cs.CL

TL;DR: Proposes ORACLE method to address semantic leakage in cross-lingual embeddings by enforcing orthogonality between semantic and language representations.

DetailsMotivation: Current disentangled representation learning methods suffer from semantic leakage where language-specific information contaminates semantic representations, hindering effective cross-lingual alignment.

Method: ORACLE training objective with intra-class clustering and inter-class separation to enforce orthogonality between semantic and language embeddings.

Result: ORACLE effectively reduces semantic leakage and enhances semantic alignment in cross-lingual retrieval and semantic textual similarity tasks.

Conclusion: The proposed orthogonality constraint learning approach successfully mitigates semantic leakage and improves cross-lingual representation alignment.

Abstract: Accurately aligning contextual representations in cross-lingual sentence embeddings is key for effective parallel data mining. A common strategy for achieving this alignment involves disentangling semantics and language in sentence embeddings derived from multilingual pre-trained models. However, we discover that current disentangled representation learning methods suffer from semantic leakage - a term we introduce to describe when a substantial amount of language-specific information is unintentionally leaked into semantic representations. This hinders the effective disentanglement of semantic and language representations, making it difficult to retrieve embeddings that distinctively represent the meaning of the sentence. To address this challenge, we propose a novel training objective, ORthogonAlity Constraint LEarning (ORACLE), tailored to enforce orthogonality between semantic and language embeddings. ORACLE builds upon two components: intra-class clustering and inter-class separation. Through experiments on cross-lingual retrieval and semantic textual similarity tasks, we demonstrate that training with the ORACLE objective effectively reduces semantic leakage and enhances semantic alignment within the embedding space.

[151] Learning by Surprise: Surplexity for Mitigating Model Collapse in Generative AI

Daniele Gambetta, Gizem Gezici, Fosca Giannotti, Dino Pedreschi, Alistair Knott, Luca Pappalardo

Main category: cs.CL

TL;DR: Model collapse occurs when AI models train on their own synthetic outputs, leading to performance degradation. This paper introduces new measures to characterize collapse from probability distributions, shows it depends on data complexity and autophagy extent, and proposes filtering by high surplexity as an effective mitigation strategy.

DetailsMotivation: As synthetic content proliferates online, generative AI models risk training on their own outputs (autophagy), causing model collapse. Current characterizations are limited and mitigation methods assume reliable knowledge of data authorship, which is problematic.

Method: Introduces new measures to characterize collapse directly from models’ next-token probability distributions. Proposes filtering training items by high surplexity to maximize model surprise, without needing to distinguish human vs AI-generated data.

Result: Experiments show collapse degree depends on initial training set complexity and autophagy extent. The surplexity-based filtering strategy is at least as effective as human-data baselines and more effective in reducing distributional skewedness.

Conclusion: Model collapse occurs when models train on unsurprising data. The proposed surplexity-based approach provides effective mitigation without requiring data authorship identification, offering more resilient training for AI systems in synthetic-data-saturated environments.

Abstract: As synthetic content increasingly infiltrates the web, generative AI models may be retrained on their own outputs: a process termed “autophagy”. This leads to model collapse: a progressive loss of performance and diversity across generations. Recent studies have examined the emergence of model collapse across various generative AI models and data types, and have proposed mitigation strategies that rely on incorporating human-authored content. However, current characterizations of model collapse remain limited, and existing mitigation methods assume reliable knowledge of whether training data is human-authored or AI-generated. In this paper, we address these gaps by introducing new measures that characterise collapse directly from a model’s next-token probability distributions, rather than from properties of AI-generated text. Using these measures, we show that the degree of collapse depends on the complexity of the initial training set, as well as on the extent of autophagy. Our experiments prompt a new suggestion: that model collapse occurs when a model trains on data that does not “surprise” it. We express this hypothesis in terms of the well-known Free Energy Principle in cognitive science. Building on this insight, we propose a practical mitigation strategy: filtering training items by high surplexity, maximising the surprise of the model. Unlike existing methods, this approach does not require distinguishing between human- and AI-generated data. Experiments across datasets and models demonstrate that our strategy is at least as effective as human-data baselines, and even more effective in reducing distributional skewedness. Our results provide a richer understanding of model collapse and point toward more resilient approaches for training generative AI systems in environments increasingly saturated with synthetic data.

[152] How Does Knowledge Selection Help Retrieval Augmented Generation?

Xiangci Li, Jessica Ouyang

Main category: cs.CL

TL;DR: Knowledge selection’s impact on RAG systems depends on generator model strength and task complexity - strong generators on clear tasks benefit more from recall, while weaker generators on ambiguous tasks need better F1 scores from selection.

DetailsMotivation: To empirically analyze how knowledge selection influences downstream generation performance in RAG systems, as prior work focused on retrieval improvement but selection's role remained unclear.

Method: Simulated different retrieval and selection conditions through controlled mixture of gold and distractor knowledge to assess impact on generation outcomes.

Result: Generator model capability and task/dataset complexity significantly influence knowledge selection impact. Strong generators on clear tasks benefit from recall improvement, while weaker generators on ambiguous tasks require better F1 scores from selection.

Conclusion: Knowledge selection’s importance varies based on system context - it provides limited benefit with strong generators on clear tasks but becomes critical for weaker generators or ambiguous tasks where F1 score matters more.

Abstract: Retrieval-augmented generation (RAG) is a powerful method for enhancing natural language generation by integrating external knowledge into a model’s output. While prior work has demonstrated the importance of improving knowledge retrieval for boosting generation quality, the role of knowledge selection, a.k.a. reranking or filtering, remains less clear. This paper empirically analyzes how knowledge selection influences downstream generation performance in RAG systems. By simulating different retrieval and selection conditions through a controlled mixture of gold and distractor knowledge, we assess the impact of these factors on generation outcomes. Our findings indicate that the downstream generator model’s capability, as well as the complexity of the task and dataset, significantly influence the impact of knowledge selection on the overall RAG system performance. In typical scenarios, improving the knowledge recall score is key to enhancing generation outcomes, with the knowledge selector providing limited benefit when a strong generator model is used on clear, well-defined tasks. For weaker generator models or more ambiguous tasks and datasets, the knowledge F1 score becomes a critical factor, and the knowledge selector plays a more prominent role in improving overall performance.

[153] Distill Visual Chart Reasoning Ability from LLMs to MLLMs

Wei He, Zhiheng Xi, Wanxu Zhao, Xiaoran Fan, Yiwen Ding, Zifei Shan, Tao Gui, Qi Zhang, Xuanjing Huang

Main category: cs.CL

TL;DR: Proposes Code-as-Intermediary Translation (CIT) method to synthesize high-quality chart Q&A data using code as intermediary, creating ReachQA dataset to enhance MLLMs’ visual reasoning abilities.

DetailsMotivation: Collecting and annotating charts and questions for multimodal reasoning is expensive, hard to scale, and often results in low-quality annotations, creating a need for cost-effective data synthesis methods.

Method: Uses code as intermediary to translate visual chart representations into textual representations, enabling text-based synthesizing techniques to generate high-quality Q&A pairs from chart-plotting code.

Result: Created ReachQA dataset with 3k reasoning-intensive charts and 20k Q&A pairs. Models fine-tuned with ReachQA perform well on chart tasks and show gains on general reasoning benchmarks.

Conclusion: CIT provides an efficient and scalable data synthesis approach that successfully distills visual reasoning abilities from LLMs to MLLMs, enhancing both chart-specific and general reasoning performance.

Abstract: Solving complex chart Q&A tasks requires advanced visual reasoning abilities in multimodal large language models (MLLMs), including recognizing key information from visual inputs and conducting reasoning over it. While fine-tuning MLLMs for reasoning is critical, collecting and annotating charts and questions is expensive, hard to scale, and often results in low-quality annotations. To address this, we propose Code-as-Intermediary Translation (CIT), a cost-effective, efficient and scalable data synthesis method for distilling visual reasoning abilities from LLMs to MLLMs. The code serves as an intermediary that translates visual chart representations into textual representations, enabling language models to understand cross-modal information and generate reasoning chains accordingly. In this way, we can employ text-based synthesizing techniques to expand chart-plotting code and generate high-quality Q&A pairs for training models. This produces ReachQA, a dataset containing 3k reasoning-intensive charts and 20k Q&A pairs to enhance both recognition and reasoning abilities of MLLMs. Experiments show that models fine-tuned with ReachQA not only perform well on chart-related tasks but also show performance gains on general reasoning benchmarks. The code and dataset are publicly available at https://github.com/hewei2001/ReachQA.

[154] A Computational Method for Measuring “Open Codes” in Qualitative Analysis

John Chen, Alexandros Lotsos, Sihan Cheng, Caiyi Wang, Lexie Zhao, Jessica Hullman, Bruce Sherin, Uri Wilensky, Michael Horn

Main category: cs.CL

TL;DR: A computational method for evaluating inductive coding in qualitative analysis using LLM-enriched codebook merging and four novel metrics (Coverage, Overlap, Novelty, Divergence) to assess human and AI coding quality.

DetailsMotivation: Traditional ground-truth metrics contradict the exploratory nature of inductive coding, and manual evaluation is labor-intensive, especially with increasing use of Generative AI in qualitative analysis.

Method: LLM-enriched algorithm merges individual codebooks, then measures each coder’s contribution using four metrics: Coverage, Overlap, Novelty, and Divergence against the merged result.

Result: Experiments revealed the merging algorithm’s impact, validated metric stability across multiple runs and LLMs, and demonstrated ability to diagnose coding issues like excessive or hallucinated codes.

Conclusion: Provides a reliable pathway for ensuring methodological rigor in human-AI qualitative analysis by offering computational evaluation of inductive coding results.

Abstract: Qualitative analysis is critical to understanding human datasets in many social science disciplines. A central method in this process is inductive coding, where researchers identify and interpret codes directly from the datasets themselves. Yet, this exploratory approach poses challenges for meeting methodological expectations (such as depth'' and variation’’), especially as researchers increasingly adopt Generative AI (GAI) for support. Ground-truth-based metrics are insufficient because they contradict the exploratory nature of inductive coding, while manual evaluation can be labor-intensive. This paper presents a theory-informed computational method for measuring inductive coding results from humans and GAI. Our method first merges individual codebooks using an LLM-enriched algorithm. It measures each coder’s contribution against the merged result using four novel metrics: Coverage, Overlap, Novelty, and Divergence. Through two experiments on a human-coded online conversation dataset, we 1) reveal the merging algorithm’s impact on metrics; 2) validate the metrics’ stability and robustness across multiple runs and different LLMs; and 3) showcase the metrics’ ability to diagnose coding issues, such as excessive or irrelevant (hallucinated) codes. Our work provides a reliable pathway for ensuring methodological rigor in human-AI qualitative analysis.

[155] From Intents to Conversations: Generating Intent-Driven Dialogues with Contrastive Learning for Multi-Turn Classification

Junhua Liu, Yong Keat Tan, Bin Fu, Kwan Hui Lim

Main category: cs.CL

TL;DR: Chain-of-Intent framework combines HMMs and LLMs to generate multilingual, intent-driven dialogues through self-play, with a contrastive learning approach for multi-turn intent classification.

DetailsMotivation: Addressing the challenge of generating large-scale, domain-specific multilingual dialogue datasets for training effective multi-turn intent classification models in conversational AI systems.

Method: Integrates Hidden Markov Models (HMMs) with LLMs to extract intent transition patterns from e-commerce chat logs, parameterize emission probabilities, and uses multi-task contrastive learning (MINT-CL) for classification.

Result: Outperforms competitive baselines in dialogue generation quality and classification accuracy, especially in multilingual settings, and releases MINT-E multilingual dialogue corpus.

Conclusion: The framework successfully generates high-quality multilingual dialogues and improves intent classification performance while reducing reliance on large annotated datasets.

Abstract: In conversational AI systems, a critical challenge in training effective multi-turn intent classification models lies in the generation of large-scale, domain-specific, multilingual dialogue datasets. In this paper, we introduce Chain-of-Intent, a novel framework that integrates Hidden Markov Models (HMMs) with Large Language Models (LLMs) to generate intent-driven, context-aware dialogues through self-play. Our method first extracts domain-specific intent transition patterns from real-world e-commerce chat logs, which guide the modeling of turn-level dynamics and intent sequences. LLMs are then employed to parameterize the emission probabilities of HMMs, enabling the generation of natural, coherent utterances aligned with predicted intents and dialogue context. We also propose MINT-CL, a multi-task contrastive learning framework for multi-turn intent classification, which improves performance while reducing dependence on large-scale annotated datasets. Empirical results demonstrate that our approach outperforms competitive baselines in dialogue generation quality and classification accuracy, particularly in multilingual settings. To facilitate future research, we release MINT-E, a comprehensive, multilingual, intent-aware multi-turn dialogue corpus derived from the e-commerce domain\footnote{The reproduced source code and dataset are available at https://github.com/junhua/chain-of-intent.

[156] Evaluating Language Models as Synthetic Data Generators

Seungone Kim, Juyoung Suk, Xiang Yue, Vijay Viswanathan, Seongyun Lee, Yizhong Wang, Kiril Gashteovski, Carolin Lawrence, Sean Welleck, Graham Neubig

Main category: cs.CL

TL;DR: AgoraBench benchmark reveals that LM data generation ability doesn’t correlate with problem-solving ability, with different LMs excelling at different data generation tasks and intrinsic data quality features being better indicators.

DetailsMotivation: The increasing use of synthetic data in LM post-training makes data generation capability crucial, but prior works lack systematic comparison of different LMs as data generators in unified settings.

Method: Proposed AgoraBench benchmark with standardized settings and metrics, synthesized 1.26M training instances using 6 LMs, and trained 99 student models to evaluate data generation capabilities.

Result: LMs show distinct strengths (GPT-4o excels at new problems, Claude-3.5-Sonnet better at enhancing existing ones), data generation ability doesn’t correlate with problem-solving ability, and intrinsic data quality features are better indicators.

Conclusion: Strategic choices in output format and cost-conscious model selection significantly impact data generation effectiveness, with multiple intrinsic data quality features serving as reliable indicators of generation capability.

Abstract: Given the increasing use of synthetic data in language model (LM) post-training, an LM’s ability to generate high-quality data has become nearly as crucial as its ability to solve problems directly. While prior works have focused on developing effective data generation methods, they lack systematic comparison of different LMs as data generators in a unified setting. To address this gap, we propose AgoraBench, a benchmark that provides standardized settings and metrics to evaluate LMs’ data generation abilities. Through synthesizing 1.26 million training instances using 6 LMs and training 99 student models, we uncover key insights about LMs’ data generation capabilities. First, we observe that LMs exhibit distinct strengths. For instance, GPT-4o excels at generating new problems, while Claude-3.5-Sonnet performs better at enhancing existing ones. Furthermore, our analysis reveals that an LM’s data generation ability doesn’t necessarily correlate with its problem-solving ability. Instead, multiple intrinsic features of data quality-including response quality, perplexity, and instruction difficulty-collectively serve as better indicators. Finally, we demonstrate that strategic choices in output format and cost-conscious model selection significantly impact data generation effectiveness.

[157] Opportunities and Challenges of Large Language Models for Low-Resource Languages in Humanities Research

Tianyang Zhong, Zhenyuan Yang, Zhengliang Liu, Ruidong Zhang, Yiheng Liu, Haiyang Sun, Yi Pan, Yiwei Li, Yifan Zhou, Hanqi Jiang, Junhao Chen, Tianming Liu

Main category: cs.CL

TL;DR: This paper evaluates how large language models (LLMs) can help preserve and study low-resource languages, addressing challenges like data scarcity while enabling linguistic, historical, and cultural research through AI-humanities integration.

DetailsMotivation: Low-resource languages contain invaluable cultural and historical knowledge but face critical challenges including data scarcity and technological limitations that hinder their preservation and study.

Method: Systematic evaluation of LLM applications in low-resource language research, analyzing technical frameworks, current methodologies, ethical considerations, and interdisciplinary approaches.

Result: Identified key challenges including data accessibility, model adaptability, and cultural sensitivity, while highlighting the transformative potential of LLMs for linguistic variation analysis, historical documentation, and cultural expression studies.

Conclusion: Interdisciplinary collaboration and development of customized models are promising avenues for advancing low-resource language research, integrating AI with humanities to preserve humanity’s linguistic and cultural heritage.

Abstract: Low-resource languages serve as invaluable repositories of human history, embodying cultural evolution and intellectual diversity. Despite their significance, these languages face critical challenges, including data scarcity and technological limitations, which hinder their comprehensive study and preservation. Recent advancements in large language models (LLMs) offer transformative opportunities for addressing these challenges, enabling innovative methodologies in linguistic, historical, and cultural research. This study systematically evaluates the applications of LLMs in low-resource language research, encompassing linguistic variation, historical documentation, cultural expressions, and literary analysis. By analyzing technical frameworks, current methodologies, and ethical considerations, this paper identifies key challenges such as data accessibility, model adaptability, and cultural sensitivity. Given the cultural, historical, and linguistic richness inherent in low-resource languages, this work emphasizes interdisciplinary collaboration and the development of customized models as promising avenues for advancing research in this domain. By underscoring the potential of integrating artificial intelligence with the humanities to preserve and study humanity’s linguistic and cultural heritage, this study fosters global efforts towards safeguarding intellectual diversity.

[158] Benchmarking LLMs for Mimicking Child-Caregiver Language in Interaction

Jing Liu, Abdellah Fourtassi

Main category: cs.CL

TL;DR: LLMs can approximate child-caregiver dialogues at basic levels but struggle with discursive patterns, alignment exaggeration, and diversity compared to humans.

DetailsMotivation: To explore how effectively LLMs can simulate early child-adult interactions and capture distinctive features of child-caregiver language.

Method: Used both static and interactive benchmarking methods to evaluate state-of-the-art LLMs like Llama 3 and GPT-4o.

Result: LLMs can approximate dialogues at word and utterance level but fail to reproduce discursive patterns, exaggerate alignment, and lack human-level diversity.

Conclusion: This work aims to initiate development of a comprehensive benchmark for LLMs in child-oriented applications, highlighting current limitations in simulating authentic child-caregiver interactions.

Abstract: LLMs can generate human-like dialogues, yet their ability to simulate early child-adult interactions remains largely unexplored. In this paper, we examined how effectively LLMs can capture the distinctive features of child-caregiver language in interaction, using both static and interactive benchmarking methods. We found that state-of-the-art LLMs like Llama 3 and GPT-4o can approximate child-caregiver dialogues at the word and utterance level, but they struggle to reproduce the child and caregiver’s discursive patterns, exaggerate alignment, and fail to reach the level of diversity shown by humans. The broader goal of this work is to initiate the development of a comprehensive benchmark for LLMs in child-oriented applications.

[159] Truthful Text Sanitization Guided by Inference Attacks

Ildikó Pilán, Benet Manzanares-Salor, David Sánchez, Pierre Lison

Main category: cs.CL

TL;DR: A novel text sanitization method using LLMs to replace PII with broader generalizations that balance privacy protection and content utility through a two-stage process of candidate generation and privacy evaluation.

DetailsMotivation: To address the challenge of balancing privacy protection (preventing personal information leakage) and utility preservation (retaining document content) in text sanitization.

Method: Two-stage approach using instruction-tuned LLMs: 1) Generate truth-preserving replacement candidates ranked by abstraction level, 2) Evaluate privacy protection through LLM inference attacks and select the most informative resistant candidate.

Result: Enhanced utility with only marginal (<1 p.p.) increase in re-identification risk compared to full suppression, and more truth-preserving than existing methods like Microsoft Presidio’s synthetic replacements.

Conclusion: The proposed generalization-based approach using LLMs effectively balances privacy and utility in text sanitization, outperforming existing methods while maintaining minimal privacy risks.

Abstract: Text sanitization aims to rewrite parts of a document to prevent disclosure of personal information. The central challenge of text sanitization is to strike a balance between privacy protection (avoiding the leakage of personal information) and utility preservation (retaining as much as possible of the document’s original content). To this end, we introduce a novel text sanitization method based on generalizations, that is, broader but still informative terms that subsume the semantic content of the original text spans. The approach relies on the use of instruction-tuned large language models (LLMs) and is divided into two stages. Given a document including text spans expressing personally identifiable information (PII), the LLM is first applied to obtain truth-preserving replacement candidates for each text span and rank those according to their abstraction level. Those candidates are then evaluated for their ability to protect privacy by conducting inference attacks with the LLM. Finally, the system selects the most informative replacement candidate shown to be resistant to those attacks. This two-stage process produces replacements that effectively balance privacy and utility. We also present novel metrics to evaluate these two aspects without needing to manually annotate documents. Results on the Text Anonymization Benchmark show that the proposed approach, implemented with Mistral 7B Instruct, leads to enhanced utility, with only a marginal (< 1 p.p.) increase in re-identification risk compared to fully suppressing the original spans. Furthermore, our approach is shown to be more truth-preserving than existing methods such as Microsoft Presidio’s synthetic replacements.

[160] Acquisition of Recursive Possessives and Recursive Locatives in Mandarin

Chenxi Fu, Xiaoyi Wang, Zaijiang Man, Caimei Yang

Main category: cs.CL

TL;DR: Mandarin-speaking children achieve adult-like proficiency in two-level recursive structures by age 6, with notable asymmetry between possessive and locative recursion acquisition.

DetailsMotivation: To understand how children acquire recursive linguistic structures, particularly recursive possessives and locatives in Mandarin, and assess the impact of structural diversity on language acquisition.

Method: Used answering question while seeing a picture task to test comprehension of two-level recursive structures among children aged 3 to 7 years.

Result: Children do not reach adult-like proficiency in two-level recursion until age 6, with significant asymmetry between recursive possessives and locatives acquisition.

Conclusion: Structural complexity and cognitive factors play primary roles in language acquisition, highlighting the importance of recursion in child language development.

Abstract: As recursion has been underlying any linguistic work for the last 60 years, the acquisition of recursive structures by children during language learning has become a focal point of inquiry. This study delves into the developmental trajectory of Mandarin-speaking children’s acquisition of recursive possessives and locatives, assessing the impact of structural diversity on language acquisition. The research contrasts the comprehension of two-level recursive structures among children aged 3 to 7 years, employing answering question while seeing a picture task to elicit responses. The findings indicate that children do not attain adult-like proficiency in two-level recursion until the age of 6, and there exists a notable asymmetry in the acquisition of recursive possessives versus locatives. These results underscore the primacy of structural complexity and cognitive factors in the acquisition process, enhancing our comprehension of the cognitive foundations of language development and the pivotal role of recursion in child language acquisition.

[161] Improving Low-Resource Machine Translation via Cross-Linguistic Transfer from Typologically Similar High-Resource Languages

Saughmon Boujkian

Main category: cs.CL

TL;DR: Transfer learning improves low-resource machine translation across diverse language families when fine-tuning from typologically similar high-resource languages, with optimal results using moderate batch sizes and careful learning rate selection.

DetailsMotivation: To test whether linguistic similarity enables efficient adaptation in transfer learning for low-resource machine translation, reducing the need for extensive training data.

Method: Fine-tuned models initially trained on typologically similar high-resource languages using limited target language data. Experiments conducted on five diverse language pairs spanning Semitic, Bantu, Romance, Slavic language families and a language isolate. Varied hyperparameters (learning rate, batch size, epochs, weight decay) to ensure robustness.

Result: Transfer learning consistently improved translation quality across all language pairs, confirming applicability beyond closely related languages. Moderate batch sizes (e.g., 32) optimal for similar pairs, smaller sizes benefited less similar pairs. Excessively high learning rates destabilized training.

Conclusion: Provides empirical evidence for transfer learning’s generalizability across language families and offers practical guidance for building machine translation systems in low-resource settings with minimal tuning effort.

Abstract: This study examines the cross-linguistic effectiveness of transfer learning for low-resource machine translation by fine-tuning models initially trained on typologically similar high-resource languages, using limited data from the target low-resource language. We hypothesize that linguistic similarity enables efficient adaptation, reducing the need for extensive training data. To test this, we conduct experiments on five typologically diverse language pairs spanning distinct families: Semitic (Modern Standard Arabic to Levantine Arabic), Bantu (Hausa to Zulu), Romance (Spanish to Catalan), Slavic (Slovak to Macedonian), and a language isolate (Eastern Armenian to Western Armenian). Results show that transfer learning consistently improves translation quality across all pairs, confirming its applicability beyond closely related languages. As a secondary analysis, we vary key hyperparameters learning rate, batch size, number of epochs, and weight decay to ensure results are not dependent on a single configuration. We find that moderate batch sizes (e.g., 32) are often optimal for similar pairs, smaller sizes benefit less similar pairs, and excessively high learning rates can destabilize training. These findings provide empirical evidence for the generalizability of transfer learning across language families and offer practical guidance for building machine translation systems in low-resource settings with minimal tuning effort.

[162] Echoes in AI: Quantifying lack of plot diversity in LLM outputs

Weijia Xu, Nebojsa Jojic, Sudha Rao, Chris Brockett, Bill Dolan

Main category: cs.CL

TL;DR: LLMs generate stories with repetitive plot elements across generations, lacking diversity compared to human creativity. A new metric called Sui Generis score measures plot uniqueness automatically.

DetailsMotivation: To investigate whether current large language models can provide diverse enough ideas to support collective creativity, particularly in story generation where originality is crucial.

Method: Evaluated GPT-4 and LLaMA-3 on story generation, introduced Sui Generis score to measure plot element uniqueness, analyzed 100 short stories, and conducted human evaluation comparing automatic scores with human judgment of surprise.

Result: LLM-generated stories contain frequently echoed plot elements across generations and different models, while human-written stories show rare plot recreation. Sui Generis scores moderately correlate with human surprise judgments.

Conclusion: Current LLMs struggle with generating truly diverse creative content, showing pattern repetition that limits their ability to enhance collective creativity compared to human originality.

Abstract: With rapid advances in large language models (LLMs), there has been an increasing application of LLMs in creative content ideation and generation. A critical question emerges: can current LLMs provide ideas that are diverse enough to truly bolster collective creativity? We examine two state-of-the-art LLMs, GPT-4 and LLaMA-3, on story generation and discover that LLM-generated stories often consist of plot elements that are echoed across a number of generations. To quantify this phenomenon, we introduce the Sui Generis score, an automatic metric that measures the uniqueness of a plot element among alternative storylines generated using the same prompt under an LLM. Evaluating on 100 short stories, we find that LLM-generated stories often contain combinations of idiosyncratic plot elements echoed frequently across generations and across different LLMs, while plots from the original human-written stories are rarely recreated or even echoed in pieces. Moreover, our human evaluation shows that the ranking of Sui Generis scores among story segments correlates moderately with human judgment of surprise level, even though score computation is completely automatic without relying on human judgment.

[163] Token Assorted: Mixing Latent and Text Tokens for Improved Language Model Reasoning

DiJia Su, Hanlin Zhu, Yingchen Xu, Jiantao Jiao, Yuandong Tian, Qinqing Zheng

Main category: cs.CL

TL;DR: Hybrid reasoning representation using latent discrete tokens from VQ-VAE to reduce reasoning trace length while maintaining performance.

DetailsMotivation: Traditional chain-of-thought reasoning produces lengthy inputs with many words for textual coherence rather than core reasoning, consuming substantial computation resources.

Method: Partially abstract initial reasoning steps using latent discrete tokens from VQ-VAE, train models with hybrid data mixing latent and text tokens, extend vocabulary with unseen latent tokens.

Result: Consistently outperforms baseline methods in various benchmarks including Keys-Finding Maze problem and logical/mathematical reasoning problems.

Conclusion: Hybrid representation with latent tokens effectively reduces reasoning trace length while maintaining or improving reasoning performance, with simple training procedure enabling fast adaptation.

Abstract: Large Language Models (LLMs) excel at reasoning and planning when trained on chainof-thought (CoT) data, where the step-by-step thought process is explicitly outlined by text tokens. However, this results in lengthy inputs where many words support textual coherence rather than core reasoning information, and processing these inputs consumes substantial computation resources. In this work, we propose a hybrid representation of the reasoning process, where we partially abstract away the initial reasoning steps using latent discrete tokens generated by VQ-VAE, significantly reducing the length of reasoning traces. We explore the use of latent trace abstractions in two scenarios: 1) training the model from scratch for the Keys-Finding Maze problem, 2) fine-tuning LLMs on this hybrid data with an extended vocabulary including unseen latent tokens, for both logical and mathematical reasoning problems. To facilitate effective learning, we introduce a simple training procedure that randomly mixes latent and text tokens, which enables fast adaptation to new latent tokens. Our approach consistently outperforms the baselines methods in various benchmarks.

[164] Injecting Domain-Specific Knowledge into Large Language Models: A Comprehensive Survey

Zirui Song, Bin Yan, Yuhan Liu, Miao Fang, Mingzhe Li, Rui Yan, Xiuying Chen

Main category: cs.CL

TL;DR: Survey paper categorizing methods to enhance LLMs with domain-specific knowledge through four approaches: dynamic injection, static embedding, modular adapters, and prompt optimization.

DetailsMotivation: General-purpose LLMs lack effectiveness in domain-specific applications requiring specialized knowledge like healthcare, chemistry, or legal analysis.

Method: Comprehensive overview and categorization of four key approaches: dynamic knowledge injection, static knowledge embedding, modular adapters, and prompt optimization.

Result: Provides analysis of trade-offs between flexibility, scalability, and efficiency for each method, along with evaluation of domain-specific LLMs vs general LLMs.

Conclusion: Highlights challenges and opportunities in specialized LLM field, maintains open-source repository for ongoing research documentation, and summarizes commonly used datasets and benchmarks.

Abstract: Large Language Models (LLMs) have demonstrated remarkable success in various tasks such as natural language understanding, text summarization, and machine translation. However, their general-purpose nature often limits their effectiveness in domain-specific applications that require specialized knowledge, such as healthcare, chemistry, or legal analysis. To address this, researchers have explored diverse methods to enhance LLMs by integrating domain-specific knowledge. In this survey, we provide a comprehensive overview of these methods, which we categorize into four key approaches: dynamic knowledge injection, static knowledge embedding, modular adapters, and prompt optimization. Each approach offers unique mechanisms to equip LLMs with domain expertise, balancing trade-offs between flexibility, scalability, and efficiency. We discuss how these methods enable LLMs to tackle specialized tasks, compare their advantages and disadvantages, evaluate domain-specific LLMs against general LLMs, and highlight the challenges and opportunities in this emerging field. For those interested in delving deeper into this area, we also summarize the commonly used datasets and benchmarks. To keep researchers updated on the latest studies, we maintain an open-source at: https://github.com/abilliyb/Knowledge_Injection_Survey_Papers, dedicated to documenting research in the field of specialized LLM.

[165] LogiDynamics: Unraveling the Dynamics of Logical Inference in Large Language Model Reasoning

Tianshi Zheng, Jiayang Cheng, Chunyang Li, Haochen Shi, Zihao Wang, Jiaxin Bai, Yangqiu Song, Ginny Y. Wong, Simon See

Main category: cs.CL

TL;DR: This paper explores how different logical inference paradigms (inductive, abductive, deductive) perform in LLMs across various dimensions of analogical reasoning tasks, and investigates advanced inference strategies for scaling up LLM reasoning capabilities.

DetailsMotivation: To understand how to optimally leverage different logical inference paradigms in large language models for reasoning tasks, particularly analogical reasoning as a fundamental cognitive task.

Method: Created a controlled evaluation environment for analogical reasoning parameterized across three dimensions: modality (textual, visual, symbolic), difficulty (easy, medium, hard), and task format (multiple-choice or free-text generation). Analyzed comparative dynamics of different inference pipelines and investigated advanced paradigms like hypothesis selection, verification, and refinement.

Result: The findings demonstrate that the comparative performance of different inference paradigms varies across the tested dimensions, and these findings generalize to broader in-context learning tasks. Advanced inference strategies show potential for scaling up logical inference in LLM reasoning.

Conclusion: This exploratory study provides a foundation for future research in enhancing LLM reasoning through systematic logical inference strategies, with resources made available for further development.

Abstract: Modern large language models (LLMs) employ various forms of logical inference, both implicitly and explicitly, when addressing reasoning tasks. Understanding how to optimally leverage these inference paradigms is critical for advancing LLMs’ reasoning capabilities. This paper adopts an exploratory approach by introducing a controlled evaluation environment for analogical reasoning – a fundamental cognitive task – that is systematically parameterized across three dimensions: modality (textual, visual, symbolic), difficulty (easy, medium, hard), and task format (multiple-choice or free-text generation). We analyze the comparative dynamics of inductive, abductive, and deductive inference pipelines across these dimensions, and demonstrate that our findings generalize to broader in-context learning tasks. Additionally, we investigate advanced paradigms such as hypothesis selection, verification, and refinement, revealing their potential to scale up logical inference in LLM reasoning. This exploratory study provides a foundation for future research in enhancing LLM reasoning through systematic logical inference strategies. Resources are available at https://github.com/HKUST-KnowComp/LogiDynamics.

[166] InsBank: Evolving Instruction Subset for Ongoing Alignment

Jiayi Shi, Yiwei Li, Shaoxiong Feng, Peiwen Yuan, Xinglin Wang, Yueqi Zhang, Chuyi Tan, Boyuan Pan, Huan Ren, Yao Hu, Kan Li

Main category: cs.CL

TL;DR: Instruction Bank (InsBank) with Progressive Instruction Bank Evolution (PIBE) framework for continuously evolving instruction data selection to maintain LLM alignment efficiently.

DetailsMotivation: Existing methods focus on selecting diverse, high-quality instruction data subsets but lack mechanisms to continuously evolve these subsets as new instruction data becomes available, creating a need for ongoing alignment of LLMs.

Method: Proposed PIBE framework with gradual data selection strategy using representation-based diversity scores to capture data relationships and maintain historical information, allowing flexible combination of diversity and quality metrics.

Result: Extensive experiments show PIBE significantly outperforms baselines in InsBank evolution and effectively extracts budget-specific subsets, demonstrating both effectiveness and adaptability.

Conclusion: The PIBE framework successfully enables continuous evolution of instruction data repositories, providing an efficient and adaptable solution for maintaining LLM alignment over time through progressive data selection strategies.

Abstract: Large language models (LLMs) typically undergo instruction tuning to enhance alignment. Recent studies emphasize that quality and diversity of instruction data are more crucial than quantity, highlighting the need to select diverse, high-quality subsets to reduce training costs. However, how to evolve these selected subsets alongside the development of new instruction data remains insufficiently explored. To achieve LLMs’ ongoing alignment, we introduce Instruction Bank (\textbf{InsBank}), a continuously updated repository that integrates the latest valuable instruction data. We further propose Progressive Instruction Bank Evolution (\textbf{PIBE}), a novel framework designed to evolve InsBank effectively and efficiently over time. PIBE employs a gradual data selection strategy to maintain long-term efficiency, leveraging a representation-based diversity score to capture relationships between data points and retain historical information for comprehensive diversity evaluation. This also allows for flexible combination of diversity and quality scores during data selection and ranking. Extensive experiments demonstrate that PIBE significantly outperforms baselines in InsBank evolution and is able to extract budget-specific subsets, demonstrating its effectiveness and adaptability.

[167] Instruction Tuning on Public Government and Cultural Data for Low-Resource Language: a Case Study in Kazakh

Nurkhan Laiyk, Daniil Orel, Rituraj Joshi, Maiya Goloburda, Yuxia Wang, Preslav Nakov, Fajri Koto

Main category: cs.CL

TL;DR: A large-scale instruction-following dataset for Kazakh language with 10,600 manually verified samples covering government and cultural knowledge, using LLM-assisted generation with GPT-4o backbone, showing consistent performance improvements when fine-tuning various models.

DetailsMotivation: Address the lack of instruction tuning resources for low-resource languages like Kazakh, particularly in government and cultural domains where text data is limited.

Method: LLM-assisted data generation using both open-weight and closed-weight models (selected GPT-4o as backbone), with full manual verification of each dataset entity. Fine-tuned Qwen, Falcon, and Gemma models on the created dataset.

Result: Consistent performance improvements in both multiple-choice and generative tasks across all fine-tuned models (Qwen, Falcon, Gemma), demonstrating effective LLM-assisted instruction tuning for low-resource languages.

Conclusion: The approach successfully creates high-quality instruction datasets for low-resource languages and shows that LLM-assisted instruction tuning can significantly enhance model performance in specialized domains like government and cultural knowledge.

Abstract: Instruction tuning in low-resource languages remains underexplored due to limited text data, particularly in government and cultural domains. To address this, we introduce and open-source a large-scale (10,600 samples) instruction-following (IFT) dataset, covering key institutional and cultural knowledge relevant to Kazakhstan. Our dataset enhances LLMs’ understanding of procedural, legal, and structural governance topics. We employ LLM-assisted data generation, comparing open-weight and closed-weight models for dataset construction, and select GPT-4o as the backbone. Each entity of our dataset undergoes full manual verification to ensure high quality. We also show that fine-tuning Qwen, Falcon, and Gemma on our dataset leads to consistent performance improvements in both multiple-choice and generative tasks, demonstrating the potential of LLM-assisted instruction tuning for low-resource languages.

[168] Multilingual != Multicultural: Evaluating Gaps Between Multilingual Capabilities and Cultural Alignment in LLMs

Jonathan Rystrøm, Hannah Rose Kirk, Scott Hale

Main category: cs.CL

TL;DR: LLMs show US-centric cultural bias despite multilingual capabilities. Study compares Gemma and OpenAI models against World Value Survey data across 4 languages, finding no consistent link between language proficiency and cultural alignment, with self-consistency being a better predictor.

DetailsMotivation: Large Language Models are becoming multilingual but may reflect US cultural values rather than local cultural representations, raising concerns about cultural bias in global applications.

Method: Used linear mixed-effects regression to compare LLM-generated response distributions against World Value Survey population-level opinion data across Danish, Dutch, English, and Portuguese languages, analyzing Google’s Gemma models (2B-27B) and OpenAI’s turbo-series.

Result: No consistent relationship between language capabilities and cultural alignment across model families. Gemma models showed positive correlation, while OpenAI models did not. Self-consistency was a stronger predictor of multicultural alignment than multilingual capabilities.

Conclusion: Achieving meaningful cultural alignment requires dedicated effort beyond improving general language capabilities, as multilingual proficiency alone does not ensure appropriate cultural representations.

Abstract: Large Language Models (LLMs) are becoming increasingly capable across global languages. However, the ability to communicate across languages does not necessarily translate to appropriate cultural representations. A key concern is US-centric bias, where LLMs reflect US rather than local cultural values. We propose a novel methodology that compares LLM-generated response distributions against population-level opinion data from the World Value Survey across four languages (Danish, Dutch, English, and Portuguese). Using a rigorous linear mixed-effects regression framework, we compare two families of models: Google’s Gemma models (2B–27B parameters) and successive iterations of OpenAI’s turbo-series. Across the families of models, we find no consistent relationships between language capabilities and cultural alignment. While the Gemma models have a positive correlation between language capability and cultural alignment across languages, the OpenAI models do not. Importantly, we find that self-consistency is a stronger predictor of multicultural alignment than multilingual capabilities. Our results demonstrate that achieving meaningful cultural alignment requires dedicated effort beyond improving general language capabilities.

[169] Automatic Input Rewriting Improves Translation with Large Language Models

Dayeon Ki, Marine Carpuat

Main category: cs.CL

TL;DR: LLM-assisted input rewriting, particularly text simplification, improves machine translation quality by making source text easier to translate while preserving original meaning.

DetailsMotivation: Users intuitively believe well-written text is easier to translate with off-the-shelf MT systems, but LLM capabilities have been primarily used for post-editing outputs rather than improving inputs.

Method: Empirical study of 21 input rewriting methods using 3 open-weight LLMs for English to 6 target languages translation, with text simplification as the primary strategy enhanced by quality estimation for translatability assessment.

Result: Text simplification was the most effective MT-agnostic rewrite strategy, and human evaluation confirmed that simplified rewrites and their MT outputs largely preserve original meaning of both source and translation.

Conclusion: LLM-assisted input rewriting represents a promising direction for improving translation quality, with text simplification showing particular effectiveness when combined with quality estimation techniques.

Abstract: Can we improve machine translation (MT) with LLMs by rewriting their inputs automatically? Users commonly rely on the intuition that well-written text is easier to translate when using off-the-shelf MT systems. LLMs can rewrite text in many ways but in the context of MT, these capabilities have been primarily exploited to rewrite outputs via post-editing. We present an empirical study of 21 input rewriting methods with 3 open-weight LLMs for translating from English into 6 target languages. We show that text simplification is the most effective MT-agnostic rewrite strategy and that it can be improved further when using quality estimation to assess translatability. Human evaluation further confirms that simplified rewrites and their MT outputs both largely preserve the original meaning of the source and MT. These results suggest LLM-assisted input rewriting as a promising direction for improving translations.

[170] A Causal Lens for Evaluating Faithfulness Metrics

Kerem Zaman, Shashank Srivastava

Main category: cs.CL

TL;DR: Causal Diagnosticity framework evaluates faithfulness metrics for LLM explanations using model-editing to create faithful/unfaithful explanation pairs across four tasks, finding Filler Tokens performs best but current metrics need improvement.

DetailsMotivation: LLM explanations may not reflect true model reasoning, and existing faithfulness metrics lack standardized evaluation, making comparisons difficult.

Method: Proposes Causal Diagnosticity framework using model-editing to generate faithful/unfaithful explanation pairs and evaluates metrics across fact-checking, analogy, object counting, and multi-hop reasoning tasks.

Result: Filler Tokens performed best overall, continuous metrics were more diagnostic than binary ones but sensitive to noise and model choice, with performance varying across tasks and models.

Conclusion: Current faithfulness metrics need improvement for robustness, highlighting the importance of standardized evaluation frameworks like Causal Diagnosticity.

Abstract: Large Language Models (LLMs) offer natural language explanations as an alternative to feature attribution methods for model interpretability. However, despite their plausibility, they may not reflect the model’s true reasoning faithfully, which is crucial for understanding the model’s true decision-making processes. Although several faithfulness metrics have been proposed, they are often evaluated in isolation, making direct, principled comparisons between them difficult. Here, we present Causal Diagnosticity, a framework that serves as a common testbed to evaluate faithfulness metrics for natural language explanations. Our framework employs the concept of diagnosticity, and uses model-editing methods to generate faithful-unfaithful explanation pairs. Our benchmark includes four tasks: fact-checking, analogy, object counting, and multi-hop reasoning. We evaluate prominent faithfulness metrics, including post-hoc explanation and chain-of-thought-based methods. We find that diagnostic performance varies across tasks and models, with Filler Tokens performing best overall. Additionally, continuous metrics are generally more diagnostic than binary ones but can be sensitive to noise and model choice. Our results highlight the need for more robust faithfulness metrics.

[171] Personalized Causal Graph Reasoning for LLMs: An Implementation for Dietary Recommendations

Zhongqi Yang, Amir Rahmani

Main category: cs.CL

TL;DR: A framework that enables LLMs to perform personalized reasoning using individual-specific causal graphs from longitudinal data, particularly effective for personalized dietary recommendations and glucose control.

DetailsMotivation: LLMs lack personalized reasoning capabilities for individual-specific data, limiting their use in domains like healthcare where decisions must adapt to personal contexts and multifactorial data.

Method: Personalized Causal Graph Reasoning framework where LLMs construct individual-specific causal graphs from longitudinal data, traverse these graphs to identify relevant causal pathways, rank them by impact, simulate outcomes, and generate tailored responses.

Result: The method reduces postprandial glucose iAUC across three time windows compared to prior approaches, with LLM-as-a-judge evaluations confirming improvements in personalization quality for nutrient-oriented dietary recommendations.

Conclusion: The framework successfully enables LLMs to perform personalized reasoning over individual-specific data, demonstrating significant improvements in healthcare applications like personalized dietary recommendations and glucose management.

Abstract: Large Language Models (LLMs) excel at general-purpose reasoning by leveraging broad commonsense knowledge, but they remain limited in tasks requiring personalized reasoning over multifactorial personal data. This limitation constrains their applicability in domains such as healthcare, where decisions must adapt to individual contexts. We introduce Personalized Causal Graph Reasoning, a framework that enables LLMs to reason over individual-specific causal graphs constructed from longitudinal data. Each graph encodes how user-specific factors influence targeted outcomes. In response to a query, the LLM traverses the graph to identify relevant causal pathways, rank them by estimated impact, simulate potential outcomes, and generate tailored responses. We implement this framework in the context of nutrient-oriented dietary recommendations, where variability in metabolic responses demands personalized reasoning. Using counterfactual evaluation, we assess the effectiveness of LLM-generated food suggestions for glucose control. Our method reduces postprandial glucose iAUC across three time windows compared to prior approaches. Additional LLM-as-a-judge evaluations further confirm improvements in personalization quality.

[172] Interpretation Gaps in LLM-Assisted Comprehension of Privacy Documents

Rinku Dewri

Main category: cs.CL

TL;DR: LLMs show promise for simplifying privacy policies but face accuracy, completeness, clarity and representation gaps that need further research.

DetailsMotivation: To explore the limitations and gaps when using large language models to simplify complex privacy policies and data practices.

Method: The article examines and exemplifies various gaps that manifest when LLMs interpret privacy policies, focusing on specific case studies or examples.

Result: Identified significant gaps in accuracy, completeness, clarity and representation when LLMs simplify privacy policies, showing current limitations.

Conclusion: While LLMs have potential to revolutionize privacy management through personal assistants and compliance automation, continued research is needed to address the identified gaps and realize their full potential.

Abstract: This article explores the gaps that can manifest when using a large language model (LLM) to obtain simplified interpretations of data practices from a complex privacy policy. We exemplify these gaps to showcase issues in accuracy, completeness, clarity and representation, while advocating for continued research to realize an LLM’s true potential in revolutionizing privacy management through personal assistants and automated compliance checking.

[173] General Table Question Answering via Answer-Formula Joint Generation

Zhongyuan Wang, Richong Zhang, Zhijie Nie, Hangyu Mao

Main category: cs.CL

TL;DR: This paper proposes TabAF, a table question answering framework that uses spreadsheet formulas as executable representations, achieving state-of-the-art performance on multiple benchmarks.

DetailsMotivation: Existing TableQA methods lack versatility for different question types and table structures, while spreadsheet formulas - a well-defined operation language for tabular data - remain unexplored for this task.

Method: Construct FormulaQA dataset with formula annotations from existing datasets, and develop TabAF framework that decodes both answers and formulas using a single LLM backbone to handle multiple task types and table structures.

Result: TabAF achieves new state-of-the-art performance on WikiTableQuestion, HiTab, and TabFact benchmarks under the same model size, demonstrating strong versatility and generalization.

Conclusion: Using spreadsheet formulas as executable representations provides an effective and versatile approach for complex table reasoning tasks across different table structures and question types.

Abstract: Advanced table question answering (TableQA) methods prompt large language models (LLMs) to generate answer text, SQL query, Python code, or custom operation, which impressively improve the complex reasoning problems in the TableQA task. However, these methods lack the versatility to cope with specific question types or table structures. In contrast, the Spreadsheet Formula, the widely used and well-defined operation language for tabular data, has not been thoroughly explored to solve TableQA. In this paper, we first attempt to use the Formula as the executable representation for solving complex reasoning on tables with different structures. Specifically, we construct \texttt{FromulaQA}, a large Formula-annotated TableQA dataset from existing datasets. In addition, we propose \texttt{TabAF}, a general table answering framework to solve multiple types of tasks over multiple types of tables simultaneously, which decodes answers and Formulas with a single LLM backbone. Extensive experiments demonstrate the versatility and generalization of \texttt{TabAF}. Under the same model size, \texttt{TabAF} achieves new state-of-the-art performance on the WikiTableQuestion, HiTab, and TabFact.

[174] UniBERT: Adversarial Training for Language-Universal Representations

Andrei-Marius Avram, Marian Lupaşcu, Dumitru-Clementin Cercel, Ionuţ Mironică, Ştefan Trăuşan-Matu

Main category: cs.CL

TL;DR: UniBERT is a compact multilingual language model that combines masked language modeling, adversarial training, and knowledge distillation to achieve better cross-lingual generalization with reduced computational demands.

DetailsMotivation: To address the computational demands of large-scale multilingual models while maintaining competitive performance across various NLP tasks, and to improve cross-lingual generalization capabilities.

Method: Uses an innovative training framework integrating three components: masked language modeling, adversarial training, and knowledge distillation. Pre-trained on a curated Wikipedia corpus spanning 107 languages.

Result: Achieves 7.72% average relative improvement over traditional baselines (which only achieved 1.17%) across four NLP tasks (named entity recognition, natural language inference, question answering, semantic textual similarity). Statistical significance confirmed with p-value = 0.0181.

Conclusion: The combination of adversarial training and knowledge distillation effectively builds scalable and robust multilingual language models, advancing cross-lingual NLP capabilities while reducing computational requirements.

Abstract: This paper presents UniBERT, a compact multilingual language model that uses an innovative training framework that integrates three components: masked language modeling, adversarial training, and knowledge distillation. Pre-trained on a meticulously curated Wikipedia corpus spanning 107 languages, UniBERT is designed to reduce the computational demands of large-scale models while maintaining competitive performance across various natural language processing tasks. Comprehensive evaluations on four tasks - named entity recognition, natural language inference, question answering, and semantic textual similarity - demonstrate that our multilingual training strategy enhanced by an adversarial objective significantly improves cross-lingual generalization. Specifically, UniBERT models show an average relative improvement of 7.72% over traditional baselines, which achieved an average relative improvement of only 1.17%, and statistical analysis confirms the significance of these gains (p-value = 0.0181). This work highlights the benefits of combining adversarial training and knowledge distillation to build scalable and robust language models, thus advancing the field of multilingual and cross-lingual natural language processing.

[175] Unmasking Deceptive Visuals: Benchmarking Multimodal Large Language Models on Misleading Chart Question Answering

Zixin Chen, Sicheng Song, Kashun Shum, Yanna Lin, Rui Sheng, Huamin Qu

Main category: cs.CL

TL;DR: A new benchmark dataset (Misleading ChartQA) with 3,026 examples evaluates multimodal LLMs’ ability to detect misleading visualizations across 21 misleader types and 10 chart formats, revealing current limitations and proposing an improved reasoning pipeline.

DetailsMotivation: Misleading visualizations remain widespread despite decades of research, posing risks to public understanding and AI safety. While MLLMs show strong chart comprehension, their capacity to detect misleading charts is unexplored.

Method: Created Misleading ChartQA benchmark with 3,026 curated examples spanning 21 misleader types and 10 chart types, each with standardized chart code, CSV data, multiple-choice questions, and labeled explanations. Benchmarked 24 state-of-the-art MLLMs and proposed a novel region-aware reasoning pipeline.

Result: The study benchmarks 24 MLLMs and analyzes their performance across different misleader types and chart formats, revealing current limitations in detecting misleading visualizations.

Conclusion: This work lays the foundation for developing MLLMs that are robust, trustworthy, and aligned with responsible visual communication demands, with the proposed region-aware reasoning pipeline showing enhanced accuracy.

Abstract: Misleading visualizations, which manipulate chart representations to support specific claims, can distort perception and lead to incorrect conclusions. Despite decades of research, they remain a widespread issue-posing risks to public understanding and raising safety concerns for AI systems involved in data-driven communication. While recent multimodal large language models (MLLMs) show strong chart comprehension abilities, their capacity to detect and interpret misleading charts remains unexplored. We introduce Misleading ChartQA benchmark, a large-scale multimodal dataset designed to evaluate MLLMs on misleading chart reasoning. It contains 3,026 curated examples spanning 21 misleader types and 10 chart types, each with standardized chart code, CSV data, multiple-choice questions, and labeled explanations, validated through iterative MLLM checks and exhausted expert human review. We benchmark 24 state-of-the-art MLLMs, analyze their performance across misleader types and chart formats, and propose a novel region-aware reasoning pipeline that enhances model accuracy. Our work lays the foundation for developing MLLMs that are robust, trustworthy, and aligned with the demands of responsible visual communication.

[176] CoRanking: Collaborative Ranking with Small and Large Ranking Agents

Wenhan Liu, Xinyu Ma, Yutao Zhu, Lixin Su, Shuaiqiang Wang, Dawei Yin, Zhicheng Dou

Main category: cs.CL

TL;DR: CoRanking is a collaborative ranking framework that combines small and large models for efficient document ranking, using a small reranker for pre-ranking and LLM for final ranking with order adjustment to mitigate positional bias.

DetailsMotivation: LLMs show superior ranking performance but suffer from efficiency issues due to large parameters and sliding window processes. There's a need for efficient ranking that maintains effectiveness while reducing computational costs.

Method: 1) Use small reranker to pre-rank and select top passages 2) Apply reinforcement learning-trained order adjuster to reorder passages for LLM preference 3) Use LLM listwise reranker only on top passages instead of full list

Result: 70% reduction in ranking latency while achieving better effectiveness compared to using only LLM listwise reranker across three IR benchmarks

Conclusion: CoRanking successfully addresses efficiency challenges in LLM-based ranking by combining small and large models with order adjustment, achieving both improved efficiency and effectiveness.

Abstract: Large Language Models (LLMs) have demonstrated superior listwise ranking performance. However, their superior performance often relies on large-scale parameters (\eg, GPT-4) and a repetitive sliding window process, which introduces significant efficiency challenges. In this paper, we propose \textbf{CoRanking}, a novel collaborative ranking framework that combines small and large ranking models for efficient and effective ranking. CoRanking first employs a small-size reranker to pre-rank all the candidate passages, bringing relevant ones to the top part of the list (\eg, top-20). Then, the LLM listwise reranker is applied to only rerank these top-ranked passages instead of the whole list, substantially enhancing overall ranking efficiency. Although more efficient, previous studies have revealed that the LLM listwise reranker have significant positional biases on the order of input passages. Directly feed the top-ranked passages from small reranker may result in the sub-optimal performance of LLM listwise reranker. To alleviate this problem, we introduce a passage order adjuster trained via reinforcement learning, which reorders the top passages from the small reranker to align with the LLM’s preferences of passage order. Extensive experiments on three IR benchmarks demonstrate that CoRanking significantly improves efficiency (reducing ranking latency by about 70%) while achieving even better effectiveness compared to using only the LLM listwise reranker.

[177] AIR: A Systematic Analysis of Annotations, Instructions, and Response Pairs in Preference Dataset

Bingxiang He, Wenbin Zhang, Jiaxi Song, Cheng Qian, Zixuan Fu, Bowen Sun, Ning Ding, Haiwen Hong, Longtao Huang, Hui Xue, Ganqu Cui, Wanxiang Che, Zhiyuan Liu, Maosong Sun

Main category: cs.CL

TL;DR: AIR framework isolates and optimizes the three core components of preference learning datasets (Annotations, Instructions, Response Pairs) to achieve +5.3 average gains with only 14k high-quality pairs.

DetailsMotivation: Current approaches conflate the three core components of preference datasets, obscuring their individual impacts and hindering systematic optimization of preference learning for LLM alignment.

Method: Proposes AIR framework that systematically isolates and optimizes each component (Annotations, Instructions, Response Pairs) while evaluating their synergistic effects through rigorous experimentation.

Result: Achieves +5.3 average gains over baseline methods with only 14k high-quality pairs, revealing actionable principles: annotation simplicity, instruction inference stability, and response pair quality optimization.

Conclusion: Shifts preference dataset design from ad hoc scaling to component-aware optimization, offering a blueprint for efficient and reproducible LLM alignment.

Abstract: Preference learning is critical for aligning large language models (LLMs) with human values, yet its success hinges on high-quality datasets comprising three core components: Preference \textbf{A}nnotations, \textbf{I}nstructions, and \textbf{R}esponse Pairs. Current approaches conflate these components, obscuring their individual impacts and hindering systematic optimization. In this work, we propose \textbf{AIR}, a component-wise analysis framework that systematically isolates and optimizes each component while evaluating their synergistic effects. Through rigorous experimentation, AIR reveals actionable principles: annotation simplicity (point-wise generative scoring), instruction inference stability (variance-based filtering across LLMs), and response pair quality (moderate margins + high absolute scores). When combined, these principles yield +5.3 average gains over baseline method, even with only 14k high-quality pairs. Our work shifts preference dataset design from ad hoc scaling to component-aware optimization, offering a blueprint for efficient, reproducible alignment.

[178] SaRoHead: Detecting Satire in a Multi-Domain Romanian News Headline Dataset

Mihnea-Alexandru Vîrlan, Răzvan-Alexandru Smădu, Dumitru-Clementin Cercel, Florin Pop, Mihaela-Claudia Cercel

Main category: cs.CL

TL;DR: The paper investigates detecting satirical tone in Romanian news headlines alone using various ML approaches, finding that Bidirectional Transformer models with meta-learning outperform other methods.

DetailsMotivation: Current approaches for Romanian language satire detection require both headline and main article content, but headlines alone should reflect the satirical tone and serve as sufficient indicators.

Method: Tested multiple baselines including standard machine learning algorithms, deep learning models, and Large Language Models (LLMs), with a focus on Bidirectional Transformer models using meta-learning Reptile approach.

Result: Bidirectional Transformer models outperformed both standard machine-learning approaches and LLMs, particularly when the meta-learning Reptile approach was employed.

Conclusion: Headlines alone contain sufficient signals for satirical tone detection in Romanian news, and advanced transformer models with meta-learning provide the best performance for this task.

Abstract: The primary goal of a news headline is to summarize an event in as few words as possible. Depending on the media outlet, a headline can serve as a means to objectively deliver a summary or improve its visibility. For the latter, specific publications may employ stylistic approaches that incorporate the use of sarcasm, irony, and exaggeration, key elements of a satirical approach. As such, even the headline must reflect the tone of the satirical main content. Current approaches for the Romanian language tend to detect the non-conventional tone (i.e., satire and clickbait) of the news content by combining both the main article and the headline. Because we consider a headline to be merely a brief summary of the main article, we investigate in this paper the presence of satirical tone in headlines alone, testing multiple baselines ranging from standard machine learning algorithms to deep learning models. Our experiments show that Bidirectional Transformer models outperform both standard machine-learning approaches and Large Language Models (LLMs), particularly when the meta-learning Reptile approach is employed.

[179] RankAlign: A Ranking View of the Generator-Validator Gap in Large Language Models

Juan Diego Rodriguez, Wenxuan Ding, Katrin Erk, Greg Durrett

Main category: cs.CL

TL;DR: The paper identifies a generator-validator gap in LLMs where models generate answers but fail to verify them consistently. It proposes RankAlign, a ranking-based training method that significantly reduces this gap and improves generalization.

DetailsMotivation: LLMs remain unreliable due to inconsistency in reporting the same information across different prompts, particularly the discrepancy between generated answers and the model's own verification of those answers.

Method: The authors define the generator-validator gap more stringently than prior work, requiring correlation over all candidate answers. They propose RankAlign, a ranking-based training method to align generator and validator outputs.

Result: A large generator-validator gap exists across various tasks (question answering, lexical semantics, next-word prediction). RankAlign significantly closes this gap and outperforms all baseline methods while generalizing well to out-of-domain tasks.

Conclusion: The generator-validator gap is a fundamental limitation in LLMs, but RankAlign provides an effective training approach to reduce this inconsistency and improve model reliability across diverse domains.

Abstract: Although large language models (LLMs) have become more capable and accurate across many tasks, some fundamental sources of unreliability remain in their behavior. One key limitation is their inconsistency at reporting the same information when prompts are changed. In this paper, we consider the discrepancy between a model’s generated answer and their own verification of that answer, the generator-validator gap. We define this gap in a more stringent way than prior work: we expect correlation of scores from a generator and a validator over the entire set of candidate answers, i.e., candidate completions that could possibly arise during ordinary language use without breaking Gricean norms. We show that according to this measure, a large gap exists in various settings, including question answering, lexical semantics tasks, and next-word prediction. We then propose RankAlign, a ranking-based training method, and show that it significantly closes the gap, surpassing all baseline methods. Moreover, this approach generalizes well to out-of-domain tasks and lexical items.

[180] AskQE: Question Answering as Automatic Evaluation for Machine Translation

Dayeon Ki, Kevin Duh, Marine Carpuat

Main category: cs.CL

TL;DR: AskQE is a question generation and answering framework that helps monolingual English speakers evaluate French machine translations without knowing French, using contrastive error detection and achieving better correlation with human ratings than existing QE metrics.

DetailsMotivation: Existing MT error detection and quality estimation techniques don't address the practical scenario where monolingual users need to evaluate translations in languages they don't understand, creating a need for accessible quality assessment tools.

Method: Uses question generation and answering framework with LLaMA-3 70B model and entailed facts to detect critical MT errors. Developed using ContraTICO dataset of synthetic MT errors in COVID-19 domain and evaluated on BioMQM dataset of natural MT errors.

Result: AskQE achieves higher Kendall’s Tau correlation and decision accuracy with human ratings compared to other quality estimation metrics when tested on the BioMQM dataset.

Conclusion: The AskQE framework successfully enables monolingual users to assess translation quality without target language knowledge, outperforming existing QE approaches and providing actionable feedback for translation acceptance decisions.

Abstract: How can a monolingual English speaker determine whether an automatic translation in French is good enough to be shared? Existing MT error detection and quality estimation (QE) techniques do not address this practical scenario. We introduce AskQE, a question generation and answering framework designed to detect critical MT errors and provide actionable feedback, helping users decide whether to accept or reject MT outputs even without the knowledge of the target language. Using ContraTICO, a dataset of contrastive synthetic MT errors in the COVID-19 domain, we explore design choices for AskQE and develop an optimized version relying on LLaMA-3 70B and entailed facts to guide question generation. We evaluate the resulting system on the BioMQM dataset of naturally occurring MT errors, where AskQE has higher Kendall’s Tau correlation and decision accuracy with human ratings compared to other QE metrics.

[181] Deep Binding of Language Model Virtual Personas: a Study on Approximating Political Partisan Misperceptions

Minwoo Kang, Suhong Moon, Seung Hyeong Lee, Ayush Raj, Joseph Suh, David M. Chan, John Canny

Main category: cs.CL

TL;DR: LLMs can simulate human survey responses but it’s unclear if they provide deep in-group perspectives or shallow out-group assumptions. The paper proposes using narrative identity theory to create detailed synthetic personas that significantly improve LLM fidelity in replicating human response patterns for in-group/out-group bias studies.

DetailsMotivation: To determine whether LLMs provide authentic in-group perspectives or just out-group assumptions when simulating human survey responses, which is critical for political science applications like polarization studies and inter-group conflict analysis.

Method: Proposes a novel methodology using narrative identity theory to create virtual personas with detailed synthetic backstories generated as multi-turn interview transcripts, producing longer, richer, and more consistent individual descriptions than previous methods.

Result: Virtual personas conditioned on the generated backstories closely replicate human response distributions (up to 87% improvement in Wasserstein Distance) and produce effect sizes that match original studies of in-group/out-group biases.

Conclusion: The work extends LLM applicability beyond estimating socially understood responses, enabling their use in a broader range of human studies requiring authentic in-group perspectives.

Abstract: Large language models (LLMs) are increasingly capable of simulating human behavior, offering cost-effective ways to estimate user responses to various surveys and polls. However, the questions in these surveys usually reflect socially understood attitudes: the patterns of attitudes of old/young, liberal/conservative, as understood by both members and non-members of those groups. It is not clear whether the LLM binding is \emph{deep}, meaning the LLM answers as a member of a particular in-group would, or \emph{shallow}, meaning the LLM responds as an out-group member believes an in-group member would. To explore this difference, we use questions that expose known in-group/out-group biases. This level of fidelity is critical for applying LLMs to various political science studies, including timely topics on polarization dynamics, inter-group conflict, and democratic backsliding. To this end, we propose a novel methodology for constructing virtual personas with synthetic user “backstories” generated as extended, multi-turn interview transcripts. This approach is justified by the theory of \emph{narrative identity} which argues that personality at the highest level is \emph{constructed} from self-narratives. Our generated backstories are longer, rich in detail, and consistent in authentically describing a singular individual, compared to previous methods. We show that virtual personas conditioned on our backstories closely replicate human response distributions (up to an 87% improvement as measured by Wasserstein Distance) and produce effect sizes that closely match those observed in the original studies of in-group/out-group biases. Altogether, our work extends the applicability of LLMs beyond estimating socially understood responses, enabling their use in a broader range of human studies.

[182] Bias Analysis and Mitigation through Protected Attribute Detection and Regard Classification

Takuma Udagawa, Yang Zhao, Hiroshi Kanayama, Bishwaranjan Bhattacharjee

Main category: cs.CL

TL;DR: Proposed efficient annotation pipeline to detect social biases in LLM pretraining corpora through protected attribute detection and regard classification, with experiments on Common Crawl.

DetailsMotivation: Pretraining data from web-crawled texts contain undesirable social biases that can be perpetuated or amplified by large language models, requiring systematic analysis and mitigation.

Method: Developed an annotation pipeline with two stages: protected attribute detection to identify diverse demographics, followed by regard classification to analyze language polarity towards each attribute.

Result: Demonstrated the effectiveness of the bias analysis and mitigation measures through experiments on Common Crawl as a representative pretraining corpus.

Conclusion: The proposed pipeline provides an efficient and effective way to investigate and address social biases in LLM pretraining data, helping to reduce bias propagation in language models.

Abstract: Large language models (LLMs) acquire general linguistic knowledge from massive-scale pretraining. However, pretraining data mainly comprised of web-crawled texts contain undesirable social biases which can be perpetuated or even amplified by LLMs. In this study, we propose an efficient yet effective annotation pipeline to investigate social biases in the pretraining corpora. Our pipeline consists of protected attribute detection to identify diverse demographics, followed by regard classification to analyze the language polarity towards each attribute. Through our experiments, we demonstrate the effect of our bias analysis and mitigation measures, focusing on Common Crawl as the most representative pretraining corpus.

[183] Improving Informally Romanized Language Identification

Adrian Benton, Alexander Gutkin, Christo Kirov, Brian Roark

Main category: cs.CL

TL;DR: Improving language identification for romanized text by using synthetic training data with natural spelling variations, achieving state-of-the-art results on 20 Indic languages.

DetailsMotivation: Romanized text from languages with non-Latin scripts has high spelling variability, making languages that are normally distinct (like Hindi and Urdu) highly confusable when written in Latin script.

Method: Developed improved methods to synthesize training sets that incorporate natural spelling variation, using synthetic samples to train language identification systems rather than relying solely on naturally occurring examples.

Result: Achieved new state-of-the-art performance on Bhasha-Abhijnaanam evaluation set: improved F1 from 74.7% to 85.4% with synthetic data alone, and 88.2% when combined with harvested text.

Conclusion: Training on synthetic samples with natural spelling variation yields higher language identification accuracy than using naturally occurring examples or higher capacity models, demonstrating effective approach for romanized text LID.

Abstract: The Latin script is often used to informally write languages with non-Latin native scripts. In many cases (e.g., most languages in India), the lack of conventional spelling in the Latin script results in high spelling variability. Such romanization renders languages that are normally easily distinguished due to being written in different scripts - Hindi and Urdu, for example - highly confusable. In this work, we increase language identification (LID) accuracy for romanized text by improving the methods used to synthesize training sets. We find that training on synthetic samples which incorporate natural spelling variation yields higher LID system accuracy than including available naturally occurring examples in the training set, or even training higher capacity models. We demonstrate new state-of-the-art LID performance on romanized text from 20 Indic languages in the Bhasha-Abhijnaanam evaluation set (Madhani et al., 2023a), improving test F1 from the reported 74.7% (using a pretrained neural model) to 85.4% using a linear classifier trained solely on synthetic data and 88.2% when also training on available harvested text.

[184] A Survey on Progress in LLM Alignment from the Perspective of Reward Design

Miaomiao Ji, Yanqiu Wu, Zhibin Wu, Shoujin Wang, Jian Yang, Mark Dras, Usman Naseem

Main category: cs.CL

TL;DR: A comprehensive survey on reward design for aligning LLMs with human values, covering mathematical formulation, construction practices, and optimization interactions, with a taxonomy of reward mechanisms.

DetailsMotivation: Reward design is crucial for bridging human feedback signals with LLM optimization to align models with human values, requiring systematic organization and practical guidance.

Method: Develops a structured taxonomy of reward mechanisms along complementary dimensions, analyzing mathematical formulations, construction practices, and interactions with optimization paradigms including RL-based and RL-free approaches.

Result: Provides conceptual clarity and practical guidance for alignment research, characterizing the progression from RL-based to RL-free optimization and from single-task to multi-objective complex settings.

Conclusion: LLM alignment advancement depends on continuous refinement of reward design strategies, with recent paradigm shifts toward more sophisticated multi-objective and complex optimization approaches.

Abstract: Reward design plays a pivotal role in aligning large language models (LLMs) with human values, serving as the bridge between feedback signals and model optimization. This survey provides a structured organization of reward modeling and addresses three key aspects: mathematical formulation, construction practices, and interaction with optimization paradigms. Building on this, it develops a macro-level taxonomy that characterizes reward mechanisms along complementary dimensions, thereby offering both conceptual clarity and practical guidance for alignment research. The progression of LLM alignment can be understood as a continuous refinement of reward design strategies, with recent developments highlighting paradigm shifts from reinforcement learning (RL)-based to RL-free optimization and from single-task to multi-objective and complex settings.

[185] Search and Refine During Think: Autonomous Retrieval-Augmented Reasoning of LLMs

Yaorui Shi, Sihang Li, Chang Wu, Zhiyuan Liu, Junfeng Fang, Hengxing Cai, An Zhang, Xiang Wang

Main category: cs.CL

TL;DR: AutoRefine is a reinforcement learning framework that improves retrieval-augmented reasoning by adding knowledge refinement steps between search calls, enabling better evidence filtering and organization.

DetailsMotivation: Large language models are limited by their knowledge reservoir, and existing retrieval-augmented methods often retrieve irrelevant or noisy information that hinders accurate reasoning.

Method: Proposes AutoRefine framework using reinforcement learning with a “search-and-refine-during-think” paradigm, incorporating knowledge refinement steps between search calls and using group relative policy optimization with retrieval-specific rewards.

Result: Significantly outperforms existing approaches on single-hop and multi-hop QA benchmarks, particularly in complex reasoning scenarios, with frequent higher-quality searches and effective evidence synthesis.

Conclusion: AutoRefine demonstrates that explicit knowledge refinement steps combined with tailored retrieval rewards can substantially improve retrieval-augmented reasoning performance, especially for complex multi-hop tasks.

Abstract: Large language models have demonstrated impressive reasoning capabilities but are inherently limited by their knowledge reservoir. Retrieval-augmented reasoning mitigates this limitation by allowing LLMs to query external resources, but existing methods often retrieve irrelevant or noisy information, hindering accurate reasoning. In this paper, we propose AutoRefine, a reinforcement learning post-training framework that adopts a new ``search-and-refine-during-think’’ paradigm. AutoRefine introduces explicit knowledge refinement steps between successive search calls, enabling the model to iteratively filter, distill, and organize evidence before generating an answer. Furthermore, we incorporate tailored retrieval-specific rewards alongside answer correctness rewards using group relative policy optimization. Experiments on single-hop and multi-hop QA benchmarks demonstrate that AutoRefine significantly outperforms existing approaches, particularly in complex, multi-hop reasoning scenarios. Detailed analysis shows that AutoRefine issues frequent, higher-quality searches and synthesizes evidence effectively.

[186] When Thinking Fails: The Pitfalls of Reasoning for Instruction-Following in LLMs

Xiaomin Li, Zhou Yu, Zhiwei Zhang, Xupeng Chen, Ziji Zhang, Yingying Zhuang, Narayanan Sadagopan, Anurag Beniwal

Main category: cs.CL

TL;DR: CoT reasoning can degrade instruction-following accuracy in LLMs, with attention diverted from key constraints. Selective reasoning strategies help recover performance.

DetailsMotivation: To uncover the surprising phenomenon that explicit chain-of-thought reasoning can significantly degrade instruction-following accuracy in large language models, despite its success in complex reasoning tasks.

Method: Evaluated 15 models on IFEval and ComplexBench benchmarks, conducted large-scale case studies and attention-based analysis, proposed constraint attention metric, and introduced four mitigation strategies: in-context learning, self-reflection, self-selective reasoning, and classifier-selective reasoning.

Result: Consistent performance drops when CoT prompting is applied, with reasoning diverting attention away from instruction-relevant tokens. Selective reasoning strategies, particularly classifier-selective reasoning, substantially recover lost performance.

Conclusion: This is the first systematic work exposing reasoning-induced failures in instruction-following and provides practical mitigation strategies, showing that selective reasoning approaches can effectively address the performance degradation caused by CoT reasoning.

Abstract: Reasoning-enhanced large language models (RLLMs), whether explicitly trained for reasoning or prompted via chain-of-thought (CoT), have achieved state-of-the-art performance on many complex reasoning tasks. However, we uncover a surprising and previously overlooked phenomenon: explicit CoT reasoning can significantly degrade instruction-following accuracy. Evaluating 15 models on two benchmarks: IFEval (with simple, rule-verifiable constraints) and ComplexBench (with complex, compositional constraints), we consistently observe performance drops when CoT prompting is applied. Through large-scale case studies and an attention-based analysis, we identify common patterns where reasoning either helps (e.g., with formatting or lexical precision) or hurts (e.g., by neglecting simple constraints or introducing unnecessary content). We propose a metric, constraint attention, to quantify model focus during generation and show that CoT reasoning often diverts attention away from instruction-relevant tokens. To mitigate these effects, we introduce and evaluate four strategies: in-context learning, self-reflection, self-selective reasoning, and classifier-selective reasoning. Our results demonstrate that selective reasoning strategies, particularly classifier-selective reasoning, can substantially recover lost performance. To our knowledge, this is the first work to systematically expose reasoning-induced failures in instruction-following and offer practical mitigation strategies.

[187] From Automation to Autonomy: A Survey on Large Language Models in Scientific Discovery

Tianshi Zheng, Zheye Deng, Hong Ting Tsang, Weiqi Wang, Jiaxin Bai, Zihao Wang, Yangqiu Song

Main category: cs.CL

TL;DR: This survey paper analyzes how Large Language Models are transforming scientific discovery from task-specific tools into autonomous agents, proposing a three-level taxonomy (Tool, Analyst, Scientist) to categorize their evolving roles and capabilities.

DetailsMotivation: To systematically chart the paradigm shift where LLMs are evolving from automation tools into autonomous agents that fundamentally redefine scientific research processes and human-AI collaboration.

Method: The authors introduce a three-level taxonomy (Tool, Analyst, Scientist) through the lens of the scientific method to delineate LLMs’ escalating autonomy and responsibilities in the research lifecycle.

Result: The survey provides a conceptual architecture for understanding LLM roles in science and identifies key challenges including robotic automation, self-improvement, and ethical governance.

Conclusion: This work offers strategic foresight to navigate and shape the future of AI-driven scientific discovery, aiming to foster both rapid innovation and responsible advancement in the field.

Abstract: Large Language Models (LLMs) are catalyzing a paradigm shift in scientific discovery, evolving from task-specific automation tools into increasingly autonomous agents and fundamentally redefining research processes and human-AI collaboration. This survey systematically charts this burgeoning field, placing a central focus on the changing roles and escalating capabilities of LLMs in science. Through the lens of the scientific method, we introduce a foundational three-level taxonomy-Tool, Analyst, and Scientist-to delineate their escalating autonomy and evolving responsibilities within the research lifecycle. We further identify pivotal challenges and future research trajectories such as robotic automation, self-improvement, and ethical governance. Overall, this survey provides a conceptual architecture and strategic foresight to navigate and shape the future of AI-driven scientific discovery, fostering both rapid innovation and responsible advancement. Github Repository: https://github.com/HKUST-KnowComp/Awesome-LLM-Scientific-Discovery.

[188] MCIP: Protecting MCP Safety via Model Contextual Integrity Protocol

Huihao Jing, Haoran Li, Wenbin Hu, Qi Hu, Heli Xu, Tianshu Chu, Peizhao Hu, Yangqiu Song

Main category: cs.CL

TL;DR: This paper addresses safety risks in Model Context Protocol (MCP) by proposing MCIP, a refined version with enhanced safety mechanisms, developing a taxonomy of unsafe behaviors, and creating benchmark data that improves LLMs’ safety performance in MCP interactions.

DetailsMotivation: MCP's decentralized architecture introduces underexplored safety risks that require systematic analysis and mitigation to ensure safe interactions between clients and servers.

Method: The authors use the MAESTRO framework to analyze MCP’s safety gaps, propose MCIP as a refined protocol, develop a fine-grained taxonomy of unsafe behaviors, create benchmark/training data, and conduct experiments on state-of-the-art LLMs.

Result: Experiments reveal LLMs’ vulnerabilities in MCP interactions and demonstrate that the proposed approach substantially improves their safety performance.

Conclusion: The paper successfully identifies and addresses MCP safety risks through a comprehensive framework that includes protocol refinement, taxonomy development, and benchmark creation, leading to significant improvements in LLM safety capabilities.

Abstract: As Model Context Protocol (MCP) introduces an easy-to-use ecosystem for users and developers, it also brings underexplored safety risks. Its decentralized architecture, which separates clients and servers, poses unique challenges for systematic safety analysis. This paper proposes a novel framework to enhance MCP safety. Guided by the MAESTRO framework, we first analyze the missing safety mechanisms in MCP, and based on this analysis, we propose the Model Contextual Integrity Protocol (MCIP), a refined version of MCP that addresses these gaps. Next, we develop a fine-grained taxonomy that captures a diverse range of unsafe behaviors observed in MCP scenarios. Building on this taxonomy, we develop benchmark and training data that support the evaluation and improvement of LLMs’ capabilities in identifying safety risks within MCP interactions. Leveraging the proposed benchmark and training data, we conduct extensive experiments on state-of-the-art LLMs. The results highlight LLMs' vulnerabilities in MCP interactions and demonstrate that our approach substantially improves their safety performance.

[189] Can Large Language Models be Effective Online Opinion Miners?

Ryang Heo, Yongsik Seo, Junseong Lee, Dongha Lee

Main category: cs.CL

TL;DR: Introduces OOMB benchmark to evaluate LLMs’ opinion mining capabilities from diverse online content, providing annotated data and opinion summaries for comprehensive assessment.

DetailsMotivation: Traditional opinion mining approaches struggle with the highly diverse, complex, and context-rich nature of user-generated online content, requiring better evaluation methods for LLMs.

Method: Developed Online Opinion Mining Benchmark (OOMB) with extensive (entity, feature, opinion) tuple annotations and opinion-centric summaries to evaluate both extractive and abstractive capabilities of LLMs.

Result: The benchmark enables comprehensive analysis of challenging aspects and LLM adaptability in opinion mining, though specific performance metrics are not detailed in the abstract.

Conclusion: OOMB lays the foundation for LLM-based opinion mining and provides directions for future research in evaluating LLMs as effective opinion miners in realistic online scenarios.

Abstract: The surge of user-generated online content presents a wealth of insights into customer preferences and market trends. However, the highly diverse, complex, and context-rich nature of such contents poses significant challenges to traditional opinion mining approaches. To address this, we introduce Online Opinion Mining Benchmark (OOMB), a novel dataset and evaluation protocol designed to assess the ability of large language models (LLMs) to mine opinions effectively from diverse and intricate online environments. OOMB provides extensive (entity, feature, opinion) tuple annotations and a comprehensive opinion-centric summary that highlights key opinion topics within each content, thereby enabling the evaluation of both the extractive and abstractive capabilities of models. Through our proposed benchmark, we conduct a comprehensive analysis of which aspects remain challenging and where LLMs exhibit adaptability, to explore whether they can effectively serve as opinion miners in realistic online scenarios. This study lays the foundation for LLM-based opinion mining and discusses directions for future research in this field.

[190] Unveil Multi-Picture Descriptions for Multilingual Mild Cognitive Impairment Detection via Contrastive Learning

Kristin Qi, Jiali Cheng, Youxiang Zhu, Hadi Amiri, Xiaohui Liang

Main category: cs.CL

TL;DR: A framework for multilingual MCI detection using contrastive learning, image modality integration, and Product of Experts to handle multiple picture descriptions

DetailsMotivation: Prior work focused only on English speakers describing single pictures, but real-world MCI detection requires handling multilingual speakers and multiple pictures with picture-dependent content

Method: Three components: supervised contrastive learning for better representations, incorporating image modality alongside speech/text, and Product of Experts to reduce spurious correlations

Result: +7.1% UAR improvement (68.1% to 75.2%) and +2.9% F1 score improvement (80.6% to 83.5%) over text baseline; contrastive learning particularly benefits text modality

Conclusion: The framework effectively addresses challenges in multilingual and multi-picture MCI detection, demonstrating significant performance improvements

Abstract: Detecting Mild Cognitive Impairment from picture descriptions is critical yet challenging, especially in multilingual and multiple picture settings. Prior work has primarily focused on English speakers describing a single picture (e.g., the ‘Cookie Theft’). The TAUKDIAL-2024 challenge expands this scope by introducing multilingual speakers and multiple pictures, which presents new challenges in analyzing picture-dependent content. To address these challenges, we propose a framework with three components: (1) enhancing discriminative representation learning via supervised contrastive learning, (2) involving image modality rather than relying solely on speech and text modalities, and (3) applying a Product of Experts (PoE) strategy to mitigate spurious correlations and overfitting. Our framework improves MCI detection performance, achieving a +7.1% increase in Unweighted Average Recall (UAR) (from 68.1% to 75.2%) and a +2.9% increase in F1 score (from 80.6% to 83.5%) compared to the text unimodal baseline. Notably, the contrastive learning component yields greater gains for the text modality compared to speech. These results highlight our framework’s effectiveness in multilingual and multi-picture MCI detection.

[191] From Tokens to Thoughts: How LLMs and Humans Trade Compression for Meaning

Chen Shani, Liron Soffer, Dan Jurafsky, Yann LeCun, Ravid Shwartz-Ziv

Main category: cs.CL

TL;DR: LLMs form broad categories similar to humans but lack fine-grained semantic distinctions, showing a bias towards statistical compression over human-like adaptive nuance.

DetailsMotivation: To understand if LLMs' internal representations achieve a human-like trade-off between compression and semantic fidelity in knowledge organization.

Method: Used information-theoretic framework based on Rate-Distortion Theory and Information Bottleneck principle to analyze token embeddings from various LLMs against human categorization benchmarks.

Result: LLMs form broad conceptual categories aligned with human judgment but struggle with fine-grained semantic distinctions. They show aggressive statistical compression bias, unlike humans who prioritize adaptive nuance and contextual richness.

Conclusion: Critical differences exist between AI and human cognitive architectures in conceptual representation, providing guidance for developing LLMs with more human-aligned representations.

Abstract: Humans organize knowledge into compact categories through semantic compression by mapping diverse instances to abstract representations while preserving meaning (e.g., robin and blue jay are both birds; most birds can fly). These concepts reflect a trade-off between expressive fidelity and representational simplicity. Large Language Models (LLMs) demonstrate remarkable linguistic abilities, yet whether their internal representations strike a human-like trade-off between compression and semantic fidelity is unclear. We introduce a novel information-theoretic framework, drawing from Rate-Distortion Theory and the Information Bottleneck principle, to quantitatively compare these strategies. Analyzing token embeddings from a diverse suite of LLMs against seminal human categorization benchmarks, we uncover key divergences. While LLMs form broad conceptual categories that align with human judgment, they struggle to capture the fine-grained semantic distinctions crucial for human understanding. More fundamentally, LLMs demonstrate a strong bias towards aggressive statistical compression, whereas human conceptual systems appear to prioritize adaptive nuance and contextual richness, even if this results in lower compressional efficiency by our measures. These findings illuminate critical differences between current AI and human cognitive architectures, guiding pathways toward LLMs with more human-aligned conceptual representations.

[192] Cog-TiPRO: Iterative Prompt Refinement with LLMs to Detect Cognitive Decline via Longitudinal Voice Assistant Commands

Kristin Qi, Youxiang Zhu, Caroline Summerour, John A. Batsis, Xiaohui Liang

Main category: cs.CL

TL;DR: Cog-TiPRO framework uses voice assistant data and AI to detect mild cognitive impairment with 73.8% accuracy through speech pattern analysis.

DetailsMotivation: Early detection of cognitive decline is crucial but traditional clinical assessments are labor-intensive and impractical for frequent monitoring. Voice assistant systems offer a non-invasive alternative for continuous monitoring.

Method: Proposed Cog-TiPRO framework combining: (1) LLM-driven iterative prompt refinement for linguistic feature extraction, (2) HuBERT-based acoustic feature extraction, and (3) transformer-based temporal modeling using iTransformer. Collected 18-month voice command data from 35 older adults.

Result: Achieved 73.80% accuracy and 72.67% F1-score in detecting MCI, outperforming baseline by 27.13%. Identified unique linguistic features characterizing cognitive decline in everyday command usage.

Conclusion: Voice assistant systems with AI analysis can effectively detect cognitive decline through speech patterns, providing a scalable non-invasive monitoring solution.

Abstract: Early detection of cognitive decline is crucial for enabling interventions that can slow neurodegenerative disease progression. Traditional diagnostic approaches rely on labor-intensive clinical assessments, which are impractical for frequent monitoring. Our pilot study investigates voice assistant systems (VAS) as non-invasive tools for detecting cognitive decline through longitudinal analysis of speech patterns in voice commands. Over an 18-month period, we collected voice commands from 35 older adults, with 15 participants providing daily at-home VAS interactions. To address the challenges of analyzing these short, unstructured and noisy commands, we propose Cog-TiPRO, a framework that combines (1) LLM-driven iterative prompt refinement for linguistic feature extraction, (2) HuBERT-based acoustic feature extraction, and (3) transformer-based temporal modeling. Using iTransformer, our approach achieves 73.80% accuracy and 72.67% F1-score in detecting MCI, outperforming its baseline by 27.13%. Through our LLM approach, we identify linguistic features that uniquely characterize everyday command usage patterns in individuals experiencing cognitive decline.

[193] Federated Retrieval-Augmented Generation: A Systematic Mapping Study

Abhijit Chakraborty, Chahana Dahal, Vivek Gupta

Main category: cs.CL

TL;DR: First systematic mapping study of Federated RAG (2020-2025), analyzing research focuses, architectures, trends, and challenges in privacy-preserving knowledge-intensive NLP.

DetailsMotivation: Address growing need for secure NLP in privacy-sensitive domains (healthcare, finance) by combining FL's distributed training with RAG's factual accuracy improvements.

Method: Conducted systematic mapping study following Kitchenham’s guidelines, developing structured classification of research focuses, contribution types, and application domains.

Result: Identified architectural patterns, temporal trends, key challenges (privacy-preserving retrieval, cross-client heterogeneity, evaluation limitations), and synthesized rapidly evolving research landscape.

Conclusion: Provides foundation for future work at RAG-federated systems intersection, identifying recurring design patterns and open questions in this emerging field.

Abstract: Federated Retrieval-Augmented Generation (Federated RAG) combines Federated Learning (FL), which enables distributed model training without exposing raw data, with Retrieval-Augmented Generation (RAG), which improves the factual accuracy of language models by grounding outputs in external knowledge. As large language models are increasingly deployed in privacy-sensitive domains such as healthcare, finance, and personalized assistance, Federated RAG offers a promising framework for secure, knowledge-intensive natural language processing (NLP). To the best of our knowledge, this paper presents the first systematic mapping study of Federated RAG, covering literature published between 2020 and 2025. Following Kitchenham’s guidelines for evidence-based software engineering, we develop a structured classification of research focuses, contribution types, and application domains. We analyze architectural patterns, temporal trends, and key challenges, including privacy-preserving retrieval, cross-client heterogeneity, and evaluation limitations. Our findings synthesize a rapidly evolving body of research, identify recurring design patterns, and surface open questions, providing a foundation for future work at the intersection of RAG and federated systems.

[194] Automated Essay Scoring Incorporating Annotations from Automated Feedback Systems

Christopher Ormerod

Main category: cs.CL

TL;DR: Incorporating feedback-oriented annotations (spelling/grammar errors and argumentative components) improves automated essay scoring accuracy using LLMs.

DetailsMotivation: To enhance automated essay scoring accuracy by integrating feedback-driven annotations that identify both surface-level errors and argumentative structure elements.

Method: Used PERSUADE corpus with two types of annotations: spelling/grammatical errors (generated by generative LLM) and argumentative components (identified by encoder-based token-classifier). Incorporated these annotations into scoring pipeline using fine-tuned encoder-based LLM classifiers.

Result: Demonstrated performance improvements in automated essay scoring when annotations are incorporated into the scoring process.

Conclusion: Feedback-oriented annotations significantly enhance AES accuracy, showing practical applicability for real-world educational scenarios through LLM-generated annotations.

Abstract: This study illustrates how incorporating feedback-oriented annotations into the scoring pipeline can enhance the accuracy of automated essay scoring (AES). This approach is demonstrated with the Persuasive Essays for Rating, Selecting, and Understanding Argumentative and Discourse Elements (PERSUADE) corpus. We integrate two types of feedback-driven annotations: those that identify spelling and grammatical errors, and those that highlight argumentative components. To illustrate how this method could be applied in real-world scenarios, we employ two LLMs to generate annotations – a generative language model used for spell correction and an encoder-based token-classifier trained to identify and mark argumentative elements. By incorporating annotations into the scoring process, we demonstrate improvements in performance using encoder-based large language models fine-tuned as classifiers.

[195] Multiple LLM Agents Debate for Equitable Cultural Alignment

Dayeon Ki, Rachel Rudinger, Tianyi Zhou, Marine Carpuat

Main category: cs.CL

TL;DR: Multi-agent debate framework using two LLMs to debate cultural scenarios improves cultural adaptability and accuracy over single-LLM approaches, enabling smaller models to match larger ones.

DetailsMotivation: LLMs need to adapt to diverse cultural contexts to serve global communities effectively, moving beyond single-model, single-turn approaches.

Method: Proposed Multi-Agent Debate framework with two variants: exclusive debate between two LLM agents, and dynamic choice between self-reflection and debate during turns. Evaluated on 7 open-weight LLMs using NormAd-ETI benchmark for social etiquette norms across 75 countries.

Result: Debate improves both overall accuracy and cultural group parity. Small LLMs (7-9B parameters) achieved accuracies comparable to much larger models (27B parameters) through multi-agent debate.

Conclusion: Multi-agent debate framework effectively enhances cultural adaptability of LLMs, demonstrating that collaborative approaches can compensate for model size limitations and improve performance across diverse cultural contexts.

Abstract: Large Language Models (LLMs) need to adapt their predictions to diverse cultural contexts to benefit diverse communities across the world. While previous efforts have focused on single-LLM, single-turn approaches, we propose to exploit the complementary strengths of multiple LLMs to promote cultural adaptability. We introduce a Multi-Agent Debate framework, where two LLM-based agents debate over a cultural scenario and collaboratively reach a final decision. We propose two variants: one where either LLM agents exclusively debate and another where they dynamically choose between self-reflection and debate during their turns. We evaluate these approaches on 7 open-weight LLMs (and 21 LLM combinations) using the NormAd-ETI benchmark for social etiquette norms in 75 countries. Experiments show that debate improves both overall accuracy and cultural group parity over single-LLM baselines. Notably, multi-agent debate enables relatively small LLMs (7-9B) to achieve accuracies comparable to that of a much larger model (27B parameters).

[196] Should I Share this Translation? Evaluating Quality Feedback for User Reliance on Machine Translation

Dayeon Ki, Kevin Duh, Marine Carpuat

Main category: cs.CL

TL;DR: Study compares four AI quality feedback methods for machine translation, finding that implicit feedback (especially question-answer tables) outperforms explicit feedback in improving user decision accuracy, appropriate reliance, and user experience.

DetailsMotivation: As AI systems become more prevalent, users need effective feedback mechanisms to use AI responsibly, especially when they lack the expertise to assess AI output quality themselves.

Method: Compared four types of quality feedback in a machine translation scenario: explicit feedback (error highlights and LLM explanations) and implicit feedback (backtranslation and question-answer tables), testing how they affect user decisions about sharing translations.

Result: All feedback types except error highlights significantly improved decision accuracy and appropriate reliance. Implicit feedback, particularly QA tables, showed the greatest gains in accuracy, reliance, and user perceptions (highest helpfulness/trust, lowest mental burden).

Conclusion: Implicit feedback mechanisms, especially question-answer tables, are more effective than explicit feedback for helping non-expert users make better decisions about AI-generated content quality.

Abstract: As people increasingly use AI systems in work and daily life, feedback mechanisms that help them use AI responsibly are urgently needed, particularly in settings where users are not equipped to assess the quality of AI predictions. We study a realistic Machine Translation (MT) scenario where monolingual users decide whether to share an MT output, first without and then with quality feedback. We compare four types of quality feedback: explicit feedback that directly give users an assessment of translation quality using (1) error highlights and (2) LLM explanations, and implicit feedback that helps users compare MT inputs and outputs through (3) backtranslation and (4) question-answer (QA) tables. We find that all feedback types, except error highlights, significantly improve both decision accuracy and appropriate reliance. Notably, implicit feedback, especially QA tables, yields significantly greater gains than explicit feedback in terms of decision accuracy, appropriate reliance, and user perceptions, receiving the highest ratings for helpfulness and trust, and the lowest for mental burden.

[197] Strategic Discourse Assessment: The Crooked Path to Innocence

Anshun Asher Zheng, Junyi Jessy Li, David I. Beaver

Main category: cs.CL

TL;DR: SDA framework combines Gricean pragmatics and game theory to analyze strategic language in adversarial settings, showing LLMs have limited understanding of strategic discourse despite model size improvements.

DetailsMotivation: Most pragmatics research focuses on cooperative communication, leaving a gap in understanding strategic language use in adversarial settings like courtroom cross-examinations.

Method: Developed SDA framework with commitment-based taxonomy of discourse moves and Gricean-based proxies. Created three metrics (BAT, PAT, NRBAT) and CPD dataset of annotated courtroom cross-examinations to evaluate LLMs.

Result: LLMs show limited pragmatic understanding of strategic language. Model size improves performance but reasoning ability hurts performance by causing overcomplication and internal confusion.

Conclusion: The SDA framework effectively assesses strategic discourse, revealing significant limitations in current LLMs’ ability to understand and process adversarial language strategies.

Abstract: Language is often used strategically, particularly in high-stakes, adversarial settings, yet most work on pragmatics and LLMs centers on cooperativity. This leaves a gap in the systematic understanding of strategic communication in adversarial settings. To address this, we introduce SDA (Strategic Discourse Assessment), a framework grounded in Gricean and game-theoretic pragmatics to assess strategic use of language. It adapts the ME Game jury function to make it empirically estimable for analyzing dialogue. Our approach incorporates two key adaptations: a commitment-based taxonomy of discourse moves, which provides a finer-grained account of strategic effects, and the use of estimable proxies grounded in Gricean maxims to operationalize abstract constructs such as credibility. Together, these adaptations build on discourse theory by treating discourse as the strategic management of commitments, enabling systematic evaluation of how conversational moves advance or undermine discourse goals. We further derive three interpretable metrics-Benefit at Turn (BAT), Penalty at Turn (PAT), and Normalized Relative Benefit at Turn (NRBAT)-to quantify the perceived strategic effects of discourse moves. We also present CPD (the Crooked Path Dataset), an annotated dataset of real courtroom cross-examinations, to demonstrate the framework’s effectiveness. Using these tools, we evaluate a range of LLMs and show that LLMs generally exhibit limited pragmatic understanding of strategic language. While model size shows an increase in performance on our metrics, reasoning ability does not help and largely hurts, introducing overcomplication and internal confusion.

[198] FinS-Pilot: A Benchmark for Online Financial RAG System

Feng Wang, Yiding Sun, Jiaxin Mao, Wei Xue, Danqing Xu

Main category: cs.CL

TL;DR: FinS-Pilot is a novel benchmark for evaluating RAG systems in financial applications, addressing data confidentiality and dynamic data integration challenges through real-world financial assistant interactions and real-time API data.

DetailsMotivation: Current financial RAG benchmarks are constrained by data confidentiality issues and lack dynamic data integration, creating a gap in specialized evaluation tools for the financial domain where professional accuracy and real-time data processing are crucial.

Method: Constructed benchmark from real-world financial assistant interactions incorporating both real-time API data and text data, organized through an intent classification framework covering critical financial domains to enable comprehensive evaluation of static knowledge and time-sensitive market information handling.

Result: Systematic experiments with multiple Chinese leading LLMs demonstrated FinS-Pilot’s effectiveness in identifying models suitable for financial applications, successfully addressing the current gap in specialized evaluation tools.

Conclusion: The work contributes both a practical evaluation framework and curated dataset to advance research in financial NLP systems, with code and dataset made accessible on GitHub for broader research community use.

Abstract: Large language models (LLMs) have demonstrated remarkable capabilities across various professional domains, with their performance typically evaluated through standardized benchmarks. In the financial field, the stringent demands for professional accuracy and real-time data processing often necessitate the use of retrieval-augmented generation (RAG) techniques. However, the development of financial RAG benchmarks has been constrained by data confidentiality issues and the lack of dynamic data integration. To address this issue, we introduce FinS-Pilot, a novel benchmark for evaluating RAG systems in online financial applications. Constructed from real-world financial assistant interactions, our benchmark incorporates both real-time API data and text data, organized through an intent classification framework covering critical financial domains. The benchmark enables comprehensive evaluation of financial assistants’ capabilities in handling both static knowledge and time-sensitive market information.Through systematic experiments with multiple Chinese leading LLMs, we demonstrate FinS-Pilot’s effectiveness in identifying models suitable for financial applications while addressing the current gap in specialized evaluation tools for the financial domain. Our work contributes both a practical evaluation framework and a curated dataset to advance research in financial NLP systems. The code and dataset are accessible on GitHub.

[199] Auto prompt sql: a resource-efficient architecture for text-to-sql translation in constrained environments

Zetong Tang, Qian Ma, Di Wu

Main category: cs.CL

TL;DR: AP-SQL bridges resource-efficient small models with large closed-source models for Text-to-SQL translation using decomposition, fine-tuning, and advanced prompt engineering techniques.

DetailsMotivation: Addressing the challenge of using resource-intensive Text-to-SQL methods in constrained environments by leveraging small open-source models while maintaining the capabilities of large closed-source models.

Method: Decomposes task into schema filtering, retrieval-augmented generation, and prompt-driven schema linking. Fine-tunes LLMs for schema selection and uses Chain-of-Thought and Graph-of-Thought prompt engineering to enhance reasoning.

Result: Comprehensive evaluations on Spider benchmarks demonstrate the effectiveness of the AP-SQL approach.

Conclusion: AP-SQL successfully bridges the gap between resource efficiency and performance in Text-to-SQL translation through innovative decomposition and prompt engineering techniques.

Abstract: Using the best Text-to-SQL methods in resource-constrained environments is challenging due to their reliance on resource-intensive open-source models. This paper introduces Auto Prompt SQL(AP-SQL), a novel architecture designed to bridge the gap between resource-efficient small open-source models and the powerful capabilities of large closed-source models for Text-to-SQL translation. Our method decomposes the task into schema filtering, retrieval-augmented text-to-SQL generation based on in-context examples, and prompt-driven schema linking and SQL generation. To improve schema selection accuracy, we fine-tune large language models. Crucially, we also explore the impact of prompt engineering throughout the process, leveraging Chain-of-Thought(CoT) and Graph-of-Thought(GoT) templates to significantly enhance the model’s reasoning for accurate SQL generation. Comprehensive evaluations on the Spider benchmarks demonstrate the effectiveness of AP-SQL.

[200] LLMEval-Med: A Real-world Clinical Benchmark for Medical LLMs with Physician Validation

Ming Zhang, Yujiong Shen, Zelin Li, Huayu Sha, Binze Hu, Yuhui Wang, Chenhao Huang, Shichun Liu, Jingqi Tong, Changhao Jiang, Mingxu Chai, Zhiheng Xi, Shihan Dou, Tao Gui, Qi Zhang, Xuanjing Huang

Main category: cs.CL

TL;DR: LLMEval-Med is a new medical benchmark for evaluating LLMs that addresses limitations of existing benchmarks by using real clinical scenarios from EHRs, covering 5 medical areas with 2,996 questions, and featuring an automated evaluation pipeline with expert-developed checklists.

DetailsMotivation: Current medical benchmarks have limitations including poor question design (mostly multiple-choice), non-clinical data sources, and inadequate evaluation of complex reasoning, which is insufficient for medical applications requiring high accuracy.

Method: Created 2,996 questions from real-world electronic health records and expert-designed clinical scenarios across 5 core medical areas. Developed an automated evaluation pipeline using LLM-as-Judge framework with expert-developed checklists, validated through human-machine agreement analysis and dynamic refinement based on expert feedback.

Result: Evaluated 13 LLMs across three categories (specialized medical models, open-source models, and closed-source models) on the LLMEval-Med benchmark, providing insights for safe and effective deployment of LLMs in medical domains.

Conclusion: LLMEval-Med addresses critical gaps in medical LLM evaluation by providing a more realistic, clinically-relevant benchmark with reliable automated assessment methods, supporting safer deployment of LLMs in healthcare applications.

Abstract: Evaluating large language models (LLMs) in medicine is crucial because medical applications require high accuracy with little room for error. Current medical benchmarks have three main types: medical exam-based, comprehensive medical, and specialized assessments. However, these benchmarks have limitations in question design (mostly multiple-choice), data sources (often not derived from real clinical scenarios), and evaluation methods (poor assessment of complex reasoning). To address these issues, we present LLMEval-Med, a new benchmark covering five core medical areas, including 2,996 questions created from real-world electronic health records and expert-designed clinical scenarios. We also design an automated evaluation pipeline, incorporating expert-developed checklists into our LLM-as-Judge framework. Furthermore, our methodology validates machine scoring through human-machine agreement analysis, dynamically refining checklists and prompts based on expert feedback to ensure reliability. We evaluate 13 LLMs across three categories (specialized medical models, open-source models, and closed-source models) on LLMEval-Med, providing valuable insights for the safe and effective deployment of LLMs in medical domains. The dataset is released in https://github.com/llmeval/LLMEval-Med.

[201] Surface Fairness, Deep Bias: A Comparative Study of Bias in Language Models

Aleksandra Sorokovikova, Pavel Chizhov, Iuliia Eremenko, Ivan P. Yamshchikov

Main category: cs.CL

TL;DR: The paper investigates bias in large language models, finding that traditional benchmarks show negligible bias, but more realistic tasks like grading user answers and salary negotiation advice reveal significant biases, especially concerning as models gain memory and personalization capabilities.

DetailsMotivation: Language models are trained on biased data containing controversial and stereotypical content, leading them to express biased viewpoints and produce different results based on user personality or assigned persona.

Method: The study uses various proxy measures of bias, including evaluating models with pre-prompted personae on MMLU benchmark, reformulating tasks to have models grade user answers, and testing salary negotiation advice scenarios.

Result: Traditional benchmark evaluation showed negligible and mostly random differences, but grading user answers revealed more significant bias, and salary negotiation advice showed pronounced bias. The problem is exacerbated as models gain memory and personalization capabilities.

Conclusion: Bias in LLMs is more pronounced in realistic application scenarios than in traditional benchmarks, and the trend toward personalized LLM assistants with memory capabilities makes this an increasingly important problem that requires attention.

Abstract: Modern language models are trained on large amounts of data. These data inevitably include controversial and stereotypical content, which contains all sorts of biases related to gender, origin, age, etc. As a result, the models express biased points of view or produce different results based on the assigned personality or the personality of the user. In this paper, we investigate various proxy measures of bias in large language models (LLMs). We find that evaluating models with pre-prompted personae on a multi-subject benchmark (MMLU) leads to negligible and mostly random differences in scores. However, if we reformulate the task and ask a model to grade the user’s answer, this shows more significant signs of bias. Finally, if we ask the model for salary negotiation advice, we see pronounced bias in the answers. With the recent trend for LLM assistant memory and personalization, these problems open up from a different angle: modern LLM users do not need to pre-prompt the description of their persona since the model already knows their socio-demographics.

[202] A Structured Bangla Dataset of Disease-Symptom Associations to Improve Diagnostic Accuracy

Abdullah Al Shafi, Rowzatul Zannat, Abdul Muntakim, Mahmudul Hasan

Main category: cs.CL

TL;DR: A structured Bangla disease-symptom dataset compiled from verified medical sources with binary symptom associations for machine learning applications.

DetailsMotivation: To address the significant gap in structured disease-symptom datasets for the Bangla language and improve diagnostic tools for underrepresented linguistic communities.

Method: Systematic compilation from online sources, medical literature, and health databases using peer-reviewed articles and clinical studies, excluding non-verified sources. Structured in tabular format with binary symptom indicators.

Result: Created a comprehensive disease-symptom dataset in Bangla language with verified medical relationships, enabling machine learning applications and clinical decision support.

Conclusion: The dataset bridges the linguistic gap in medical informatics and provides foundation for developing multilingual disease prediction tools, with potential for future expansion to include region-specific diseases and refined symptom associations.

Abstract: Disease-symptom datasets are significant and in demand for medical research, disease diagnosis, clinical decision-making, and AI-driven health management applications. These datasets help identify symptom patterns associated with specific diseases, thus improving diagnostic accuracy and enabling early detection. The dataset presented in this study systematically compiles disease-symptom relationships from various online sources, medical literature, and publicly available health databases. The data was gathered through analyzing peer-reviewed medical articles, clinical case studies, and disease-symptom association reports. Only the verified medical sources were included in the dataset, while those from non-peer-reviewed and anecdotal sources were excluded. The dataset is structured in a tabular format, where the first column represents diseases, and the remaining columns represent symptoms. Each symptom cell contains a binary value, indicating whether a symptom is associated with a disease. Thereby, this structured representation makes the dataset very useful for a wide range of applications, including machine learning-based disease prediction, clinical decision support systems, and epidemiological studies. Although there are some advancements in the field of disease-symptom datasets, there is a significant gap in structured datasets for the Bangla language. This dataset aims to bridge that gap by facilitating the development of multilingual medical informatics tools and improving disease prediction models for underrepresented linguistic communities. Further developments should include region-specific diseases and further fine-tuning of symptom associations for better diagnostic performance

[203] TPTT: Transforming Pretrained Transformers into Titans

Fabien Furfaro

Main category: cs.CL

TL;DR: TPTT framework converts pretrained Transformers to use linear attention (LiZA) and memory gating (MaG) for efficient long-context inference without full retraining, showing potential efficiency and accuracy improvements.

DetailsMotivation: Address quadratic computational/memory requirements of Transformer LLMs that limit efficient inference on long contexts and deployment in resource-limited environments.

Method: Augment pretrained Transformers with linearized attention (LiZA) and internal memory gating (MaG) via parameter-efficient fine-tuning (LoRA), supporting conversion to pure linear attention using DeltaProduct mechanism.

Result: Models with ~1B parameters showed up to 20% relative increase in Exact Match scores on MMLU benchmark, with efficient training using modest computational resources.

Conclusion: TPTT enables adaptation of pretrained LLMs for long-context tasks with limited overhead, though further studies on larger models and benchmarks are needed to evaluate generality.

Abstract: Transformer-based large language models (LLMs) have achieved strong performance across many natural language processing tasks. Nonetheless, their quadratic computational and memory requirements, particularly in self-attention layers, pose challenges for efficient inference on long contexts and for deployment in resource-limited environments. We present TPTT (Transforming Pretrained Transformers into Titans), a framework designed to augment pretrained Transformers with linearized attention (LiZA) and internal memory gating via Memory as Gate (MaG), applied without full retraining. TPTT supports parameter-efficient fine-tuning (LoRA) and integrates with standard toolkits such as Hugging Face Transformers. We evaluated TPTT on several pretrained models, including Llama-1B, OlMoE-1B-7B, Qwen2.5-1.5B, Gemma3-270m, OpenELM-1.3B, and Mistral-7B, in order to assess applicability across architectures of different scales. Experiments on models with approximately 1 billion parameters, evaluated primarily on the MMLU benchmark, suggest potential improvements in both efficiency and accuracy compared to baseline models. For example, Titans-Llama-1B exhibited up to a 20% relative increase in Exact Match scores in one-shot evaluation. An additional finding is that it is possible to convert a quadratic-attention model into a purely linear-attention model using the DeltaProduct mechanism. All training runs were carried out with modest computational resources. These preliminary findings indicate that TPTT may help adapt pretrained LLMs for long-context tasks with limited overhead. Further studies on larger models and a broader set of benchmarks will be necessary to evaluate the generality and robustness of the framework. Code is available at https://github.com/fabienfrfr/tptt . Python package at https://pypi.org/project/tptt/ .

[204] Flow-Modulated Scoring for Semantic-Aware Knowledge Graph Completion

Siyuan Li, Ruitong Liu, Yan Wen, Te Sun, Andi Zhang, Yanbiao Ma, Xiaoshuai Hao

Main category: cs.CL

TL;DR: FMS framework combines static semantic context learning with dynamic flow modeling to achieve state-of-the-art performance in knowledge graph completion with high parameter efficiency.

DetailsMotivation: Existing methods struggle to capture both rich semantic context and dynamic nature of relations simultaneously using static scoring functions over learned embeddings.

Method: Two-stage approach: 1) Semantic Context Learning module for context-aware entity embeddings, 2) Conditional Flow-Matching module to model dynamic flow between entities that modulates a base static score.

Result: Achieves near-perfect performance: 99.8% MRR and 99.7% Hits@1 on FB15k-237 with only 0.35M parameters, 99.9% MRR on WN18RR, and 25.2% relative MRR gain in transductive entity prediction.

Conclusion: FMS provides a highly effective and parameter-efficient paradigm for knowledge graph completion by unifying dynamic flow mechanisms with rich static contexts.

Abstract: Knowledge graph completion demands effective modeling of multifaceted semantic relationships between entities. Yet, prevailing methods, which rely on static scoring functions over learned embeddings, struggling to simultaneously capture rich semantic context and the dynamic nature of relations. To overcome this limitation, we propose the Flow-Modulated Scoring (FMS) framework, conceptualizing a relation as a dynamic evolutionary process governed by its static semantic environment. FMS operates in two stages: it first learns context-aware entity embeddings via a Semantic Context Learning module, and then models a dynamic flow between them using a Conditional Flow-Matching module. This learned flow dynamically modulates a base static score for the entity pair. By unifying context-rich static representations with a conditioned dynamic flow, FMS achieves a more comprehensive understanding of relational semantics. Extensive experiments demonstrate that FMS establishes a new state of the art across both canonical knowledge graph completion tasks: relation prediction and entity prediction. On the standard relation prediction benchmark FB15k-237, FMS achieves a near-perfect MRR of 99.8% and Hits@1 of 99.7% using a mere 0.35M parameters, while also attaining a 99.9% MRR on WN18RR. Its dominance extends to entity prediction, where it secures a 25.2% relative MRR gain in the transductive setting and substantially outperforms all baselines in challenging inductive settings. By unifying a dynamic flow mechanism with rich static contexts, FMS offers a highly effective and parameter-efficient new paradigm for knowledge graph completion. Code published at: https://github.com/yuanwuyuan9/FMS.

[205] MedVAL: Toward Expert-Level Medical Text Validation with Language Models

Asad Aali, Vasiliki Bikia, Maya Varma, Nicole Chiou, Sophie Ostmeier, Arnav Singhvi, Magdalini Paschali, Ashwin Kumar, Andrew Johnston, Karimar Amador-Martinez, Eduardo Juan Perez Guerrero, Paola Naovi Cruz Rivera, Sergios Gatidis, Christian Bluethgen, Eduardo Pontes Reis, Eddy D. Zandee van Rilland, Poonam Laxmappa Hosamani, Kevin R Keet, Minjoung Go, Evelyn Ling, David B. Larson, Curtis Langlotz, Roxana Daneshjou, Jason Hom, Sanmi Koyejo, Emily Alsentzer, Akshay S. Chaudhari

Main category: cs.CL

TL;DR: MedVAL is a self-supervised framework that trains LMs to evaluate medical text accuracy without physician labels, achieving 83% F1 score alignment with physicians and improving GPT-4o by 8%.

DetailsMotivation: Current LM evaluation in clinical settings relies on costly manual physician review and lacks expert references, while existing LM-as-judge approaches miss clinically significant errors.

Method: Proposes MedVAL framework using synthetic data to train evaluator LMs for factual consistency assessment without physician labels or reference outputs, and introduces MedVAL-Bench dataset with physician annotations.

Result: MedVAL fine-tuning significantly improves alignment with physicians (66% to 83% F1), achieves 86% safety classification, and boosts GPT-4o performance by 8% across 6 medical tasks and 10 LMs.

Conclusion: MedVAL provides the first evidence of LMs approaching expert-level medical text validation, enabling scalable, risk-aware clinical integration with open-sourced code, benchmark, and models.

Abstract: With the growing use of language models (LMs) in clinical environments, there is an immediate need to evaluate the accuracy and safety of LM-generated medical text. Currently, such evaluation relies solely on manual physician review. However, detecting errors in LM-generated text is challenging because

  1. manual review is costly and 2) expert-composed reference outputs are often unavailable in real-world settings. While the “LM-as-judge” paradigm (a LM evaluating another LM) offers scalable evaluation, even frontier LMs can miss subtle but clinically significant errors. To address these challenges, we propose MedVAL, a self-supervised framework that leverages synthetic data to train evaluator LMs to assess whether LM-generated medical outputs are factually consistent with inputs, without requiring physician labels or reference outputs. To evaluate LM performance, we introduce MedVAL-Bench, a dataset containing 840 outputs annotated by physicians, following a physician-defined taxonomy of risk levels and error categories. Across 6 diverse medical tasks and 10 state-of-the-art LMs spanning open-source, proprietary, and medically adapted models, MedVAL fine-tuning significantly improves (p < 0.001) alignment with physicians on both seen and unseen tasks, increasing average F1 scores from 66% to 83%, with per-sample safety classification scores up to 86%. MedVAL improves the performance of even the best-performing proprietary LM (GPT-4o) by 8%. To support a scalable, risk-aware pathway towards clinical integration, we open-source the 1) codebase (https://github.com/StanfordMIMI/MedVAL), 2) MedVAL-Bench (https://huggingface.co/datasets/stanfordmimi/MedVAL-Bench), and 3) MedVAL-4B (https://huggingface.co/stanfordmimi/MedVAL-4B), the best-performing open-source LM. Our research provides the first evidence of LMs approaching expert-level validation ability for medical text.

[206] Agentic-R1: Distilled Dual-Strategy Reasoning

Weihua Du, Pranjal Aggarwal, Sean Welleck, Yiming Yang

Main category: cs.CL

TL;DR: DualDistill framework distills multiple reasoning strategies into a single model that dynamically selects between tool-based computation and text-based reasoning for optimal problem-solving.

DetailsMotivation: Current approaches have limitations - long chain-of-thought models are slow and error-prone with natural language traces, while tool-augmented agents struggle with complex logical tasks despite handling arithmetic well.

Method: Fine-tuning framework that distills complementary reasoning strategies from multiple teachers into a unified student model (Agentic-R1) that dynamically selects optimal strategy per query.

Result: Improves accuracy across computation-intensive and standard benchmarks, demonstrating robust and efficient reasoning through multi-strategy distillation.

Conclusion: The approach effectively combines tool-based computation for arithmetic/algorithmic problems with text-based reasoning for abstract tasks, achieving superior performance across diverse reasoning tasks.

Abstract: Current long chain-of-thought (long-CoT) models excel at mathematical reasoning but rely on slow and error-prone natural language traces. Tool-augmented agents address arithmetic via code execution, but often falter on complex logical tasks. We introduce a fine-tuning framework, DualDistill, that distills complementary reasoning strategies from multiple teachers into a unified student model. Using this approach, we train Agentic-R1, which dynamically selects the optimal strategy for each query, invoking tools for arithmetic and algorithmic problems, and using text-based reasoning for abstract ones. Our method improves accuracy across a range of tasks, including both computation-intensive and standard benchmarks, demonstrating the effectiveness of multi-strategy distillation in achieving robust and efficient reasoning. Our project is available at https://github.com/StigLidu/DualDistill

[207] How Important is `Perfect’ English for Machine Translation Prompts?

Patrícia Schmidtová, Niyati Bafna, Seth Aycock, Gianluca Vico, Wiktor Kamzela, Katharina Hämmerl, Vilém Zouhar

Main category: cs.CL

TL;DR: LLMs’ translation performance is highly sensitive to prompt errors, with different noise types affecting quality differently, but models can still translate even with overwhelming noise that humans can’t read.

DetailsMotivation: To systematically evaluate how humanly plausible and synthetic errors in user prompts affect LLMs' performance on machine translation and translation evaluation tasks.

Method: Quantitative analysis and qualitative insights into how models respond to increasing noise in user prompts, testing different noise types including character-level, phrasal perturbations, and combined noisers.

Result: Prompt quality strongly affects translation performance - good prompts with errors can underperform poor prompts without errors. Character-level and combined noise degrade performance more than phrasal perturbations. LLMs can translate even with overwhelming random noise that makes prompts illegible to humans.

Conclusion: Lower prompt quality mainly leads to poorer instruction following rather than directly affecting translation quality itself, showing LLMs’ robustness to extreme noise but sensitivity to prompt errors.

Abstract: Large language models (LLMs) have achieved top results in recent machine translation evaluations, but they are also known to be sensitive to errors and perturbations in their prompts. We systematically evaluate how both humanly plausible and synthetic errors in user prompts affect LLMs’ performance on two related tasks: Machine translation and machine translation evaluation. We provide both a quantitative analysis and qualitative insights into how the models respond to increasing noise in the user prompt. The prompt quality strongly affects the translation performance: With many errors, even a good prompt can underperform a minimal or poor prompt without errors. However, different noise types impact translation quality differently, with character-level and combined noisers degrading performance more than phrasal perturbations. Qualitative analysis reveals that lower prompt quality largely leads to poorer instruction following, rather than directly affecting translation quality itself. Further, LLMs can still translate in scenarios with overwhelming random noise that would make the prompt illegible to humans.

[208] Cross-Domain Transfer and Few-Shot Learning for Personal Identifiable Information Recognition

Junhong Ye, Xu Yuan, Xinying Qiu

Main category: cs.CL

TL;DR: Cross-domain PII recognition study shows legal data transfers well to biography domains, medical data resists transfer, fusion benefits are domain-specific, and 10% training data suffices for low-specialization domains.

DetailsMotivation: To investigate effective strategies for PII recognition through cross-domain model transfer, multi-domain data fusion, and sample-efficient learning to improve automated text anonymization.

Method: Evaluated models using annotated corpora from healthcare (I2B2), legal (TAB), and biography (Wikipedia) across four dimensions: in-domain performance, cross-domain transferability, fusion, and few-shot learning.

Result: Legal-domain data transfers effectively to biographical texts, medical domains resist incoming transfer, fusion benefits are domain-specific, and high-quality recognition is achievable with only 10% of training data in low-specialization domains.

Conclusion: Domain characteristics significantly impact PII recognition transferability, with legal data showing good cross-domain applicability while medical data remains domain-specific, and minimal training data can be sufficient for less specialized domains.

Abstract: Accurate recognition of personally identifiable information (PII) is central to automated text anonymization. This paper investigates the effectiveness of cross-domain model transfer, multi-domain data fusion, and sample-efficient learning for PII recognition. Using annotated corpora from healthcare (I2B2), legal (TAB), and biography (Wikipedia), we evaluate models across four dimensions: in-domain performance, cross-domain transferability, fusion, and few-shot learning. Results show legal-domain data transfers well to biographical texts, while medical domains resist incoming transfer. Fusion benefits are domain-specific, and high-quality recognition is achievable with only 10% of training data in low-specialization domains.

[209] Label Unification for Cross-Dataset Generalization in Cybersecurity NER

Maciej Jalocha, Johan Hausted Schmidt, William Michelseen

Main category: cs.CL

TL;DR: This paper addresses the lack of standardized labels in cybersecurity NER by investigating label unification across four datasets, but finds that unified models generalize poorly and proposed alternative architectures provide limited improvements.

DetailsMotivation: The cybersecurity NER field lacks standardized labels, making it difficult to combine datasets and limiting data resource usability across different cybersecurity datasets.

Method: Performed coarse-grained label unification across four cybersecurity datasets, conducted pairwise cross-dataset evaluations using BiLSTM models, and proposed alternative architectures including a multihead model with weight sharing and a graph-based transfer model built on BERT-base-NER.

Result: Models trained on unified datasets generalized poorly across different datasets. The multihead model provided only marginal improvements over unified training, and the graph-based transfer model showed no significant performance gains compared to standard BERT-base-NER.

Conclusion: Label unification in cybersecurity NER faces significant challenges, and the proposed alternative architectures did not substantially improve cross-dataset generalization, indicating the need for more effective approaches to handle dataset heterogeneity in this domain.

Abstract: The field of cybersecurity NER lacks standardized labels, making it challenging to combine datasets. We investigate label unification across four cybersecurity datasets to increase data resource usability. We perform a coarse-grained label unification and conduct pairwise cross-dataset evaluations using BiLSTM models. Qualitative analysis of predictions reveals errors, limitations, and dataset differences. To address unification limitations, we propose alternative architectures including a multihead model and a graph-based transfer model. Results show that models trained on unified datasets generalize poorly across datasets. The multihead model with weight sharing provides only marginal improvements over unified training, while our graph-based transfer model built on BERT-base-NER shows no significant performance gains compared BERT-base-NER.

[210] Bottom-up Domain-specific Superintelligence: A Reliable Knowledge Graph is What We Need

Bhishma Dedhia, Yuval Kansal, Niraj K. Jha

Main category: cs.CL

TL;DR: The paper presents a bottom-up approach using knowledge graphs to train language models for domain-specific expertise, demonstrating medical superintelligence through KG-grounded task generation and fine-tuning.

DetailsMotivation: Traditional top-down training on general corpora is insufficient for acquiring deep domain expertise, requiring a bottom-up approach that composes simple domain concepts into complex ones using knowledge graphs.

Method: A task generation pipeline synthesizes reasoning tasks directly from KG primitives, creating a KG-grounded curriculum. The QwQ-32B model is fine-tuned on 24,000 medical reasoning tasks derived from a medical knowledge graph.

Result: QwQ-Med-3 significantly outperforms state-of-the-art reasoning models on ICD-Bench across 15 medical domains and shows enhanced performance on medical QA benchmarks, utilizing acquired primitives for hardest tasks.

Conclusion: The approach demonstrates that domain-specific superintelligence can be achieved through composable KG-grounded training, suggesting AGI may emerge from efficient domain-specific agents rather than broad general training.

Abstract: Language models traditionally used for cross-domain generalization have recently demonstrated task-specific reasoning. However, their top-down training approach on general corpora is insufficient for acquiring abstractions needed for deep domain expertise. This may require a bottom-up approach that acquires expertise by learning to compose simple domain concepts into more complex ones. A knowledge graph (KG) provides this compositional structure, where domain primitives are represented as head-relation-tail edges and their paths encode higher-level concepts. We present a task generation pipeline that synthesizes tasks directly from KG primitives, enabling models to acquire and compose them for reasoning. We fine-tune language models on the resultant KG-grounded curriculum to demonstrate domain-specific superintelligence. While broadly applicable, we validate our approach in medicine, where reliable KGs exist. Using a medical KG, we curate 24,000 reasoning tasks paired with thinking traces derived from diverse medical primitives. We fine-tune the QwQ-32B model on this curriculum to obtain QwQ-Med-3 that takes a step towards medical superintelligence. We also introduce ICD-Bench, an evaluation suite to quantify reasoning abilities across 15 medical domains. Our experiments demonstrate that QwQ-Med-3 significantly outperforms state-of-the-art reasoning models on ICD-Bench categories. Further analysis reveals that QwQ-Med-3 utilizes acquired primitives to widen the performance gap on the hardest tasks of ICD-Bench. Finally, evaluation on medical question-answer benchmarks shows that QwQ-Med-3 transfers acquired expertise to enhance the base model’s performance. While the industry’s approach to artificial general intelligence (AGI) emphasizes broad expertise, we envision a future in which AGI emerges from the composable interaction of efficient domain-specific superintelligent agents.

[211] From Disagreement to Understanding: The Case for Ambiguity Detection in NLI

Chathuri Jayaweera, Bonnie J. Dorr

Main category: cs.CL

TL;DR: Annotation disagreement in NLI is meaningful variation from ambiguity, not noise. The paper proposes ambiguity-aware NLI with detection/classification before inference, introduces a unified taxonomy, and calls for new annotated resources.

DetailsMotivation: Current NLI approaches treat annotation disagreement as noise, but the authors argue it reflects meaningful human interpretation differences, particularly from ambiguity in premises/hypotheses.

Method: Proposes a framework that incorporates ambiguity detection and classification prior to inference. Introduces a unified taxonomy synthesizing existing taxonomies and motivates targeted detection methods.

Result: The paper presents a conceptual framework and taxonomy but acknowledges current lack of datasets explicitly annotated for ambiguity and subtypes, identifying this as a research opportunity.

Conclusion: Shifting to ambiguity-aware NLI through new annotated resources and unsupervised approaches will enable more robust, explainable, and human-aligned NLI systems that better capture divergent human perspectives.

Abstract: This position paper argues that annotation disagreement in Natural Language Inference (NLI) is not mere noise but often reflects meaningful variation, especially when triggered by ambiguity in the premise or hypothesis. While underspecified guidelines and annotator behavior contribute to variation, content-based ambiguity provides a process-independent signal of divergent human perspectives. We call for a shift toward ambiguity-aware NLI that first identifies ambiguous input pairs, classifies their types, and only then proceeds to inference. To support this shift, we present a framework that incorporates ambiguity detection and classification prior to inference. We also introduce a unified taxonomy that synthesizes existing taxonomies, illustrates key subtypes with examples, and motivates targeted detection methods that better align models with human interpretation. Although current resources lack datasets explicitly annotated for ambiguity and subtypes, this gap presents an opportunity: by developing new annotated resources and exploring unsupervised approaches to ambiguity detection, we enable more robust, explainable, and human-aligned NLI systems.

[212] Towards Compute-Optimal Many-Shot In-Context Learning

Shahriar Golchin, Yanfei Chen, Rujun Han, Manan Gandhi, Tianli Yu, Swaroop Mishra, Mihai Surdeanu, Rishabh Agarwal, Chen-Yu Lee, Tomas Pfister

Main category: cs.CL

TL;DR: Two efficient demonstration selection strategies for many-shot in-context learning that combine similarity-based selection with cached random/centroid demonstrations to improve performance while reducing inference costs.

DetailsMotivation: Current many-shot ICL approaches often use random demonstration selection due to high inference costs and caching benefits, but better selection strategies could improve performance without significant computational overhead.

Method: 1) Combine a small set of similarity-based demonstrations with a larger cached set of random demonstrations. 2) Replace random demonstrations with centroid-based demonstrations from k-means clustering of test samples.

Result: Strategies consistently outperform random selection and match top-performing approaches while reducing inference cost by up to 10x across Gemini Pro and Flash models on multiple datasets.

Conclusion: The proposed hybrid demonstration selection methods effectively balance performance and computational efficiency in many-shot ICL, enabling better results with significantly reduced inference costs.

Abstract: Long-context large language models (LLMs) are able to process inputs containing up to several million tokens. In the scope of in-context learning (ICL), this translates into using hundreds/thousands of demonstrations in the input prompt, enabling many-shot ICL. In practice, a fixed set of demonstrations is often selected at random in many-shot settings due to (1) high inference costs, (2) the benefits of caching and reusing computations, and (3) the similar performance offered by this strategy compared to others when scaled. In this work, we propose two straightforward strategies for demonstration selection in many-shot ICL that improve performance with minimal computational overhead. Our first method combines a small number of demonstrations, selected based on their similarity to each test sample, with a disproportionately larger set of random demonstrations that are cached. The second strategy improves the first by replacing random demonstrations with those selected using centroids derived from test sample representations via k-means clustering. Our experiments with Gemini Pro and Flash across several datasets indicate that our strategies consistently outperform random selection and surpass or match the most performant selection approach while supporting caching and reducing inference cost by up to an order of magnitude. We also show that adjusting the proportion of demonstrations selected based on different criteria can balance performance and inference cost in many-shot ICL.

[213] Persona Vectors: Monitoring and Controlling Character Traits in Language Models

Runjin Chen, Andy Arditi, Henry Sleight, Owain Evans, Jack Lindsey

Main category: cs.CL

TL;DR: The paper identifies ‘persona vectors’ in LLM activation space that represent personality traits like evil, sycophancy, and hallucination propensity, enabling monitoring and control of personality shifts during deployment and training.

DetailsMotivation: Large language models sometimes deviate from their intended helpful, harmless, and honest Assistant persona, requiring methods to monitor and control these undesirable personality traits.

Method: Extract automated persona vectors from model activation space using natural-language descriptions, apply these vectors to predict and control personality shifts during finetuning, and develop preventative steering methods and data filtering techniques.

Result: Persona vectors effectively monitor personality fluctuations at deployment and strongly correlate with both intended and unintended personality changes after finetuning. Shifts can be mitigated through post-hoc intervention or prevented with new steering methods.

Conclusion: Persona vectors provide an automated, scalable approach to monitor and control LLM personality traits, enabling detection of undesirable changes and prevention methods at both dataset and individual sample levels.

Abstract: Large language models interact with users through a simulated ‘Assistant’ persona. While the Assistant is typically trained to be helpful, harmless, and honest, it sometimes deviates from these ideals. In this paper, we identify directions in the model’s activation space-persona vectors-underlying several traits, such as evil, sycophancy, and propensity to hallucinate. We confirm that these vectors can be used to monitor fluctuations in the Assistant’s personality at deployment time. We then apply persona vectors to predict and control personality shifts that occur during training. We find that both intended and unintended personality changes after finetuning are strongly correlated with shifts along the relevant persona vectors. These shifts can be mitigated through post-hoc intervention, or avoided in the first place with a new preventative steering method. Moreover, persona vectors can be used to flag training data that will produce undesirable personality changes, both at the dataset level and the individual sample level. Our method for extracting persona vectors is automated and can be applied to any personality trait of interest, given only a natural-language description.

[214] MAQuA: Adaptive Question-Asking for Multidimensional Mental Health Screening using Item Response Theory

Vasudha Varadarajan, Hui Xu, Rebecca Astrid Boehme, Mariam Marlan Mirstrom, Sverker Sikstrom, H. Andrew Schwartz

Main category: cs.CL

TL;DR: MAQuA is an adaptive mental health screening framework that reduces assessment questions by 50-87% compared to random ordering while maintaining accuracy across multiple symptom domains.

DetailsMotivation: Large language models offer opportunities for scalable mental health assessment, but excessive querying burdens users and is inefficient for screening across multiple symptom dimensions simultaneously.

Method: Combines multi-outcome modeling on language responses with item response theory (IRT) and factor analysis to select the most informative questions across multiple dimensions at each turn.

Result: Reduces assessment questions by 50-87% (71% fewer for depression, 85% fewer for eating disorders) while achieving stable scores. Performs robustly across internalizing (depression, anxiety) and externalizing (substance use, eating disorder) domains.

Conclusion: MAQuA is a powerful and efficient tool for scalable, nuanced mental health screening that advances LLM integration into clinical workflows by reducing patient burden while maintaining diagnostic accuracy.

Abstract: Recent advances in large language models (LLMs) offer new opportunities for scalable, interactive mental health assessment, but excessive querying by LLMs burdens users and is inefficient for real-world screening across transdiagnostic symptom profiles. We introduce MAQuA, an adaptive question-asking framework for simultaneous, multidimensional mental health screening. Combining multi-outcome modeling on language responses with item response theory (IRT) and factor analysis, MAQuA selects the questions with most informative responses across multiple dimensions at each turn to optimize diagnostic information, improving accuracy and potentially reducing response burden. Empirical results on a novel dataset reveal that MAQuA reduces the number of assessment questions required for score stabilization by 50-87% compared to random ordering (e.g., achieving stable depression scores with 71% fewer questions and eating disorder scores with 85% fewer questions). MAQuA demonstrates robust performance across both internalizing (depression, anxiety) and externalizing (substance use, eating disorder) domains, with early stopping strategies further reducing patient time and burden. These findings position MAQuA as a powerful and efficient tool for scalable, nuanced, and interactive mental health screening, advancing the integration of LLM-based agents into real-world clinical workflows.

[215] Weakly Supervised Fine-grained Span-Level Framework for Chinese Radiology Report Quality Assurance

Kaiyu Wang, Lin Mu, Zhiyao Yang, Ximing Li, Xiaotang Zhou Wanfu Gao, Huimao Zhang

Main category: cs.CL

TL;DR: Sqator is an automated system that evaluates radiology report quality by analyzing fine-grained text span revisions between junior and senior doctor reports, achieving competitive QA scores while reducing senior doctor workload.

DetailsMotivation: Manual quality assurance of radiology reports by senior doctors is labor-intensive and potentially inaccurate due to factors like diagnosis bias and varying doctor abilities.

Method: Span-level Quality Assurance EvaluaTOR (Sqator) analyzes semantic differences at fine-grained text span level, measuring importance of revised spans between junior and senior reports, then merges all revised span scores for final QA assessment.

Result: Evaluation on 12,013 radiology reports shows Sqator achieves competitive QA scores, with revised span importance scores consistent with senior doctors’ judgments.

Conclusion: Sqator provides an effective automated solution for radiology report quality assurance that reduces senior doctor workload while maintaining accuracy comparable to human evaluation.

Abstract: Quality Assurance (QA) for radiology reports refers to judging whether the junior reports (written by junior doctors) are qualified. The QA scores of one junior report are given by the senior doctor(s) after reviewing the image and junior report. This process requires intensive labor costs for senior doctors. Additionally, the QA scores may be inaccurate for reasons like diagnosis bias, the ability of senior doctors, and so on. To address this issue, we propose a Span-level Quality Assurance EvaluaTOR (Sqator) to mark QA scores automatically. Unlike the common document-level semantic comparison method, we try to analyze the semantic difference by exploring more fine-grained text spans. Specifically, Sqator measures QA scores by measuring the importance of revised spans between junior and senior reports, and outputs the final QA scores by merging all revised span scores. We evaluate Sqator using a collection of 12,013 radiology reports. Experimental results show that Sqator can achieve competitive QA scores. Moreover, the importance scores of revised spans can be also consistent with the judgments of senior doctors.

[216] MMReview: A Multidisciplinary and Multimodal Benchmark for LLM-Based Peer Review Automation

Xian Gao, Jiacheng Ruan, Zongyun Zhang, Jingsheng Gao, Ting Liu, Yuzhuo Fu

Main category: cs.CL

TL;DR: MMReview is a comprehensive multimodal benchmark for evaluating LLMs’ peer review capabilities across 17 research domains with 240 papers and 13 tasks.

DetailsMotivation: Current LLM-based review systems lack a unified evaluation benchmark, especially for handling multimodal content like figures and tables in peer review scenarios.

Method: Created MMReview benchmark with 240 papers across 17 domains, featuring multimodal content and expert-written reviews. Designed 13 tasks in 4 categories to evaluate LLMs and MLLMs on review generation, outcome formulation, human alignment, and robustness.

Result: Extensive experiments on 16 open-source and 5 closed-source models demonstrate the benchmark’s thoroughness in assessing automated peer review capabilities.

Conclusion: MMReview provides a standardized foundation for developing automated peer review systems and represents a critical step forward in evaluating AI-assisted academic review.

Abstract: With the rapid growth of academic publications, peer review has become an essential yet time-consuming responsibility within the research community. Large Language Models (LLMs) have increasingly been adopted to assist in the generation of review comments; however, current LLM-based review tasks lack a unified evaluation benchmark to rigorously assess the models’ ability to produce comprehensive, accurate, and human-aligned assessments, particularly in scenarios involving multimodal content such as figures and tables. To address this gap, we propose \textbf{MMReview}, a comprehensive benchmark that spans multiple disciplines and modalities. MMReview includes multimodal content and expert-written review comments for 240 papers across 17 research domains within four major academic disciplines: Artificial Intelligence, Natural Sciences, Engineering Sciences, and Social Sciences. We design a total of 13 tasks grouped into four core categories, aimed at evaluating the performance of LLMs and Multimodal LLMs (MLLMs) in step-wise review generation, outcome formulation, alignment with human preferences, and robustness to adversarial input manipulation. Extensive experiments conducted on 16 open-source models and 5 advanced closed-source models demonstrate the thoroughness of the benchmark. We envision MMReview as a critical step toward establishing a standardized foundation for the development of automated peer review systems.

[217] NVIDIA Nemotron Nano 2: An Accurate and Efficient Hybrid Mamba-Transformer Reasoning Model

NVIDIA, :, Aarti Basant, Abhijit Khairnar, Abhijit Paithankar, Abhinav Khattar, Adithya Renduchintala, Aditya Malte, Akhiad Bercovich, Akshay Hazare, Alejandra Rico, Aleksander Ficek, Alex Kondratenko, Alex Shaposhnikov, Alexander Bukharin, Ali Taghibakhshi, Amelia Barton, Ameya Sunil Mahabaleshwarkar, Amy Shen, Andrew Tao, Ann Guan, Anna Shors, Anubhav Mandarwal, Arham Mehta, Arun Venkatesan, Ashton Sharabiani, Ashwath Aithal, Ashwin Poojary, Ayush Dattagupta, Balaram Buddharaju, Banghua Zhu, Barnaby Simkin, Bilal Kartal, Bita Darvish Rouhani, Bobby Chen, Boris Ginsburg, Brandon Norick, Brian Yu, Bryan Catanzaro, Charles Wang, Charlie Truong, Chetan Mungekar, Chintan Patel, Chris Alexiuk, Christian Munley, Christopher Parisien, Dan Su, Daniel Afrimi, Daniel Korzekwa, Daniel Rohrer, Daria Gitman, David Mosallanezhad, Deepak Narayanan, Dima Rekesh, Dina Yared, Dmytro Pykhtar, Dong Ahn, Duncan Riach, Eileen Long, Elliott Ning, Eric Chung, Erick Galinkin, Evelina Bakhturina, Gargi Prasad, Gerald Shen, Haifeng Qian, Haim Elisha, Harsh Sharma, Hayley Ross, Helen Ngo, Herman Sahota, Hexin Wang, Hoo Chang Shin, Hua Huang, Iain Cunningham, Igor Gitman, Ivan Moshkov, Jaehun Jung, Jan Kautz, Jane Polak Scowcroft, Jared Casper, Jian Zhang, Jiaqi Zeng, Jimmy Zhang, Jinze Xue, Jocelyn Huang, Joey Conway, John Kamalu, Jonathan Cohen, Joseph Jennings, Julien Veron Vialard, Junkeun Yi, Jupinder Parmar, Kari Briski, Katherine Cheung, Katherine Luna, Keith Wyss, Keshav Santhanam, Kezhi Kong, Krzysztof Pawelec, Kumar Anik, Kunlun Li, Kushan Ahmadian, Lawrence McAfee, Laya Sleiman, Leon Derczynski, Luis Vega, Maer Rodrigues de Melo, Makesh Narsimhan Sreedhar, Marcin Chochowski, Mark Cai, Markus Kliegl, Marta Stepniewska-Dziubinska, Matvei Novikov, Mehrzad Samadi, Meredith Price, Meriem Boubdir, Michael Boone, Michael Evans, Michal Bien, Michal Zawalski, Miguel Martinez, Mike Chrzanowski, Mohammad Shoeybi, Mostofa Patwary, Namit Dhameja, Nave Assaf, Negar Habibi, Nidhi Bhatia, Nikki Pope, Nima Tajbakhsh, Nirmal Kumar Juluru, Oleg Rybakov, Oleksii Hrinchuk, Oleksii Kuchaiev, Oluwatobi Olabiyi, Pablo Ribalta, Padmavathy Subramanian, Parth Chadha, Pavlo Molchanov, Peter Dykas, Peter Jin, Piotr Bialecki, Piotr Januszewski, Pradeep Thalasta, Prashant Gaikwad, Prasoon Varshney, Pritam Gundecha, Przemek Tredak, Rabeeh Karimi Mahabadi, Rajen Patel, Ran El-Yaniv, Ranjit Rajan, Ria Cheruvu, Rima Shahbazyan, Ritika Borkar, Ritu Gala, Roger Waleffe, Ruoxi Zhang, Russell J. Hewett, Ryan Prenger, Sahil Jain, Samuel Kriman, Sanjeev Satheesh, Saori Kaji, Sarah Yurick, Saurav Muralidharan, Sean Narenthiran, Seonmyeong Bak, Sepehr Sameni, Seungju Han, Shanmugam Ramasamy, Shaona Ghosh, Sharath Turuvekere Sreenivas, Shelby Thomas, Shizhe Diao, Shreya Gopal, Shrimai Prabhumoye, Shubham Toshniwal, Shuoyang Ding, Siddharth Singh, Siddhartha Jain, Somshubra Majumdar, Soumye Singhal, Stefania Alborghetti, Syeda Nahida Akter, Terry Kong, Tim Moon, Tomasz Hliwiak, Tomer Asida, Tony Wang, Tugrul Konuk, Twinkle Vashishth, Tyler Poon, Udi Karpas, Vahid Noroozi, Venkat Srinivasan, Vijay Korthikanti, Vikram Fugro, Vineeth Kalluru, Vitaly Kurin, Vitaly Lavrukhin, Wasi Uddin Ahmad, Wei Du, Wonmin Byeon, Ximing Lu, Xin Dong, Yashaswi Karnati, Yejin Choi, Yian Zhang, Ying Lin, Yonggan Fu, Yoshi Suhara, Zhen Dong, Zhiyu Li, Zhongbo Zhu, Zijia Chen

Main category: cs.CL

TL;DR: Nemotron-Nano-9B-v2 is a hybrid Mamba-Transformer model that achieves state-of-the-art reasoning accuracy with 6x higher throughput than similar-sized models, enabling 128k token inference on a single A10G GPU.

DetailsMotivation: To increase throughput for reasoning workloads while maintaining high accuracy by replacing most self-attention layers with more efficient Mamba-2 layers for faster long-sequence generation.

Method: Pre-trained a 12B parameter base model on 20T tokens using FP8 training, then used Minitron strategy for compression and distillation to enable 128k token inference on limited GPU memory.

Result: Achieves on-par or better accuracy than Qwen3-8B on reasoning benchmarks with up to 6x higher inference throughput for 8k input + 16k output token scenarios.

Conclusion: The hybrid Mamba-Transformer architecture successfully balances accuracy and efficiency, making high-performance reasoning models more accessible on consumer-grade hardware.

Abstract: We introduce Nemotron-Nano-9B-v2, a hybrid Mamba-Transformer language model designed to increase throughput for reasoning workloads while achieving state-of-the-art accuracy compared to similarly-sized models. Nemotron-Nano-9B-v2 builds on the Nemotron-H architecture, in which the majority of the self-attention layers in the common Transformer architecture are replaced with Mamba-2 layers, to achieve improved inference speed when generating the long thinking traces needed for reasoning. We create Nemotron-Nano-9B-v2 by first pre-training a 12-billion-parameter model (Nemotron-Nano-12B-v2-Base) on 20 trillion tokens using an FP8 training recipe. After aligning Nemotron-Nano-12B-v2-Base, we employ the Minitron strategy to compress and distill the model with the goal of enabling inference on up to 128k tokens on a single NVIDIA A10G GPU (22GiB of memory, bfloat16 precision). Compared to existing similarly-sized models (e.g., Qwen3-8B), we show that Nemotron-Nano-9B-v2 achieves on-par or better accuracy on reasoning benchmarks while achieving up to 6x higher inference throughput in reasoning settings like 8k input and 16k output tokens. We are releasing Nemotron-Nano-9B-v2, Nemotron-Nano12B-v2-Base, and Nemotron-Nano-9B-v2-Base checkpoints along with the majority of our pre- and post-training datasets on Hugging Face.

[218] MedResearcher-R1: Expert-Level Medical Deep Researcher via A Knowledge-Informed Trajectory Synthesis Framework

Ailing Yu, Lan Yao, Jingnan Liu, Zhe Chen, Jiajun Yin, Yuan Wang, Xinhao Liao, Zhiling Ye, Ji Li, Yun Yue, Hansong Xiao, Hualei Zhou, Chunxiao Guo, Peng Wei, Junwei Liu, Jinjie Gu

Main category: cs.CL

TL;DR: Medical deep research agent that addresses LLM limitations in medical domain through specialized knowledge synthesis and retrieval tools, achieving state-of-the-art performance on medical benchmarks.

DetailsMotivation: General-purpose LLM agents struggle with medical domain challenges due to insufficient medical knowledge and lack of specialized retrieval tools for medical contexts.

Method: Two core innovations: 1) Data synthesis framework using medical knowledge graphs to generate complex multi-hop QA pairs, 2) Integration of custom private medical retrieval engine with general-purpose tools. Two-stage training with supervised fine-tuning and online reinforcement learning.

Result: Generated 2100+ diverse trajectories across 12 medical specialties, MedResearcher-R1-32B model established new state-of-the-art results on medical benchmarks while maintaining competitive general performance.

Conclusion: Strategic domain-specific innovations in architecture, tool design, and training data enable smaller open-source models to outperform larger proprietary systems in specialized domains like medicine.

Abstract: Recent developments in Large Language Model (LLM)-based agents have shown impressive capabilities spanning multiple domains, exemplified by deep research systems that demonstrate superior performance on complex information-seeking and synthesis tasks. While general-purpose deep research agents have shown impressive capabilities, they struggle significantly with medical domain challenges, as evidenced by leading proprietary systems achieving limited accuracy on complex medical benchmarks. The key limitations are: (1) the model lacks sufficient dense medical knowledge for clinical reasoning, and (2) the framework is constrained by the absence of specialized retrieval tools tailored for medical contexts. We present a medical deep research agent that addresses these challenges through two core innovations. First, we develop a novel data synthesis framework using medical knowledge graphs, extracting the longest chains from subgraphs around rare medical entities to generate complex multi-hop question-answer pairs. Second, we integrate a custom-built private medical retrieval engine alongside general-purpose tools, enabling accurate medical information synthesis. Our approach generates 2100+ diverse trajectories across 12 medical specialties, each averaging 4.2 tool interactions. Through a two-stage training paradigm combining supervised fine-tuning and online reinforcement learning with composite rewards, our MedResearcher-R1-32B model demonstrates exceptional performance, establishing new state-of-the-art results on medical benchmarks while maintaining competitive performance on general deep research tasks. Our work demonstrates that strategic domain-specific innovations in architecture, tool design, and training data construction can enable smaller open-source models to outperform much larger proprietary systems in specialized domains.

[219] Reward-Shifted Speculative Sampling Is An Efficient Test-Time Weak-to-Strong Aligner

Bolian Li, Yanran Wu, Xinyu Luo, Ruqi Zhang

Main category: cs.CL

TL;DR: Reward-shifted speculative sampling algorithm uses aligned draft model to achieve test-time alignment efficiency without modifying target model, reducing inference costs while maintaining performance.

DetailsMotivation: Test-time alignment techniques for LLMs incur substantial inference costs, limiting practical application. Need efficient method to align models with human preferences during inference.

Method: Propose reward-shifted speculative sampling (SSS) algorithm where draft model is aligned with human preferences while target model remains unchanged. Modify acceptance criterion and bonus token distribution to exploit distributional shift.

Result: Achieves superior gold reward scores at significantly reduced inference cost in test-time weak-to-strong alignment experiments.

Conclusion: The algorithm effectively addresses efficiency bottleneck of test-time alignment while maintaining alignment quality, validating both effectiveness and efficiency.

Abstract: Aligning large language models (LLMs) with human preferences has become a critical step in their development. Recent research has increasingly focused on test-time alignment, where additional compute is allocated during inference to enhance LLM safety and reasoning capabilities. However, these test-time alignment techniques often incur substantial inference costs, limiting their practical application. We are inspired by the speculative sampling acceleration, which leverages a small draft model to efficiently predict future tokens, to address the efficiency bottleneck of test-time alignment. We introduce the reward-shifted speculative sampling (SSS) algorithm, in which the draft model is aligned with human preferences, while the target model remains unchanged. We theoretically demonstrate that the distributional shift between the aligned draft model and the unaligned target model can be exploited to recover the RLHF optimal solution without actually obtaining it, by modifying the acceptance criterion and bonus token distribution. Our algorithm achieves superior gold reward scores at a significantly reduced inference cost in test-time weak-to-strong alignment experiments, thereby validating both its effectiveness and efficiency.

[220] SparK: Query-Aware Unstructured Sparsity with Recoverable KV Cache Channel Pruning

Huanxuan Liao, Yixing Xu, Shizhu He, Guanchen Li, Xuanwu Yin, Dong Li, Emad Barsoum, Jun Zhao, Kang Liu

Main category: cs.CL

TL;DR: SPARK is a training-free KV cache pruning method that applies channel-level sparsity to reduce memory usage while maintaining model accuracy, enabling longer sequence processing within the same memory budget.

DetailsMotivation: Long-context LLM inference faces KV cache bottlenecks where memory grows linearly and attention computation scales quadratically with sequence length. Existing methods focus on temporal compression but ignore fine-grained channel-level importance variations.

Method: Proposes SPARK - applies unstructured sparsity by pruning KV cache at channel level, dynamically restoring pruned entries during attention computation. Training-free plug-and-play approach orthogonal to existing compression techniques.

Result: Reduces KV cache storage by over 30% compared to eviction methods while preserving/improving accuracy. With 80% pruning ratio, maintains performance with less than 5% degradation vs baseline. Enables longer sequences within same memory budget.

Conclusion: SPARK effectively addresses KV cache bottleneck through channel-level sparsity, demonstrating robust performance preservation while significantly reducing memory requirements, and is compatible with existing compression techniques for further acceleration.

Abstract: Long-context inference in large language models (LLMs) is increasingly constrained by the KV cache bottleneck: memory usage grows linearly with sequence length, while attention computation scales quadratically. Existing approaches address this issue by compressing the KV cache along the temporal axis through strategies such as token eviction or merging to reduce memory and computational overhead. However, these methods often neglect fine-grained importance variations across feature dimensions (i.e., the channel axis), thereby limiting their ability to effectively balance efficiency and model accuracy. In reality, we observe that channel saliency varies dramatically across both queries and positions: certain feature channels carry near-zero information for a given query, while others spike in relevance. To address this oversight, we propose SPARK, a training-free plug-and-play method that applies unstructured sparsity by pruning KV at the channel level, while dynamically restoring the pruned entries during attention score computation. Notably, our approach is orthogonal to existing KV compression and quantization techniques, making it compatible for integration with them to achieve further acceleration. By reducing channel-level redundancy, SPARK enables processing of longer sequences within the same memory budget. For sequences of equal length, SPARK not only preserves or improves model accuracy but also reduces KV cache storage by over 30% compared to eviction-based methods. Furthermore, even with an aggressive pruning ratio of 80%, SPARK maintains performance with less degradation than 5% compared to the baseline eviction method, demonstrating its robustness and effectiveness. Our code will be available at https://github.com/Xnhyacinth/SparK.

[221] Enhancing Cryptocurrency Sentiment Analysis with Multimodal Features

Chenghao Liu, Aniket Mahanti, Ranesh Naha, Guanghao Wang, Erwann Sbai

Main category: cs.CL

TL;DR: Multimodal analysis comparing TikTok (video) and Twitter (text) sentiment for cryptocurrency markets using LLMs, showing TikTok influences short-term trends while Twitter aligns with long-term dynamics, with cross-platform integration improving forecasting by 20%.

DetailsMotivation: Video content on platforms like TikTok remains underexplored despite containing richer emotional and contextual sentiment than text alone, while most prior research focused only on text-based platforms like Twitter for cryptocurrency market analysis.

Method: Used large language models to extract insights from both video (TikTok) and text (Twitter) data, investigating dynamic dependencies and spillover effects between social media sentiment and cryptocurrency market indicators.

Result: TikTok’s video-based sentiment significantly influences speculative assets and short-term market trends, while Twitter’s text-based sentiment aligns more closely with long-term dynamics. Cross-platform sentiment integration improves forecasting accuracy by up to 20%.

Conclusion: Video-based social media platforms like TikTok provide valuable sentiment signals distinct from text-based platforms, and integrating multimodal sentiment data from both video and text sources significantly enhances cryptocurrency market forecasting capabilities.

Abstract: As cryptocurrencies gain popularity, the digital asset marketplace becomes increasingly significant. Understanding social media signals offers valuable insights into investor sentiment and market dynamics. Prior research has predominantly focused on text-based platforms such as Twitter. However, video content remains underexplored, despite potentially containing richer emotional and contextual sentiment that is not fully captured by text alone. In this study, we present a multimodal analysis comparing TikTok and Twitter sentiment, using large language models to extract insights from both video and text data. We investigate the dynamic dependencies and spillover effects between social media sentiment and cryptocurrency market indicators. Our results reveal that TikTok’s video-based sentiment significantly influences speculative assets and short-term market trends, while Twitter’s text-based sentiment aligns more closely with long-term dynamics. Notably, the integration of cross-platform sentiment signals improves forecasting accuracy by up to 20%.

[222] ObjexMT: Objective Extraction and Metacognitive Calibration for LLM-as-a-Judge under Multi-Turn Jailbreaks

Hyunjun Kim, Junwoo Ha, Sangyoon Yu, Haon Park

Main category: cs.CL

TL;DR: ObjexMT benchmark tests LLM judges’ ability to extract latent objectives from multi-turn conversations and assess their own confidence, revealing that current models often misinfer objectives with high confidence.

DetailsMotivation: LLM-as-a-Judge evaluation lacks a decisive test for qualification - whether models can recover latent conversation objectives and know when their inferences are trustworthy, especially given challenges with irrelevant context, long context, and multi-turn jailbreaks.

Method: ObjexMT benchmark requires models to return a one-sentence base objective and self-reported confidence from multi-turn transcripts. Accuracy is measured via LLM-judge semantic similarity to gold objectives with human-aligned threshold calibration. Metacognition is evaluated with ECE, Brier, Wrong-at-High-Conf, and risk-coverage metrics.

Result: Claude-sonnet-4 achieved best objective-extraction accuracy (0.515) and calibration (ECE 0.296; Brier 0.324). GPT-4.1 and Qwen3-235B-A22B-FP8 tied at 0.441 accuracy but were overconfident (mean confidence ≈0.88 vs accuracy ≈0.44; Wrong-at-0.90 ≈48-52%). Performance varied significantly across datasets (≈0.167-0.865).

Conclusion: LLM judges often misinfer objectives with high confidence when objectives are not explicit. The paper recommends exposing objectives when feasible and gating decisions by confidence otherwise. ObjexMT provides an actionable test for LLM judge qualification.

Abstract: LLM-as-a-Judge (LLMaaJ) now underpins scalable evaluation, yet we lack a decisive test of a judge’s qualification: can it recover a conversation’s latent objective and know when that inference is trustworthy? LLMs degrade under irrelevant or long context; multi-turn jailbreaks further hide goals across turns. We introduce ObjexMT, a benchmark for objective extraction and metacognition. Given a multi-turn transcript, a model must return a one-sentence base objective and a self-reported confidence. Accuracy is computed via LLM-judge semantic similarity to gold objectives, converted to binary correctness by a single human-aligned threshold calibrated once on N = 100 items ($\tau^*=0.61$). Metacognition is evaluated with ECE, Brier, Wrong-at-High-Conf, and risk-coverage. Across gpt-4.1, claude-sonnet-4, and Qwen3-235B-A22B-FP8 on SafeMTData_Attack600, SafeMTData_1K, MHJ, and CoSafe, claude-sonnet-4 attains the best objective-extraction accuracy (0.515) and calibration (ECE 0.296; Brier 0.324); gpt-4.1 and Qwen3-235B-A22B-FP8 tie at 0.441 but are overconfident (mean confidence $\approx$0.88 vs. accuracy $\approx$0.44; Wrong-at-0.90 $\approx$48-52%). Performance varies by dataset ($\approx$0.167-0.865). ObjexMT thus supplies an actionable test for LLM judges: when objectives are not explicit, judges often misinfer them with high confidence. We recommend exposing objectives when feasible and gating decisions by confidence otherwise. Code and data at https://github.com/hyunjun1121/ObjexMT_dataset.

[223] Text Meets Topology: Rethinking Out-of-distribution Detection in Text-Rich Networks

Danny Wang, Ruihong Qiu, Guangdong Bai, Zi Huang

Main category: cs.CL

TL;DR: TextTopoOOD framework for OOD detection in text-rich networks addresses complex textual-structural diversity through four scenarios and proposes TNT-OOD model with cross-attention and HyperNetwork for better ID/OOD distinction.

DetailsMotivation: Existing OOD detection methods in text-rich networks overlook the intricate interplay between textual features and topological structures, failing to address diverse OOD scenarios like attribute-level shifts, structural shifts, and thematic label shifts.

Method: Proposes TextTopoOOD framework with four OOD scenarios: attribute-level shifts via text augmentations, structural shifts through edge rewiring, thematically-guided label shifts, and domain-based divisions. Introduces TNT-OOD model with cross-attention module to fuse local structure into text representations and HyperNetwork for node-specific transformations.

Result: Experiments on 11 datasets across four OOD scenarios demonstrate the framework’s effectiveness in evaluating OOD detection challenges in text-rich networks.

Conclusion: The proposed TextTopoOOD framework and TNT-OOD model successfully address the complex interplay between text and topology in OOD detection, providing a comprehensive evaluation approach for diverse OOD scenarios in text-rich networks.

Abstract: Out-of-distribution (OOD) detection remains challenging in text-rich networks, where textual features intertwine with topological structures. Existing methods primarily address label shifts or rudimentary domain-based splits, overlooking the intricate textual-structural diversity. For example, in social networks, where users represent nodes with textual features (name, bio) while edges indicate friendship status, OOD may stem from the distinct language patterns between bot and normal users. To address this gap, we introduce the TextTopoOOD framework for evaluating detection across diverse OOD scenarios: (1) attribute-level shifts via text augmentations and embedding perturbations; (2) structural shifts through edge rewiring and semantic connections; (3) thematically-guided label shifts; and (4) domain-based divisions. Furthermore, we propose TNT-OOD to model the complex interplay between Text aNd Topology using: 1) a novel cross-attention module to fuse local structure into node-level text representations, and 2) a HyperNetwork to generate node-specific transformation parameters. This aligns topological and semantic features of ID nodes, enhancing ID/OOD distinction across structural and textual shifts. Experiments on 11 datasets across four OOD scenarios demonstrate the nuanced challenge of TextTopoOOD for evaluating OOD detection in text-rich networks.

[224] Affective Polarization across European Parliaments

Bojan Evkoski, Igor Mozetič, Nikola Ljubešić, Petra Kralj Novak

Main category: cs.CL

TL;DR: Automated analysis of parliamentary speeches in 6 European countries reveals consistent affective polarization, with parliamentarians showing more negativity towards opposing groups than their own, driven by reciprocity mechanisms.

DetailsMotivation: To examine the presence and patterns of affective polarization in European parliaments using automated natural language processing techniques on parliamentary speeches.

Method: Used comprehensive corpus of parliamentary speeches from six European countries, employed NLP techniques to estimate parliamentarian sentiment, compared negativity levels in references to opposing groups vs own groups.

Result: Found consistent affective polarization across all six European parliaments. Activity correlates with negativity but no difference in polarization between less and more active MPs. Reciprocity is a contributing mechanism in affective polarization.

Conclusion: Affective polarization is a consistent feature across European parliaments, with reciprocity playing a key role in driving negative interactions between opposing political groups.

Abstract: Affective polarization, characterized by increased negativity and hostility towards opposing groups, has become a prominent feature of political discourse worldwide. Our study examines the presence of this type of polarization in a selection of European parliaments in a fully automated manner. Utilizing a comprehensive corpus of parliamentary speeches from the parliaments of six European countries, we employ natural language processing techniques to estimate parliamentarian sentiment. By comparing the levels of negativity conveyed in references to individuals from opposing groups versus one’s own, we discover patterns of affectively polarized interactions. The findings demonstrate the existence of consistent affective polarization across all six European parliaments. Although activity correlates with negativity, there is no observed difference in affective polarization between less active and more active members of parliament. Finally, we show that reciprocity is a contributing mechanism in affective polarization between parliamentarians across all six parliaments.

[225] Continuously Steering LLMs Sensitivity to Contextual Knowledge with Proxy Models

Yilin Wang, Heng Wang, Yuyang Bai, Minnan Luo

Main category: cs.CL

TL;DR: CSKS is a lightweight framework that enables continuous control over LLMs’ sensitivity to contextual knowledge without modifying model weights, using two small proxy models to shift output distributions.

DetailsMotivation: Address knowledge conflicts in LLMs where parametric knowledge contradicts contextual knowledge, overcoming limitations of previous methods that are inefficient, ineffective for large models, or not workable for black-box models.

Method: Tune two small proxy models and use the difference in their output distributions to shift the original LLM’s distribution without weight modification, allowing continuous sensitivity adjustment.

Result: Achieves continuous and precise control over LLMs’ sensitivity to contextual knowledge, enabling both increased and reduced sensitivity, allowing flexible prioritization of contextual or parametric knowledge.

Conclusion: CSKS provides an effective, lightweight solution for steering LLMs’ knowledge sensitivity without model modification, validated through synthetic data and real conflict datasets.

Abstract: In Large Language Models (LLMs) generation, there exist knowledge conflicts and scenarios where parametric knowledge contradicts knowledge provided in the context. Previous works studied tuning, decoding algorithms, or locating and editing context-aware neurons to adapt LLMs to be faithful to new contextual knowledge. However, they are usually inefficient or ineffective for large models, not workable for black-box models, or unable to continuously adjust LLMs’ sensitivity to the knowledge provided in the context. To mitigate these problems, we propose CSKS (Continuously Steering Knowledge Sensitivity), a simple framework that can steer LLMs’ sensitivity to contextual knowledge continuously at a lightweight cost. Specifically, we tune two small LMs (i.e. proxy models) and use the difference in their output distributions to shift the original distribution of an LLM without modifying the LLM weights. In the evaluation process, we not only design synthetic data and fine-grained metrics to measure models’ sensitivity to contextual knowledge but also use a real conflict dataset to validate CSKS’s practical efficacy. Extensive experiments demonstrate that our framework achieves continuous and precise control over LLMs’ sensitivity to contextual knowledge, enabling both increased sensitivity and reduced sensitivity, thereby allowing LLMs to prioritize either contextual or parametric knowledge as needed flexibly. Our data and code are available at https://github.com/OliveJuiceLin/CSKS.

[226] How Can Input Reformulation Improve Tool Usage Accuracy in a Complex Dynamic Environment? A Study on $τ$-bench

Venkatesh Mishra, Amir Saeidi, Satyam Raj, Mutsumi Nakamura, Jayanth Srinivasa, Gaowen Liu, Ali Payani, Chitta Baral

Main category: cs.CL

TL;DR: IRMA framework improves LLM tool-use agents by automatically reformulating queries with domain rules and tool suggestions, achieving significant performance gains over existing methods.

DetailsMotivation: LLM-based agents struggle with consistent reasoning, policy adherence, and information extraction in multi-turn conversational environments with tool calls.

Method: Proposed Input-Reformulation Multi-Agent (IRMA) framework that automatically reformulates user queries augmented with relevant domain rules and tool suggestions.

Result: IRMA outperforms ReAct, Function Calling, and Self-Reflection by 16.1%, 12.7%, and 19.1% respectively in overall pass scores.

Conclusion: IRMA demonstrates superior reliability and consistency compared to other methods in dynamic environments requiring tool use.

Abstract: Recent advances in reasoning and planning capabilities of large language models (LLMs) have enabled their potential as autonomous agents capable of tool use in dynamic environments. However, in multi-turn conversational environments like $\tau$-bench, these agents often struggle with consistent reasoning, adherence to domain-specific policies, and extracting correct information over a long horizon of tool-calls and conversation. To capture and mitigate these failures, we conduct a comprehensive manual analysis of the common errors occurring in the conversation trajectories. We then experiment with reformulations of inputs to the tool-calling agent for improvement in agent decision making. Finally, we propose the Input-Reformulation Multi-Agent (IRMA) framework, which automatically reformulates user queries augmented with relevant domain rules and tool suggestions for the tool-calling agent to focus on. The results show that IRMA significantly outperforms ReAct, Function Calling, and Self-Reflection by 16.1%, 12.7%, and 19.1%, respectively, in overall pass^5 scores. These findings highlight the superior reliability and consistency of IRMA compared to other methods in dynamic environments.

[227] Re-Representation in Sentential Relation Extraction with Sequence Routing Algorithm

Ramazan Ali Bahrami, Ramin Yahyapour

Main category: cs.CL

TL;DR: Proposes capsule networks with dynamic routing for sentential relation extraction, achieving SOTA on common datasets but identifying noise in Wikidata labels and re-representation as key challenges.

DetailsMotivation: To improve sentential relation extraction performance using dynamic routing in capsules and understand why models perform well on some datasets but poorly on others like Wikidata.

Method: Uses capsule networks with dynamic routing mechanism for sentential relation extraction, tested on multiple datasets including Tacred, Tacredrev, Retacred, Conll04, and Wikidata.

Result: Outperforms state-of-the-art on common datasets but shows low performance on Wikidata due to label noise. Demonstrates better re-representation capability compared to vanilla models.

Conclusion: Noise in distantly supervised datasets and re-representation capability are significant challenges in sentential relation extraction that need to be addressed.

Abstract: Sentential relation extraction (RE) is an important task in natural language processing (NLP). In this paper we propose to do sentential RE with dynamic routing in capsules. We first show that the proposed approach outperform state of the art on common sentential relation extraction datasets Tacred, Tacredrev, Retacred, and Conll04. We then investigate potential reasons for its good performance on the mentioned datasets, and yet low performance on another similar, yet larger sentential RE dataset, Wikidata. As such, we identify noise in Wikidata labels as one of the reasons that can hinder performance. Additionally, we show associativity of better performance with better re-representation, a term from neuroscience referred to change of representation in human brain to improve the match at comparison time. As example, in the given analogous terms King:Queen::Man:Woman, at comparison time, and as a result of re-representation, the similarity between related head terms (King,Man), and tail terms (Queen,Woman) increases. As such, our observation show that our proposed model can do re-representation better than the vanilla model compared with. To that end, beside noise in the labels of the distantly supervised RE datasets, we propose re-representation as a challenge in sentential RE.

[228] Language Models and Logic Programs for Trustworthy Financial Reasoning

William Jurayj, Nils Holzenberger, Benjamin Van Durme

Main category: cs.CL

TL;DR: A neuro-symbolic approach combining LLMs with symbolic solvers achieves high accuracy in tax filing tasks, reducing costs below real-world averages while ensuring auditability.

DetailsMotivation: Tax filing requires complex reasoning with high accuracy to avoid costly penalties, making pure LLMs unsuitable due to reliability and auditability concerns.

Method: Integrates LLMs with symbolic solvers, translates plain-text tax rules into formal logic programs, and uses intelligently retrieved exemplars for formal case representations.

Result: Dramatically improved performance on the SARA dataset with costs reduced well below real-world averages of $270 and 13 hours per filing.

Conclusion: Neuro-symbolic architectures show promise for equitable access to reliable tax assistance by combining the strengths of LLMs and symbolic reasoning.

Abstract: According to the United States Internal Revenue Service, ‘’the average American spends $$270$ and 13 hours filing their taxes’’. Even beyond the U.S., tax filing requires complex reasoning, combining application of overlapping rules with numerical calculations. Because errors can incur costly penalties, any automated system must deliver high accuracy and auditability, making modern large language models (LLMs) poorly suited for this task. We propose an approach that integrates LLMs with a symbolic solver to calculate tax obligations. We evaluate variants of this system on the challenging StAtutory Reasoning Assessment (SARA) dataset, and include a novel method for estimating the cost of deploying such a system based on real-world penalties for tax errors. We further show how combining up-front translation of plain-text rules into formal logic programs, combined with intelligently retrieved exemplars for formal case representations, can dramatically improve performance on this task and reduce costs to well below real-world averages. Our results demonstrate the promise and economic feasibility of neuro-symbolic architectures for increasing equitable access to reliable tax assistance.

[229] Quantifying Label-Induced Bias in Large Language Model Self- and Cross-Evaluations

Muskan Saraf, Sajjad Rezvani Boroujeni, Justin Beaudry, Hossein Abedi, Tom Bush

Main category: cs.CL

TL;DR: LLM evaluation bias study shows model identity labels significantly influence scoring, with Claude labels boosting scores and Gemini labels depressing them, causing up to 50 percentage point shifts in preference rankings.

DetailsMotivation: To investigate how model identity labels influence LLM evaluation judgments and whether self- and cross-model evaluations are biased by perceived authorship.

Method: Tested ChatGPT, Gemini, and Claude evaluating blog posts under four label conditions (no labels, true labels, two false labels) using preference voting and quality ratings (Coherence, Informativeness, Conciseness) expressed as percentages.

Result: Strong asymmetrical bias found - Claude labels consistently boosted scores while Gemini labels depressed them. False labels frequently reversed rankings with 50pp preference vote shifts and 12pp quality rating changes. Gemini’s self-scores collapsed under true labels while Claude’s self-preference intensified.

Conclusion: Perceived model identity heavily distorts LLM evaluation judgments, requiring blind or multimodel evaluation protocols to ensure fair benchmarking.

Abstract: Large language models (LLMs) are increasingly used to evaluate outputs, yet their judgments may be influenced. This study examines bias in self- and cross-model evaluations by ChatGPT, Gemini, and Claude under four conditions: no labels, true labels, and two false-label scenarios. Blog posts authored by each model were evaluated by all three using both overall preference voting and quality ratings for Coherence, Informativeness, and Conciseness, with all scores expressed as percentages for direct comparison. Results reveal striking asymmetries: the “Claude” label consistently boosts scores, while the “Gemini” label consistently depresses them, regardless of actual content. False labels frequently reversed rankings, producing shifts of up to 50 percentage points in preference votes and up to 12 percentage points in converted quality ratings. Gemini’s self-scores collapsed under true labels, while Claude’s self-preference intensified. These findings show that perceived model identity can heavily distort high-level judgments and subtly influence detailed quality ratings, underscoring the need for blind or multimodel evaluation protocols to ensure fairness in LLM benchmarking.

cs.CV

[230] Deep Learning-Driven Multimodal Detection and Movement Analysis of Objects in Culinary

Tahoshin Alam Ishat

Main category: cs.CV

TL;DR: Fine-tuned YOLOv8 segmentation, LSTM hand motion analysis, and Whisper ASR to extract cooking data for TinyLLaMa to generate step-by-step recipes

DetailsMotivation: Extend computer vision applications to daily kitchen activities by creating a robust system that can analyze cooking procedures and generate instructional guides

Method: Combined YOLOv8 for object segmentation, LSTM for hand motion sequence analysis, and Whisper ASR for audio processing to extract comprehensive cooking data, then used TinyLLaMa LLM for recipe prediction and text generation

Result: Developed a task-specific system that performs well in complex kitchen environments, demonstrating practical computer vision applications for daily life activities

Conclusion: This work successfully extends computer vision capabilities to kitchen tasks and shows potential for many other crucial daily life applications through multimodal AI integration

Abstract: This is a research exploring existing models and fine tuning them to combine a YOLOv8 segmentation model, a LSTM model trained on hand point motion sequence and a ASR (whisper-base) to extract enough data for a LLM (TinyLLaMa) to predict the recipe and generate text creating a step by step guide for the cooking procedure. All the data were gathered by the author for a robust task specific system to perform best in complex and challenging environments proving the extension and endless application of computer vision in daily activities such as kitchen work. This work extends the field for many more crucial task of our day to day life.

[231] AMMKD: Adaptive Multimodal Multi-teacher Distillation for Lightweight Vision-Language Models

Yuqi Li, Chuanguang Yang, Junhao Dong, Zhengtao Yao, Haoyan Xu, Zeyu Dong, Hansheng Zeng, Zhulin An, Yingli Tian

Main category: cs.CV

TL;DR: AMMKD is a knowledge distillation framework that creates lightweight image-text retrieval models by combining multi-modal feature fusion, multi-teacher distillation with pre-computed text features, and adaptive optimization to handle teacher conflicts.

DetailsMotivation: Large VLP models are too computationally expensive for mobile deployment, requiring lightweight yet effective alternatives for image-text retrieval tasks.

Method: Uses feature fusion network, multi-teacher distillation with pre-computed text features from CLIP teachers, KL scatter for distribution matching, and adaptive dynamic weighting based on gradient space diversity.

Result: Achieves superior performance on three benchmark datasets while significantly reducing model complexity.

Conclusion: AMMKD provides an effective and flexible solution for deploying lightweight image-text retrieval models on resource-constrained devices.

Abstract: The success of large-scale visual language pretraining (VLP) models has driven widespread adoption of image-text retrieval tasks. However, their deployment on mobile devices remains limited due to large model sizes and computational complexity. We propose Adaptive Multi-Modal Multi-Teacher Knowledge Distillation (AMMKD), a novel framework that integrates multi-modal feature fusion, multi-teacher distillation, and adaptive optimization to deliver lightweight yet effective retrieval models. Specifically, our method begins with a feature fusion network that extracts and merges discriminative features from both the image and text modalities. To reduce model parameters and further improve performance, we design a multi-teacher knowledge distillation framework to pre-train two CLIP teacher models. We decouple modalities by pre-computing and storing text features as class vectors via the teacher text encoder to enhance efficiency. To better align teacher and student outputs, we apply KL scatter for probability distribution matching. Finally, we design an adaptive dynamic weighting scheme that treats multi-teacher distillation as a multi-objective optimization problem. By leveraging gradient space diversity, we dynamically adjust the influence of each teacher, reducing conflicts and guiding the student toward more optimal learning directions. Extensive experiments on three benchmark datasets demonstrate that AMMKD achieves superior performance while significantly reducing model complexity, validating its effectiveness and flexibility.

[232] ARTPS: Depth-Enhanced Hybrid Anomaly Detection and Learnable Curiosity Score for Autonomous Rover Target Prioritization

Poyraz Baydemir

Main category: cs.CV

TL;DR: ARTPS is a hybrid AI system for autonomous planetary exploration that combines depth estimation, anomaly detection, and learnable curiosity scoring, achieving state-of-the-art performance with AUROC 0.94 and reducing false positives by 23%.

DetailsMotivation: To develop an autonomous exploration system for planetary surfaces that can effectively prioritize targets by integrating multiple AI components for improved accuracy and reduced false positives.

Method: Hybrid approach combining monocular depth estimation using Vision Transformers, multi-component anomaly detection, and weighted curiosity scoring that balances known value, anomaly signals, depth variance, and surface roughness.

Result: State-of-the-art performance with AUROC of 0.94, AUPRC of 0.89, F1-Score of 0.87 on Mars rover datasets, and 23% reduction in false positives while maintaining high detection sensitivity across diverse terrain types.

Conclusion: The hybrid fusion approach demonstrates significant improvements in target prioritization accuracy and provides a comprehensive framework for autonomous planetary exploration with reduced false positive rates.

Abstract: We present ARTPS (Autonomous Rover Target Prioritization System), a novel hybrid AI system that combines depth estimation, anomaly detection, and learnable curiosity scoring for autonomous exploration of planetary surfaces. Our approach integrates monocular depth estimation using Vision Transformers with multi-component anomaly detection and a weighted curiosity score that balances known value, anomaly signals, depth variance, and surface roughness. The system achieves state-of-the-art performance with AUROC of 0.94, AUPRC of 0.89, and F1-Score of 0.87 on Mars rover datasets. We demonstrate significant improvements in target prioritization accuracy through ablation studies and provide comprehensive analysis of component contributions. The hybrid fusion approach reduces false positives by 23% while maintaining high detection sensitivity across diverse terrain types.

[233] Performance is not All You Need: Sustainability Considerations for Algorithms

Xiang Li, Chong Zhang, Hongpeng Wang, Shreyank Narayana Gowda, Yushi Li, Xiaobo Jin

Main category: cs.CV

TL;DR: Proposes a two-dimensional sustainability evaluation system with FMS and ASC metrics to balance AI performance and energy consumption, validated across multiple multimodal tasks.

DetailsMotivation: Address high carbon emissions from deep learning training by creating quantitative metrics that integrate energy efficiency with algorithm performance, moving beyond traditional performance-only evaluation.

Method: Developed two novel metrics: Sustainable Harmonic Mean (FMS) that combines accumulated energy consumption and performance via harmonic mean, and Area Under Sustainability Curve (ASC) that constructs performance-power consumption curves. Tested across image classification, segmentation, pose estimation, and batch/online learning tasks.

Result: The evaluation system successfully provides quantitative basis for cross-task algorithm assessment and facilitates the transition of green AI research from theory to practice. The framework code is made available for industry adoption.

Conclusion: The proposed sustainability evaluation system offers a practical approach to measure and compare algorithm energy efficiency, supporting the establishment of industry standards for green AI and promoting environmentally conscious deep learning practices.

Abstract: This work focuses on the high carbon emissions generated by deep learning model training, specifically addressing the core challenge of balancing algorithm performance and energy consumption. It proposes an innovative two-dimensional sustainability evaluation system. Different from the traditional single performance-oriented evaluation paradigm, this study pioneered two quantitative indicators that integrate energy efficiency ratio and accuracy: the sustainable harmonic mean (FMS) integrates accumulated energy consumption and performance parameters through the harmonic mean to reveal the algorithm performance under unit energy consumption; the area under the sustainability curve (ASC) constructs a performance-power consumption curve to characterize the energy efficiency characteristics of the algorithm throughout the cycle. To verify the universality of the indicator system, the study constructed benchmarks in various multimodal tasks, including image classification, segmentation, pose estimation, and batch and online learning. Experiments demonstrate that the system can provide a quantitative basis for evaluating cross-task algorithms and promote the transition of green AI research from theory to practice. Our sustainability evaluation framework code can be found here, providing methodological support for the industry to establish algorithm energy efficiency standards.

[234] MESTI-MEGANet: Micro-expression Spatio-Temporal Image and Micro-expression Gradient Attention Networks for Micro-expression Recognition

Luu Tu Nguyen, Vu Tram Anh Khuong, Thanh Ha Le, Thi Duyen Ngo

Main category: cs.CV

TL;DR: Proposes MESTI (novel dynamic image modality) and MEGANet (with Gradient Attention block) for micro-expression recognition, achieving state-of-the-art results on CASMEII and SAMM datasets.

DetailsMotivation: Traditional input modalities like Apex Frame, Optical Flow, and Dynamic Image fail to adequately capture subtle and fleeting micro-expressions, leading to suboptimal performance in micro-expression recognition.

Method: Introduces MESTI (Micro-expression Spatio-Temporal Image) that transforms video sequences into single images while preserving micro-movement characteristics. Presents MEGANet with a novel Gradient Attention block to enhance fine-grained motion feature extraction.

Result: MESTI outperforms existing input modalities across three CNN architectures. Replacing inputs of previous MER networks with MESTI consistently improves performance. MEGANet achieves state-of-the-art results on CASMEII and SAMM datasets, with the MEGANet+MESTI combination setting the highest accuracy benchmark.

Conclusion: MESTI is a superior input modality and MEGANet is an advanced recognition network that paves the way for more effective micro-expression recognition systems in various applications.

Abstract: Micro-expression recognition (MER) is a challenging task due to the subtle and fleeting nature of micro-expressions. Traditional input modalities, such as Apex Frame, Optical Flow, and Dynamic Image, often fail to adequately capture these brief facial movements, resulting in suboptimal performance. In this study, we introduce the Micro-expression Spatio-Temporal Image (MESTI), a novel dynamic input modality that transforms a video sequence into a single image while preserving the essential characteristics of micro-movements. Additionally, we present the Micro-expression Gradient Attention Network (MEGANet), which incorporates a novel Gradient Attention block to enhance the extraction of fine-grained motion features from micro-expressions. By combining MESTI and MEGANet, we aim to establish a more effective approach to MER. Extensive experiments were conducted to evaluate the effectiveness of MESTI, comparing it with existing input modalities across three CNN architectures (VGG19, ResNet50, and EfficientNetB0). Moreover, we demonstrate that replacing the input of previously published MER networks with MESTI leads to consistent performance improvements. The performance of MEGANet, both with MESTI and Dynamic Image, is also evaluated, showing that our proposed network achieves state-of-the-art results on the CASMEII and SAMM datasets. The combination of MEGANet and MESTI achieves the highest accuracy reported to date, setting a new benchmark for micro-expression recognition. These findings underscore the potential of MESTI as a superior input modality and MEGANet as an advanced recognition network, paving the way for more effective MER systems in a variety of applications.

[235] PRINTER:Deformation-Aware Adversarial Learning for Virtual IHC Staining with In Situ Fidelity

Yizhe Yuan, Bingsen Xue, Bangzheng Pu, Chengxiang Wang, Cheng Jin

Main category: cs.CV

TL;DR: PRINTER is a weakly-supervised framework that addresses spatial misalignment in tumor analysis by decoupling content and staining patterns, using prototype-driven transfer and deformation-aware adversarial learning to achieve accurate virtual IHC staining while preserving H&E details.

DetailsMotivation: Current methods for correlating H&E morphology with IHC biomarker expression suffer from spatial misalignment in consecutive sections, which compromises pathological interpretation of tumor spatial heterogeneity.

Method: A weakly-supervised framework with three innovations: 1) Prototype-driven staining pattern transfer with content-style decoupling, 2) GapBridge cyclic registration-synthesis framework for deformable structural alignment between H&E and IHC domains, and 3) Deformation-aware adversarial learning where generator and registration network jointly optimize a style-focused discriminator.

Result: Extensive experiments show PRINTER achieves superior performance in preserving H&E staining details and virtual staining fidelity, outperforming state-of-the-art methods.

Conclusion: PRINTER provides a robust and scalable solution for virtual staining that advances computational pathology by enabling accurate correlation between H&E morphology and IHC biomarker expression without spatial misalignment issues.

Abstract: Tumor spatial heterogeneity analysis requires precise correlation between Hematoxylin and Eosin H&E morphology and immunohistochemical (IHC) biomarker expression, yet current methods suffer from spatial misalignment in consecutive sections, severely compromising in situ pathological interpretation. In order to obtain a more accurate virtual staining pattern, We propose PRINTER, a weakly-supervised framework that integrates PRototype-drIven content and staiNing patTERn decoupling and deformation-aware adversarial learning strategies designed to accurately learn IHC staining patterns while preserving H&E staining details. Our approach introduces three key innovations: (1) A prototype-driven staining pattern transfer with explicit content-style decoupling; and (2) A cyclic registration-synthesis framework GapBridge that bridges H&E and IHC domains through deformable structural alignment, where registered features guide cross-modal style transfer while synthesized outputs iteratively refine the registration;(3) Deformation-Aware Adversarial Learning: We propose a training framework where a generator and deformation-aware registration network jointly adversarially optimize a style-focused discriminator. Extensive experiments demonstrate that PRINTER effectively achieves superior performance in preserving H&E staining details and virtual staining fidelity, outperforming state-of-the-art methods. Our work provides a robust and scalable solution for virtual staining, advancing the field of computational pathology.

[236] Scaffold Diffusion: Sparse Multi-Category Voxel Structure Generation with Discrete Diffusion

Justin Jung

Main category: cs.CV

TL;DR: Scaffold Diffusion is a discrete diffusion language model that generates realistic sparse multi-category 3D voxel structures by treating voxels as tokens, overcoming challenges of cubic memory scaling and class imbalance in sparse voxel generation.

DetailsMotivation: Generating realistic sparse multi-category 3D voxel structures is challenging due to cubic memory scaling and significant class imbalance caused by sparsity. Existing methods struggle with these issues.

Method: Treats voxels as tokens and uses a discrete diffusion language model to generate 3D voxel structures, extending discrete diffusion beyond sequential domains to create spatially coherent 3D structures.

Result: Outperforms prior baselines and auto-regressive formulations, producing realistic and coherent structures even when trained on data with over 98% sparsity. Evaluated on Minecraft house structures from 3D-Craft dataset.

Conclusion: Discrete diffusion is a promising framework for 3D sparse voxel generative modeling, successfully addressing the challenges of sparsity and class imbalance in voxel structure generation.

Abstract: Generating realistic sparse multi-category 3D voxel structures is difficult due to the cubic memory scaling of voxel structures and moreover the significant class imbalance caused by sparsity. We introduce Scaffold Diffusion, a generative model designed for sparse multi-category 3D voxel structures. By treating voxels as tokens, Scaffold Diffusion uses a discrete diffusion language model to generate 3D voxel structures. We show that discrete diffusion language models can be extended beyond inherently sequential domains such as text to generate spatially coherent 3D structures. We evaluate on Minecraft house structures from the 3D-Craft dataset and demonstrate that, unlike prior baselines and an auto-regressive formulation, Scaffold Diffusion produces realistic and coherent structures even when trained on data with over 98% sparsity. We provide an interactive viewer where readers can visualize generated samples and the generation process. Our results highlight discrete diffusion as a promising framework for 3D sparse voxel generative modeling.

[237] Identity-Preserving Text-to-Video Generation via Training-Free Prompt, Image, and Guidance Enhancement

Jiayi Gao, Changcheng Hua, Qingchao Chen, Yuxin Peng, Yang Liu

Main category: cs.CV

TL;DR: TPIGE is a training-free framework that enhances text-to-video generation by mutually refining prompts and reference images, then using spatiotemporal guidance to improve identity preservation and video quality without costly fine-tuning.

DetailsMotivation: Address data scarcity and high tuning costs in identity-preserving text-to-video generation by developing a training-free approach that bridges semantic gaps between text prompts and reference images.

Method: Three-step framework: 1) Face Aware Prompt Enhancement using GPT-4o to add facial details, 2) Prompt Aware Reference Image Enhancement using identity-preserving image generator, 3) ID-Aware Spatiotemporal Guidance Enhancement using unified gradients during generation.

Result: Outperforms prior work, achieves state-of-the-art performance, wins first place in ACM Multimedia 2025 challenge, validated on 1000-video test set with automatic and human evaluations.

Conclusion: TPIGE framework provides significant performance gains at minimal cost, demonstrating strong generality and effectiveness for identity-preserving text-to-video generation without requiring expensive fine-tuning.

Abstract: Identity-preserving text-to-video (IPT2V) generation creates videos faithful to both a reference subject image and a text prompt. While fine-tuning large pretrained video diffusion models on ID-matched data achieves state-of-the-art results on IPT2V, data scarcity and high tuning costs hinder broader improvement. We thus introduce a Training-Free Prompt, Image, and Guidance Enhancement (TPIGE) framework that bridges the semantic gap between the video description and the reference image and design sampling guidance that enhances identity preservation and video quality, achieving performance gains at minimal cost.Specifically, we first propose Face Aware Prompt Enhancement, using GPT-4o to enhance the text prompt with facial details derived from the reference image. We then propose Prompt Aware Reference Image Enhancement, leveraging an identity-preserving image generator to refine the reference image, rectifying conflicts with the text prompt. The above mutual refinement significantly improves input quality before video generation. Finally, we propose ID-Aware Spatiotemporal Guidance Enhancement, utilizing unified gradients to optimize identity preservation and video quality jointly during generation.Our method outperforms prior work and is validated by automatic and human evaluations on a 1000 video test set, winning first place in the ACM Multimedia 2025 Identity-Preserving Video Generation Challenge, demonstrating state-of-the-art performance and strong generality. The code is available at https://github.com/Andyplus1/IPT2V.git.

[238] Dual-Stage Global and Local Feature Framework for Image Dehazing

Anas M. Ali, Anis Koubaa, Bilel Benjdira

Main category: cs.CV

TL;DR: Proposes SGLC framework for high-resolution image dehazing by combining global contextual information with local fine-grained details through GFG and LFE components, showing significant PSNR improvements.

DetailsMotivation: Existing dehazing models perform poorly on high-resolution images due to difficulty combining global context with local details, leading practitioners to use suboptimal downsampling or patching approaches.

Method: SGLC framework with two components: Global Features Generator (GFG) for broad contextual understanding and Local Features Enhancer (LFE) for refining localized details and pixel-level features. Integrated with Uformer architecture.

Result: Experimental results show considerable improvement in peak signal-to-noise ratio (PSNR) on high-resolution datasets, demonstrating effectiveness for large-scale imagery dehazing.

Conclusion: SGLC provides a model-agnostic solution that enables any dehazing network to effectively combine scene-level cues with granular details, significantly improving visual fidelity in high-resolution environments.

Abstract: Addressing the challenge of removing atmospheric fog or haze from digital images, known as image dehazing, has recently gained significant traction in the computer vision community. Although contemporary dehazing models have demonstrated promising performance, few have thoroughly investigated high-resolution imagery. In such scenarios, practitioners often resort to downsampling the input image or processing it in smaller patches, which leads to a notable performance degradation. This drop is primarily linked to the difficulty of effectively combining global contextual information with localized, fine-grained details as the spatial resolution grows. In this chapter, we propose a novel framework, termed the Streamlined Global and Local Features Combinator (SGLC), to bridge this gap and enable robust dehazing for high-resolution inputs. Our approach is composed of two principal components: the Global Features Generator (GFG) and the Local Features Enhancer (LFE). The GFG produces an initial dehazed output by focusing on broad contextual understanding of the scene. Subsequently, the LFE refines this preliminary output by enhancing localized details and pixel-level features, thereby capturing the interplay between global appearance and local structure. To evaluate the effectiveness of SGLC, we integrated it with the Uformer architecture, a state-of-the-art dehazing model. Experimental results on high-resolution datasets reveal a considerable improvement in peak signal-to-noise ratio (PSNR) when employing SGLC, indicating its potency in addressing haze in large-scale imagery. Moreover, the SGLC design is model-agnostic, allowing any dehazing network to be augmented with the proposed global-and-local feature fusion mechanism. Through this strategy, practitioners can harness both scene-level cues and granular details, significantly improving visual fidelity in high-resolution environments.

[239] Enhancing Partially Relevant Video Retrieval with Robust Alignment Learning

Long Zhang, Peipei Song, Jianfeng Dong, Kun Li, Xun Yang

Main category: cs.CV

TL;DR: RAL framework uses probabilistic modeling with Gaussian distributions and confidence gates to handle data uncertainty in partially relevant video retrieval, improving robustness against spurious correlations.

DetailsMotivation: Existing PRVR methods struggle with data uncertainty from query ambiguity and partial video relevance, making them vulnerable to distractor videos with spurious similarities.

Method: Proposes Robust Alignment Learning (RAL) with: 1) probabilistic modeling encoding videos/queries as multivariate Gaussian distributions for uncertainty quantification, 2) learnable confidence gates to dynamically weight query word similarity.

Result: RAL demonstrates effectiveness across diverse retrieval backbones as a plug-and-play solution through extensive experiments.

Conclusion: The framework successfully addresses data uncertainty in PRVR through probabilistic modeling and dynamic confidence weighting, improving retrieval robustness.

Abstract: Partially Relevant Video Retrieval (PRVR) aims to retrieve untrimmed videos partially relevant to a given query. The core challenge lies in learning robust query-video alignment against spurious semantic correlations arising from inherent data uncertainty: 1) query ambiguity, where the query incompletely characterizes the target video and often contains uninformative tokens, and 2) partial video relevance, where abundant query-irrelevant segments introduce contextual noise in cross-modal alignment. Existing methods often focus on enhancing multi-scale clip representations and retrieving the most relevant clip. However, the inherent data uncertainty in PRVR renders them vulnerable to distractor videos with spurious similarities, leading to suboptimal performance. To fill this research gap, we propose Robust Alignment Learning (RAL) framework, which explicitly models the uncertainty in data. Key innovations include: 1) we pioneer probabilistic modeling for PRVR by encoding videos and queries as multivariate Gaussian distributions. This not only quantifies data uncertainty but also enables proxy-level matching to capture the variability in cross-modal correspondences; 2) we consider the heterogeneous informativeness of query words and introduce learnable confidence gates to dynamically weight similarity. As a plug-and-play solution, RAL can be seamlessly integrated into the existing architectures. Extensive experiments across diverse retrieval backbones demonstrate its effectiveness.

[240] Self-supervised large-scale kidney abnormality detection in drug safety assessment studies

Ivan Slootweg, Natalia P. García-De-La-Puente, Geert Litjens, Salma Dammak

Main category: cs.CV

TL;DR: First large-scale self-supervised model for kidney abnormality detection in drug safety studies, showing UNI foundation model features alone are insufficient but self-supervised methods can achieve better-than-chance performance.

DetailsMotivation: Kidney abnormality detection in preclinical drug development requires examining thousands of whole-slide images, which is time-consuming and costly. Most slides are normal, so automated detection could significantly reduce workload and costs.

Method: Used UNI foundation model features and tested a simple k-nearest neighbor classifier. Then applied self-supervised learning methods on the same features for abnormality detection across 158 drug safety assessment studies.

Result: k-NN classifier performed at chance level, showing UNI features alone are insufficient. Self-supervised method achieved AUC of 0.62 and negative predictive value of 89%, demonstrating better-than-chance performance.

Conclusion: Self-supervised methods show promise for kidney abnormality detection and could be used to rule out normal slides, reducing time and costs in drug safety assessment studies with further development.

Abstract: Kidney abnormality detection is required for all preclinical drug development. It involves a time-consuming and costly examination of hundreds to thousands of whole-slide images per drug safety study, most of which are normal, to detect any subtle changes indicating toxic effects. In this study, we present the first large-scale self-supervised abnormality detection model for kidney toxicologic pathology, spanning drug safety assessment studies from 158 compounds. We explore the complexity of kidney abnormality detection on this scale using features extracted from the UNI foundation model (FM) and show that a simple k-nearest neighbor classifier on these features performs at chance, demonstrating that the FM-generated features alone are insufficient for detecting abnormalities. We then demonstrate that a self-supervised method applied to the same features can achieve better-than-chance performance, with an area under the receiver operating characteristic curve of 0.62 and a negative predictive value of 89%. With further development, such a model can be used to rule out normal slides in drug safety assessment studies, reducing the costs and time associated with drug development.

[241] Waste-Bench: A Comprehensive Benchmark for Evaluating VLLMs in Cluttered Environments

Muhammad Ali, Salman Khan

Main category: cs.CV

TL;DR: The paper introduces a novel waste classification dataset for evaluating Vision Large Language Models (VLLMs) in complex environments with deformed objects, and provides comprehensive analysis showing VLLMs need improved robustness for challenging real-world scenarios.

DetailsMotivation: While LLMs perform well on standard natural images, their capabilities haven't been thoroughly explored in cluttered datasets with complex environments and deformed objects, particularly for waste classification tasks.

Method: The authors created a novel dataset specifically designed for waste classification in real-world scenarios with complex environments and deformed objects, and developed an in-depth evaluation approach to assess VLLM robustness and accuracy.

Result: The comprehensive analysis revealed that current VLLMs lack sufficient robustness to perform effectively in complex environments with deformed objects, highlighting performance gaps in challenging conditions.

Conclusion: There is a critical need for further advancements in VLLM robustness to enable better performance in complex real-world environments, and the introduced dataset provides valuable insights for future research in this direction.

Abstract: Recent advancements in Large Language Models (LLMs) have paved the way for Vision Large Language Models (VLLMs) capable of performing a wide range of visual understanding tasks. While LLMs have demonstrated impressive performance on standard natural images, their capabilities have not been thoroughly explored in cluttered datasets where there is complex environment having deformed shaped objects. In this work, we introduce a novel dataset specifically designed for waste classification in real-world scenarios, characterized by complex environments and deformed shaped objects. Along with this dataset, we present an in-depth evaluation approach to rigorously assess the robustness and accuracy of VLLMs. The introduced dataset and comprehensive analysis provide valuable insights into the performance of VLLMs under challenging conditions. Our findings highlight the critical need for further advancements in VLLM’s robustness to perform better in complex environments. The dataset and code for our experiments will be made publicly available.

[242] SoccerHigh: A Benchmark Dataset for Automatic Soccer Video Summarization

Artur Díaz-Juan, Coloma Ballester, Gloria Haro

Main category: cs.CV

TL;DR: Introduces a curated soccer video summarization dataset and baseline model to address the lack of publicly available sports highlight datasets, achieving 0.3956 F1 score with a new length-constrained evaluation metric.

DetailsMotivation: Address the gap in publicly available datasets for sports video summarization, particularly for soccer highlight generation, to support video editors in the sports media industry by reducing manual effort.

Method: Curated a dataset with shot boundaries for 237 soccer matches from Spanish, French, and Italian leagues using broadcast footage from SoccerNet. Developed a baseline model specifically designed for soccer video summarization.

Result: The baseline model achieved an F1 score of 0.3956 on the test set. Proposed a new length-constrained evaluation metric for more objective assessment of generated summaries.

Conclusion: Provides a valuable benchmark dataset and baseline approach for soccer video summarization research, with publicly available dataset and code to facilitate further development in sports highlight generation.

Abstract: Video summarization aims to extract key shots from longer videos to produce concise and informative summaries. One of its most common applications is in sports, where highlight reels capture the most important moments of a game, along with notable reactions and specific contextual events. Automatic summary generation can support video editors in the sports media industry by reducing the time and effort required to identify key segments. However, the lack of publicly available datasets poses a challenge in developing robust models for sports highlight generation. In this paper, we address this gap by introducing a curated dataset for soccer video summarization, designed to serve as a benchmark for the task. The dataset includes shot boundaries for 237 matches from the Spanish, French, and Italian leagues, using broadcast footage sourced from the SoccerNet dataset. Alongside the dataset, we propose a baseline model specifically designed for this task, which achieves an F1 score of 0.3956 in the test set. Furthermore, we propose a new metric constrained by the length of each target summary, enabling a more objective evaluation of the generated content. The dataset and code are available at https://ipcv.github.io/SoccerHigh/.

[243] Category-level Text-to-Image Retrieval Improved: Bridging the Domain Gap with Diffusion Models and Vision Encoders

Faizan Farooq Khan, Vladan Stojnić, Zakaria Laskar, Mohamed Elhoseiny, Giorgos Tolias

Main category: cs.CV

TL;DR: A two-step text-to-image retrieval method that converts text queries to visual queries using diffusion models, then uses vision models for image similarity, outperforming text-only approaches.

DetailsMotivation: Vision-and-language models like CLIP map text and images to distant regions in representation space, limiting retrieval performance for semantic category queries.

Method: Two-step approach: 1) Transform text query into visual query using generative diffusion model, 2) Estimate image-to-image similarity with vision model. Includes aggregation network to combine multiple generated images and fuse similarity scores across modalities.

Result: Extensive evaluations show the approach consistently outperforms retrieval methods relying solely on text queries.

Conclusion: The proposed method effectively bridges the modality gap between text and images by leveraging generative diffusion models and vision encoders, improving text-to-image retrieval performance.

Abstract: This work explores text-to-image retrieval for queries that specify or describe a semantic category. While vision-and-language models (VLMs) like CLIP offer a straightforward open-vocabulary solution, they map text and images to distant regions in the representation space, limiting retrieval performance. To bridge this modality gap, we propose a two-step approach. First, we transform the text query into a visual query using a generative diffusion model. Then, we estimate image-to-image similarity with a vision model. Additionally, we introduce an aggregation network that combines multiple generated images into a single vector representation and fuses similarity scores across both query modalities. Our approach leverages advancements in vision encoders, VLMs, and text-to-image generation models. Extensive evaluations show that it consistently outperforms retrieval methods relying solely on text queries. Source code is available at: https://github.com/faixan-khan/cletir

[244] Safe-LLaVA: A Privacy-Preserving Vision-Language Dataset and Benchmark for Biometric Safety

Younggun Kim, Sirnam Swetha, Fazil Kagdi, Mubarak Shah

Main category: cs.CV

TL;DR: PRISM benchmark and Safe-LLaVA dataset address biometric privacy leakage in MLLMs by providing evaluation standards and privacy-preserving training data.

DetailsMotivation: MLLMs often infer and reveal sensitive biometric attributes without explicit requests, raising privacy concerns in real-world applications, but no existing benchmarks or datasets address this issue.

Method: Introduce PRISM benchmark to evaluate MLLMs on refusing biometric queries and detecting implicit leakage, and create Safe-LLaVA dataset by systematically removing biometric information from LLaVA data.

Result: Evaluation reveals widespread biometric leakage across MLLMs, and models fine-tuned on Safe-LLaVA show significant reduction in privacy violations.

Conclusion: PRISM and Safe-LLaVA establish new standards for privacy-aligned development and evaluation of multimodal language models, addressing critical biometric privacy concerns.

Abstract: Multimodal Large Language Models (MLLMs) have demonstrated remarkable capabilities in vision-language tasks. However, these models often infer and reveal sensitive biometric attributes - such as race, gender, age, body weight, and eye color - even when such information is not explicitly requested. This raises critical concerns, particularly in real-world applications and socially-sensitive domains. Despite increasing awareness, no publicly available dataset or benchmark exists to comprehensively evaluate or mitigate biometric leakage in MLLMs. To address this gap, we introduce PRISM (Privacy-aware Evaluation of Responses in Sensitive Modalities), a new benchmark designed to assess MLLMs on two fronts: (1) refuse biometric-related queries and (2) implicit biometric leakage in general responses while maintaining semantic faithfulness. Further, we conduct a detailed audit of the widely used LLaVA datasets and uncover extensive biometric leakage across pretraining and instruction data. To address this, we present Safe-LLaVA dataset, the first privacy-preserving MLLM training dataset constructed by systematically removing explicit and implicit biometric information from LLaVA dataset. Our evaluations on PRISM reveal biometric leakages across MLLMs for different attributes, highlighting the detailed privacy-violations. We also fine-tune a model on Safe-LLaVA dataset and show that it substantially reduces the biometric leakages. Together, Safe-LLaVA & PRISM set a new standard for privacy-aligned development and evaluation of MLLMs. The Safe-LLaVA dataset & PRISM benchmark are publicly available at https://huggingface.co/datasets/kyh9191/Safe-LLaVA, and the source code is available at https://github.com/Kimyounggun99/Safe-LLaVA.git.

[245] Beyond Pixels: Introducing Geometric-Semantic World Priors for Video-based Embodied Models via Spatio-temporal Alignment

Jinzhou Tang, Jusheng zhang, Sidi Liu, Waikit Xiu, Qinhan Lv, Xiying Li

Main category: cs.CV

TL;DR: VEME is a cross-modal alignment method that enhances vision-language models for embodied tasks by learning ego-centric world models with spatio-temporal reasoning, improving navigation and question answering in unseen environments.

DetailsMotivation: Current vision-language models lack spatio-temporal reasoning and adaptation capabilities for dynamic, open-set embodied tasks like navigation and question answering, due to inadequate modeling of fine-grained spatio-temporal cues and physical world comprehension.

Method: Proposes VEME framework with: 1) cross-modal alignment bridging objects, spatial representations and visual semantics with spatio-temporal cues; 2) dynamic implicit cognitive map activated by world embedding for task-relevant memory recall; 3) instruction-based navigation leveraging embodied priors for long-term planning.

Result: Achieves 1%-3% accuracy and exploration efficiency improvement on VSI-Bench and VLN-CE benchmarks compared to traditional approaches.

Conclusion: The method significantly improves reasoning and planning in dynamic environments by embedding geometry-aware spatio-temporal episodic experiences, enhancing generalization in unseen scenes.

Abstract: Achieving human-like reasoning in deep learning models for complex tasks in unknown environments remains a critical challenge in embodied intelligence. While advanced vision-language models (VLMs) excel in static scene understanding, their limitations in spatio-temporal reasoning and adaptation to dynamic, open-set tasks like task-oriented navigation and embodied question answering (EQA) persist due to inadequate modeling of fine-grained spatio-temporal cues and physical world comprehension. To address this, we propose VEME, a novel cross-modal alignment method that enhances generalization in unseen scenes by learning an ego-centric, experience-centered world model. Our framework integrates three key components: (1) a cross-modal alignment framework bridging objects, spatial representations, and visual semantics with spatio-temporal cues to enhance VLM in-context learning; (2) a dynamic, implicit cognitive map activated by world embedding to enable task-relevant geometric-semantic memory recall; and (3) an instruction-based navigation and reasoning framework leveraging embodied priors for long-term planning and efficient exploration. By embedding geometry-aware spatio-temporal episodic experiences, our method significantly improves reasoning and planning in dynamic environments. Experimental results on VSI-Bench and VLN-CE demonstrate 1%-3% accuracy and exploration efficiency improvement compared to traditional approaches.

[246] Multimodal Deep Learning for Phyllodes Tumor Classification from Ultrasound and Clinical Data

Farhan Fuad Abir, Abigail Elliott Daly, Kyle Anderman, Tolga Ozmen, Laura J. Brattain

Main category: cs.CV

TL;DR: Multimodal deep learning framework combining ultrasound images and clinical data improves classification of rare phyllodes tumors, outperforming single-modality approaches and potentially reducing unnecessary surgeries.

DetailsMotivation: Phyllodes tumors are difficult to distinguish from benign fibroadenomas preoperatively using standard methods, leading to unnecessary surgical excisions that could be avoided with better diagnostic tools.

Method: Dual-branch neural network that extracts and fuses features from breast ultrasound images and patient metadata from 81 subjects, using class-aware sampling and subject-stratified 5-fold cross-validation to handle class imbalance and prevent data leakage.

Result: Multimodal method outperformed unimodal baselines, with ConvNeXt and ResNet18 achieving best performance (AUC-ROC: 0.9427/0.9349; F1-scores: 0.6720/0.7294) in classifying benign vs borderline/malignant tumors.

Conclusion: Multimodal AI shows potential as a non-invasive diagnostic tool that could reduce unnecessary biopsies and improve clinical decision-making in breast tumor management.

Abstract: Phyllodes tumors (PTs) are rare fibroepithelial breast lesions that are difficult to classify preoperatively due to their radiological similarity to benign fibroadenomas. This often leads to unnecessary surgical excisions. To address this, we propose a multimodal deep learning framework that integrates breast ultrasound (BUS) images with structured clinical data to improve diagnostic accuracy. We developed a dual-branch neural network that extracts and fuses features from ultrasound images and patient metadata from 81 subjects with confirmed PTs. Class-aware sampling and subject-stratified 5-fold cross-validation were applied to prevent class imbalance and data leakage. The results show that our proposed multimodal method outperforms unimodal baselines in classifying benign versus borderline/malignant PTs. Among six image encoders, ConvNeXt and ResNet18 achieved the best performance in the multimodal setting, with AUC-ROC scores of 0.9427 and 0.9349, and F1-scores of 0.6720 and 0.7294, respectively. This study demonstrates the potential of multimodal AI to serve as a non-invasive diagnostic tool, reducing unnecessary biopsies and improving clinical decision-making in breast tumor management.

[247] GraViT: Transfer Learning with Vision Transformers and MLP-Mixer for Strong Gravitational Lens Discovery

René Parlange, Juan C. Cuevas-Tello, Octavio Valenzuela, Omar de J. Cabrera-Rosas, Tomás Verdugo, Anupreeta More, Anton T. Jaelani

Main category: cs.CV

TL;DR: GraViT is a PyTorch pipeline using Vision Transformers and MLP-Mixer for automated gravitational lens detection, achieving state-of-the-art performance through extensive pretraining and transfer learning strategies.

DetailsMotivation: The LSST survey is predicted to find ~100,000 gravitational lenses in the next decade, requiring automated classifiers to handle the massive data volume and enable cosmological parameter inference.

Method: Fine-tuned ten Vision Transformer and MLP-Mixer architectures using datasets from HOLISMOKES VI and SuGOHI X, with comprehensive analysis of transfer learning impact including data quality, model architecture, training strategies, and ensemble predictions.

Result: Achieved state-of-the-art classification performance for gravitational lens detection, benchmarking against convolutional baselines with complexity and inference-time analysis.

Conclusion: GraViT demonstrates the effectiveness of pretrained Vision Transformers and MLP-Mixer models for automated gravitational lens detection, providing a scalable solution for the upcoming LSST survey data processing.

Abstract: Gravitational lensing offers a powerful probe into the properties of dark matter and is crucial to infer cosmological parameters. The Legacy Survey of Space and Time (LSST) is predicted to find O(10^5) gravitational lenses over the next decade, demanding automated classifiers. In this work, we introduce GraViT, a PyTorch pipeline for gravitational lens detection that leverages extensive pretraining of state-of-the-art Vision Transformer (ViT) models and MLP-Mixer. We assess the impact of transfer learning on classification performance by examining data quality (source and sample size), model architecture (selection and fine-tuning), training strategies (augmentation, normalization, and optimization), and ensemble predictions. This study reproduces the experiments in a previous systematic comparison of neural networks and provides insights into the detectability of strong gravitational lenses on that common test sample. We fine-tune ten architectures using datasets from HOLISMOKES VI and SuGOHI X, and benchmark them against convolutional baselines, discussing complexity and inference-time analysis.

[248] A High-Accuracy Fast Hough Transform with Linear-Log-Cubed Computational Complexity for Arbitrary-Shaped Images

Danil Kazimirov, Dmitry Nikolaev

Main category: cs.CV

TL;DR: FHT2SP algorithm combines fast computation with high accuracy for Hough Transform, achieving near-optimal complexity while maintaining constant-bounded error independent of image size.

DetailsMotivation: Existing fast Hough Transform algorithms either have optimal complexity but reduced accuracy that worsens with scale, or have high accuracy but near-cubic computational cost. There's a need for an algorithm that achieves both speed and accuracy.

Method: The FHT2SP algorithm extends Brady’s superpixel concept to arbitrary shapes beyond power-of-two constraints and integrates it into the FHT2DT algorithm. It uses appropriately sized superpixels to balance computational efficiency and approximation accuracy.

Result: For w×h images, FHT2SP achieves near-optimal computational complexity O(wh ln³ w) while maintaining constant-bounded approximation error independent of image size. The error is controllable via a meta-parameter.

Conclusion: FHT2SP successfully bridges the gap between fast but inaccurate and accurate but slow Hough Transform algorithms, providing both computational efficiency and high accuracy with theoretical guarantees.

Abstract: The Hough transform (HT) is a fundamental tool across various domains, from classical image analysis to neural networks and tomography. Two key aspects of the algorithms for computing the HT are their computational complexity and accuracy - the latter often defined as the error of approximation of continuous lines by discrete ones within the image region. The fast HT (FHT) algorithms with optimal linearithmic complexity - such as the Brady-Yong algorithm for power-of-two-sized images - are well established. Generalizations like $FHT2DT$ extend this efficiency to arbitrary image sizes, but with reduced accuracy that worsens with scale. Conversely, accurate HT algorithms achieve constant-bounded error but require near-cubic computational cost. This paper introduces $FHT2SP$ algorithm - a fast and highly accurate HT algorithm. It builds on our development of Brady’s superpixel concept, extending it to arbitrary shapes beyond the original power-of-two square constraint, and integrates it into the $FHT2DT$ algorithm. With an appropriate choice of the superpixel’s size, for an image of shape $w \times h$, the $FHT2SP$ algorithm achieves near-optimal computational complexity $\mathcal{O}(wh \ln^3 w)$, while keeping the approximation error bounded by a constant independent of image size, and controllable via a meta-parameter. We provide theoretical and experimental analyses of the algorithm’s complexity and accuracy.

[249] Image Quality Enhancement and Detection of Small and Dense Objects in Industrial Recycling Processes

Oussama Messai, Abbass Zein-Eddine, Abdelouahid Bentamou, Mickaël Picq, Nicolas Duquesne, Stéphane Puydarrieux, Yann Gavet

Main category: cs.CV

TL;DR: This paper addresses small, dense, overlapping object detection and image quality enhancement in noisy industrial environments using supervised deep learning methods, evaluated on a new 10k+ image dataset.

DetailsMotivation: To overcome challenges in detecting small, dense, overlapping objects in computer vision and improve image quality in noisy industrial settings, which are critical for industrial applications.

Method: Analysis of supervised deep learning methods using a newly developed dataset with over 10k images and 120k instances. Evaluation of performance, accuracy, and computational efficiency. Introduction of a lightweight fully connected convolutional network model for image quality improvement.

Result: Identification of the most reliable detection systems and specific challenges they address in industrial applications. Development of a dataset repository and proposed model available on GitHub.

Conclusion: The study provides valuable insights into effective object detection and image enhancement methods for industrial environments, with suggested future directions for model improvement and effectiveness enhancement.

Abstract: This paper tackles two key challenges: detecting small, dense, and overlapping objects (a major hurdle in computer vision) and improving the quality of noisy images, especially those encountered in industrial environments. [1, 2]. Our focus is on evaluating methods built on supervised deep learning. We perform an analysis of these methods, using a newly de- veloped dataset comprising over 10k images and 120k in- stances. By evaluating their performance, accuracy, and com- putational efficiency, we identify the most reliable detection systems and highlight the specific challenges they address in industrial applications. This paper also examines the use of deep learning models to improve image quality in noisy industrial environments. We introduce a lightweight model based on a fully connected convolutional network. Addition- ally, we suggest potential future directions for further enhanc- ing the effectiveness of the model. The repository of the dataset and proposed model can be found at: https://github.com/o-messai/SDOOD, https://github.com/o-messai/DDSRNet

[250] Generative AI for Industrial Contour Detection: A Language-Guided Vision System

Liang Gong, Tommy, Wang, Sara Chaker, Yanchen Dong, Fouad Bousetouane, Brenden Morton, Mark Mendez

Main category: cs.CV

TL;DR: Language-guided generative vision system for CAD-level precision remnant contour detection in manufacturing, using conditional GAN and VLM refinement to overcome noise and variability issues in industrial vision systems.

DetailsMotivation: Industrial computer vision systems struggle with noise, material variability, and uncontrolled imaging conditions, limiting classical edge detectors and handcrafted pipelines.

Method: Three-stage system: 1) Data acquisition/preprocessing, 2) Contour generation using conditional GAN, 3) Multimodal contour refinement through vision-language modeling with standardized prompts in human-in-the-loop process.

Result: Improved contour fidelity on FabTrack datasets, enhancing edge continuity and geometric alignment while reducing manual tracing. GPT-image-1 outperformed Gemini 2.0 Flash in structural accuracy and perceptual quality.

Conclusion: VLM-guided generative workflows show promise for advancing industrial computer vision beyond classical pipeline limitations, achieving CAD-level precision in manufacturing applications.

Abstract: Industrial computer vision systems often struggle with noise, material variability, and uncontrolled imaging conditions, limiting the effectiveness of classical edge detectors and handcrafted pipelines. In this work, we present a language-guided generative vision system for remnant contour detection in manufacturing, designed to achieve CAD-level precision. The system is organized into three stages: data acquisition and preprocessing, contour generation using a conditional GAN, and multimodal contour refinement through vision-language modeling, where standardized prompts are crafted in a human-in-the-loop process and applied through image-text guided synthesis. On proprietary FabTrack datasets, the proposed system improved contour fidelity, enhancing edge continuity and geometric alignment while reducing manual tracing. For the refinement stage, we benchmarked several vision-language models, including Google’s Gemini 2.0 Flash, OpenAI’s GPT-image-1 integrated within a VLM-guided workflow, and open-source baselines. Under standardized conditions, GPT-image-1 consistently outperformed Gemini 2.0 Flash in both structural accuracy and perceptual quality. These findings demonstrate the promise of VLM-guided generative workflows for advancing industrial computer vision beyond the limitations of classical pipelines.

[251] Language-Aware Information Maximization for Transductive Few-Shot CLIP

Ghassen Baklouti, Maxime Zanella, Ismail Ben Ayed

Main category: cs.CV

TL;DR: LIMO is a novel transductive few-shot learning method for vision-language models that combines information maximization, KL divergence regularization, and cross-entropy loss with parameter-efficient fine-tuning, achieving state-of-the-art performance.

DetailsMotivation: Transductive few-shot learning has been well-studied for vision-only models but remains underdeveloped for vision-language models (VLMs). Recent methods show potential but require VLM-tailored approaches to fully leverage transduction benefits.

Method: Proposes Language-aware Information MaximizatiOn (LIMO) loss with three components: mutual information between vision inputs and text descriptions, KL divergence regularization from zero-shot predictions, and standard cross-entropy loss. Also explores parameter-efficient fine-tuning strategies.

Result: LIMO significantly outperforms recent transductive few-shot CLIP methods by a large margin and achieves substantial gains over best-performing inductive methods. Parameter-efficient fine-tuning provides substantial performance boosts.

Conclusion: The method demonstrates the effectiveness of information-theoretic approaches and parameter-efficient fine-tuning for transductive few-shot learning in VLMs, establishing new state-of-the-art performance benchmarks.

Abstract: Transductive few-shot learning has triggered an abundant literature focusing on vision-only models, but is still at a nascent stage within the recent context of foundational vision-language models (VLMs). Only a few recent methods addressed the problem, pointing to the potential of tranduction in VLMs and to the need for VLM-tailored methods. Building on this momentum, we leverage information-theoretic concepts and recent progress in parameter-efficient fine-tuning (PEFT), developing a highly competitive transductive few-shot CLIP method. Specifically, we introduce a novel Language-aware Information MaximizatiOn (LIMO) loss integrating three complementary terms: (i) the mutual information between the vision inputs and the textual class descriptions; (ii) a Kullback-Leibler (KL) divergence penalizing deviation of the network’s probabilistic outputs from the text-driven zero-shot predictions; and (iii) a standard cross-entropy loss based on the labeled shots. Furthermore, we challenge the commonly followed fine-tuning practices in the context of transductive few-shot learning, and explore PEFT strategies, completely overlooked in this context. Surprisingly, we observe substantial boosts in performances, which points to the potential of adapting a subset of the model’s parameters in the transductive few-shot setting. We report comprehensive evaluations, which show that LIMO outperforms the very recent transductive few-shot CLIP methods by a large margin and yields significant gains over the best-performing inductive methods. Our code is publicly available at:[ \href{https://github.com/ghassenbaklouti/LIMO}{\text{here}} ]

[252] MorphGen: Morphology-Guided Representation Learning for Robust Single-Domain Generalization in Histopathological Cancer Classification

Hikmat Khan, Syed Farhan Alam Zaidi, Pir Masoom Shah, Kiruthika Balakrishnan, Rabia Khan, Muhammad Waqas, Jia Wu

Main category: cs.CV

TL;DR: MorphGen uses nuclear morphology and spatial organization features with contrastive learning to improve domain generalization in histopathology, making AI more robust to staining variations and imaging differences across institutions.

DetailsMotivation: Pathologists rely on domain-invariant morphological cues that remain diagnostic across different settings, while current ML systems struggle with heterogeneity in whole slide images caused by variations in tissue preparation, staining, and imaging conditions.

Method: MorphGen integrates histopathology images, augmentations, and nuclear segmentation masks in a supervised contrastive learning framework, aligning latent representations of images and nuclear masks. It also incorporates stochastic weight averaging (SWA) for enhanced robustness.

Result: The method demonstrates resilience to image corruptions and adversarial attacks, with attention maps showing primary reliance on nuclear morphology, cellular composition, and spatial cell organization for classification.

Conclusion: Explicitly modeling biologically robust nuclear morphology and spatial organization enables learning cancer representations that are resilient to domain shifts, addressing critical vulnerabilities in current deep learning systems for digital pathology.

Abstract: Domain generalization in computational histopathology is hindered by heterogeneity in whole slide images (WSIs), caused by variations in tissue preparation, staining, and imaging conditions across institutions. Unlike machine learning systems, pathologists rely on domain-invariant morphological cues such as nuclear atypia (enlargement, irregular contours, hyperchromasia, chromatin texture, spatial disorganization), structural atypia (abnormal architecture and gland formation), and overall morphological atypia that remain diagnostic across diverse settings. Motivated by this, we hypothesize that explicitly modeling biologically robust nuclear morphology and spatial organization will enable the learning of cancer representations that are resilient to domain shifts. We propose MorphGen (Morphology-Guided Generalization), a method that integrates histopathology images, augmentations, and nuclear segmentation masks within a supervised contrastive learning framework. By aligning latent representations of images and nuclear masks, MorphGen prioritizes diagnostic features such as nuclear and morphological atypia and spatial organization over staining artifacts and domain-specific features. To further enhance out-of-distribution robustness, we incorporate stochastic weight averaging (SWA), steering optimization toward flatter minima. Attention map analyses revealed that MorphGen primarily relies on nuclear morphology, cellular composition, and spatial cell organization within tumors or normal regions for final classification. Finally, we demonstrate resilience of the learned representations to image corruptions (such as staining artifacts) and adversarial attacks, showcasing not only OOD generalization but also addressing critical vulnerabilities in current deep learning systems for digital pathology. Code, datasets, and trained models are available at: https://github.com/hikmatkhan/MorphGen

[253] SequencePAR: Understanding Pedestrian Attributes via A Sequence Generation Paradigm

Jiandong Jin, Xiao Wang, Yin Lin, Chenglong Li, Lili Huang, Aihua Zheng, Jin Tang

Main category: cs.CV

TL;DR: SequencePAR is a novel sequence generation approach for pedestrian attribute recognition that uses a language-image pre-trained model and Transformer decoder to generate attributes, achieving state-of-the-art performance on multiple datasets.

DetailsMotivation: Current PAR methods using multi-label/multi-task frameworks struggle with imbalanced data and noisy samples. The success of generative models inspired a new sequence generation paradigm to address these limitations.

Method: Extracts pedestrian features using language-image pre-trained model, embeds attribute set into query tokens guided by text prompts, and uses Transformer decoder with masked multi-head attention to generate attributes while preventing next-attribute prediction during training.

Result: Achieved 84.92% accuracy, 90.44% precision, 90.73% recall, and 90.46% F1-score on PETA dataset, with extensive experiments validating effectiveness across multiple PAR datasets.

Conclusion: SequencePAR provides an effective generative approach for pedestrian attribute recognition that outperforms traditional methods and handles data imbalance issues better.

Abstract: Current pedestrian attribute recognition (PAR) algorithms use multi-label or multi-task learning frameworks with specific classification heads. These models often struggle with imbalanced data and noisy samples. Inspired by the success of generative models, we propose Sequence Pedestrian Attribute Recognition (SequencePAR), a novel sequence generation paradigm for PAR. SequencePAR extracts pedestrian features using a language-image pre-trained model and embeds the attribute set into query tokens guided by text prompts. A Transformer decoder generates human attributes by integrating visual features and attribute query tokens. The masked multi-head attention layer in the decoder prevents the model from predicting the next attribute during training. The extensive experiments on multiple PAR datasets validate the effectiveness of SequencePAR. Specifically, we achieve 84.92%, 90.44%, 90.73%, and 90.46% in accuracy, precision, recall, and F1-score on the PETA dataset. The source code and pre-trained models are available at https://github.com/Event-AHU/OpenPAR.

[254] Towards Adaptive Visual Token Pruning for Large Multimodal Models

Hao Zhang, Mengsi Lyu, Chenrui He, Yulong Ao, Yonghua Lin

Main category: cs.CV

TL;DR: A visual token pruning method for Large Multimodal Models that reduces computational costs by selectively removing redundant visual tokens while preserving cross-modal alignment and information diversity.

DetailsMotivation: Large Multimodal Models suffer from high computational and memory costs due to increased token counts during inference. Existing token pruning methods are either costly or use suboptimal importance metrics, leading to redundant retained tokens.

Method: Proposes a two-pronged visual token pruning strategy: 1) removes visual tokens semantically misaligned with textual tokens using mutual information, 2) prunes redundant visual tokens by maximizing expected pairwise distances in embedding space using a greedy algorithm.

Result: Achieves 88.9% token reduction on models like LLaVA-1.5-7B and LLaVA-NEXT-7B while maintaining strong performance, resulting in 56.7% improvement in inference speed.

Conclusion: The proposed visual token pruning method effectively reduces computational overhead while preserving model performance by focusing on cross-modal alignment and intra-modal diversity preservation.

Abstract: Large Multimodal Models (LMMs) have achieved significant success across various tasks. These models usually encode visual inputs into dense token sequences, which are then concatenated with textual tokens and jointly processed by a language model. However, the increased token count substantially raises computational and memory costs during inference. Token pruning has emerged as a promising approach to address this issue. Existing token pruning methods often rely on costly calibration or suboptimal importance metrics, leading to redundant retained tokens. In this paper, we analyze the redundancy differences between visual and textual tokens and propose pruning exclusively on visual tokens. Based on this, we propose a visual token pruning strategy that explicitly preserves both cross-modal alignment and intra-modal informational diversity. We introduce a mutual information-based token pruning strategy that removes visual tokens semantically misaligned with textual tokens, effectively preserving the alignment between the visual and textual modalities. To further improve the representational quality of the retained tokens, we additionally prune redundant visual tokens by maximizing the expected pairwise distances in the embedding space, which is solved efficiently with a greedy algorithm. Extensive experiments demonstrate that our method maintains strong performance while reducing tokens by 88.9% on models such as LLaVA-1.5-7B and LLaVA-NEXT-7B, resulting in a 56.7% improvement in inference speed.

[255] Multimodal LLMs Can Reason about Aesthetics in Zero-Shot

Ruixiang Jiang, Changwen Chen

Main category: cs.CV

TL;DR: This paper addresses the challenge of achieving genuine artistic impact in AI-generated art by using Multimodal LLMs for aesthetic judgment, revealing their tendency for hallucinations and proposing ArtCoT to suppress them through evidence-based reasoning.

DetailsMotivation: Current computational methods lack sophisticated aesthetic sensibility needed for meaningful artistic impact, focusing only on visual appeal while overlooking deeper cognitive processes involved in art appreciation.

Method: Investigates how MLLMs’ reasoning capabilities can be elicited for aesthetic judgment, identifies hallucination issues, and proposes ArtCoT - an evidence-based, objective reasoning process to suppress hallucinations and improve aesthetic reasoning.

Result: MLLMs prompted with ArtCoT produce multifaceted, in-depth aesthetic reasoning that aligns significantly better with human judgment compared to standard approaches.

Conclusion: The work demonstrates that evidence-based reasoning can improve MLLMs’ aesthetic judgment capabilities, with applications in AI art tutoring and reward models for image generation, paving the way for AI systems that truly understand and appreciate art.

Abstract: The rapid technical progress of generative art (GenArt) has democratized the creation of visually appealing imagery. However, achieving genuine artistic impact - the kind that resonates with viewers on a deeper, more meaningful level - remains formidable as it requires a sophisticated aesthetic sensibility. This sensibility involves a multifaceted cognitive process extending beyond mere visual appeal, which is often overlooked by current computational methods. This paper pioneers an approach to capture this complex process by investigating how the reasoning capabilities of Multimodal LLMs (MLLMs) can be effectively elicited to perform aesthetic judgment. Our analysis reveals a critical challenge: MLLMs exhibit a tendency towards hallucinations during aesthetic reasoning, characterized by subjective opinions and unsubstantiated artistic interpretations. We further demonstrate that these hallucinations can be suppressed by employing an evidence-based and objective reasoning process, as substantiated by our proposed baseline, ArtCoT. MLLMs prompted by this principle produce multifaceted, in-depth aesthetic reasoning that aligns significantly better with human judgment. These findings have direct applications in areas such as AI art tutoring and as reward models for image generation. Ultimately, we hope this work paves the way for AI systems that can truly understand, appreciate, and contribute to art that aligns with human aesthetic values. Project homepage: https://github.com/songrise/MLLM4Art.

[256] CryptoFace: End-to-End Encrypted Face Recognition

Wei Ao, Vishnu Naresh Boddeti

Main category: cs.CV

TL;DR: CryptoFace is the first end-to-end encrypted face recognition system using fully homomorphic encryption (FHE) to protect biometric data throughout all processing stages without exposing raw images or features.

DetailsMotivation: Face recognition systems suffer from significant privacy risks due to unauthorized access to sensitive biometric data, requiring secure processing methods.

Method: Uses a mixture of shallow patch convolutional networks with patch-based processing to support higher-dimensional tensors while reducing multiplicative depth and inference latency. Implements parallel FHE evaluation for near-resolution-independent latency.

Result: Significantly accelerates inference and increases verification accuracy compared to state-of-the-art FHE neural networks adapted for face recognition on standard benchmarks.

Conclusion: CryptoFace enables secure face recognition systems with robust and provable security, facilitating privacy-preserving authentication applications.

Abstract: Face recognition is central to many authentication, security, and personalized applications. Yet, it suffers from significant privacy risks, particularly arising from unauthorized access to sensitive biometric data. This paper introduces CryptoFace, the first end-to-end encrypted face recognition system with fully homomorphic encryption (FHE). It enables secure processing of facial data across all stages of a face-recognition process–feature extraction, storage, and matching–without exposing raw images or features. We introduce a mixture of shallow patch convolutional networks to support higher-dimensional tensors via patch-based processing while reducing the multiplicative depth and, thus, inference latency. Parallel FHE evaluation of these networks ensures near-resolution-independent latency. On standard face recognition benchmarks, CryptoFace significantly accelerates inference and increases verification accuracy compared to the state-of-the-art FHE neural networks adapted for face recognition. CryptoFace will facilitate secure face recognition systems requiring robust and provable security. The code is available at https://github.com/human-analysis/CryptoFace.

[257] Generative Frame Sampler for Long Video Understanding

Linli Yao, Haoning Wu, Kun Ouyang, Yuanxing Zhang, Caiming Xiong, Bei Chen, Xu Sun, Junnan Li

Main category: cs.CV

TL;DR: GenS is a plug-and-play generative frame sampler that helps VideoLLMs efficiently process long videos by identifying question-relevant frames, achieving state-of-the-art performance on long-form video benchmarks.

DetailsMotivation: Long-form video understanding poses substantial computational burden due to thousands of frames, making it challenging for VideoLLMs to process efficiently.

Method: Built upon a lightweight VideoLLM, GenS leverages vision-language capabilities to identify question-relevant frames. The method uses a large-scale video instruction dataset (GenS-Video-150K) with dense frame relevance annotations for effective retrieval.

Result: GenS consistently boosts performance of various VideoLLMs. Open-source models achieve SOTA results: LLaVA-Video-72B reaches 66.8 (+4.3) on LongVideoBench and 77.0 (+2.7) on MLVU. Aria obtains 39.2 on HourVideo, surpassing Gemini-1.5-pro by 1.9 points.

Conclusion: GenS effectively addresses computational challenges in long-form video perception and significantly enhances VideoLLM performance across both open-source and proprietary models.

Abstract: Despite recent advances in Video Large Language Models (VideoLLMs), effectively understanding long-form videos remains a significant challenge. Perceiving lengthy videos containing thousands of frames poses substantial computational burden. To mitigate this issue, this paper introduces Generative Frame Sampler (GenS), a plug-and-play module integrated with VideoLLMs to facilitate efficient lengthy video perception. Built upon a lightweight VideoLLM, GenS leverages its inherent vision-language capabilities to identify question-relevant frames. To facilitate effective retrieval, we construct GenS-Video-150K, a large-scale video instruction dataset with dense frame relevance annotations. Extensive experiments demonstrate that GenS consistently boosts the performance of various VideoLLMs, including open-source models (Qwen2-VL-7B, Aria-25B, VILA-40B, LLaVA-Video-7B/72B) and proprietary assistants (GPT-4o, Gemini). When equipped with GenS, open-source VideoLLMs achieve impressive state-of-the-art results on long-form video benchmarks: LLaVA-Video-72B reaches 66.8 (+4.3) on LongVideoBench and 77.0 (+2.7) on MLVU, while Aria obtains 39.2 on HourVideo surpassing the Gemini-1.5-pro by 1.9 points. We will release all datasets and models at https://generative-sampler.github.io.

[258] LUT-Fuse: Towards Extremely Fast Infrared and Visible Image Fusion via Distillation to Learnable Look-Up Tables

Xunpeng Yi, Yibing Zhang, Xinyu Xiang, Qinglong Yan, Han Xu, Jiayi Ma

Main category: cs.CV

TL;DR: LUT-Fuse: A novel infrared and visible image fusion method using learnable lookup tables with distillation strategy, achieving 10x faster speed than current SOTA while maintaining performance.

DetailsMotivation: Current infrared and visible image fusion research focuses on performance improvement but neglects real-time applicability on low-power devices. There's a need for efficient fusion methods that work well on mobile platforms.

Method: Proposes a look-up table structure with low-order approximation encoding and high-level joint contextual scene encoding. Uses efficient LUT distillation strategy instead of traditional quantization methods, transferring knowledge from a multi-modal fusion network (MM-Net) to the MM-LUT model.

Result: Achieves less than one-tenth of the time compared to current lightweight SOTA fusion algorithms. Maintains high operational speed across various scenarios, even on low-power mobile devices. Extensive experiments validate superiority, reliability, and stability.

Conclusion: LUT-Fuse provides an extremely fast and efficient solution for infrared and visible image fusion that is suitable for real-time applications on resource-constrained devices while maintaining competitive performance.

Abstract: Current advanced research on infrared and visible image fusion primarily focuses on improving fusion performance, often neglecting the applicability on real-time fusion devices. In this paper, we propose a novel approach that towards extremely fast fusion via distillation to learnable lookup tables specifically designed for image fusion, termed as LUT-Fuse. Firstly, we develop a look-up table structure that utilizing low-order approximation encoding and high-level joint contextual scene encoding, which is well-suited for multi-modal fusion. Moreover, given the lack of ground truth in multi-modal image fusion, we naturally proposed the efficient LUT distillation strategy instead of traditional quantization LUT methods. By integrating the performance of the multi-modal fusion network (MM-Net) into the MM-LUT model, our method achieves significant breakthroughs in efficiency and performance. It typically requires less than one-tenth of the time compared to the current lightweight SOTA fusion algorithms, ensuring high operational speed across various scenarios, even in low-power mobile devices. Extensive experiments validate the superiority, reliability, and stability of our fusion approach. The code is available at https://github.com/zyb5/LUT-Fuse.

[259] TriPSS: A Tri-Modal Keyframe Extraction Framework Using Perceptual, Structural, and Semantic Representations

Mert Can Cakmak, Nitin Agarwal, Diwash Poudel

Main category: cs.CV

TL;DR: TriPSS is a tri-modal framework for keyframe extraction that combines color, structural, and semantic features using PCA fusion and HDBSCAN clustering, achieving state-of-the-art results on video summarization benchmarks.

DetailsMotivation: Efficient keyframe extraction is crucial for video summarization and retrieval, but capturing the full semantic and visual richness of video content remains challenging, requiring a multi-modal approach.

Method: Integrates perceptual features from CIELAB color space, structural embeddings from ResNet-50, and semantic context from frame-level captions generated by LLaMA-3.2-11B-Vision-Instruct. Uses PCA for modality fusion and HDBSCAN for adaptive video segmentation, with quality assessment and duplicate filtering refinement.

Result: Achieves state-of-the-art performance on TVSum20 and SumMe benchmarks, significantly outperforming both unimodal and prior multimodal approaches.

Conclusion: TriPSS effectively captures complementary visual and semantic cues, establishing it as an effective solution for video summarization, retrieval, and large-scale multimedia understanding.

Abstract: Efficient keyframe extraction is critical for video summarization and retrieval, yet capturing the full semantic and visual richness of video content remains challenging. We introduce TriPSS, a tri-modal framework that integrates perceptual features from the CIELAB color space, structural embeddings from ResNet-50, and semantic context from frame-level captions generated by LLaMA-3.2-11B-Vision-Instruct. These modalities are fused using principal component analysis to form compact multi-modal embeddings, enabling adaptive video segmentation via HDBSCAN clustering. A refinement stage incorporating quality assessment and duplicate filtering ensures the final keyframe set is both concise and semantically diverse. Evaluations on the TVSum20 and SumMe benchmarks show that TriPSS achieves state-of-the-art performance, significantly outperforming both unimodal and prior multimodal approaches. These results highlight TriPSS’ ability to capture complementary visual and semantic cues, establishing it as an effective solution for video summarization, retrieval, and large-scale multimedia understanding.

[260] Target-Oriented Single Domain Generalization

Marzi Heidari, Yuhong Guo

Main category: cs.CV

TL;DR: TO-SDG leverages target domain textual descriptions to guide model generalization without target data, using STAR module to inject target semantics into source features via CLIP and spectral projection.

DetailsMotivation: Existing SDG methods neglect available textual descriptions of target environments, which could significantly enhance generalization under distribution shifts without requiring target data.

Method: STAR module uses target-anchored subspace from text embeddings to recenter image features, spectral projection to retain target-aligned directions, vision-language distillation, and feature-space Mixup for smooth transitions.

Result: Experiments across image classification and object detection benchmarks demonstrate STAR’s superiority over existing methods.

Conclusion: Minimal textual metadata significantly enhances generalization under data constraints, enabling robust model deployment in unseen target environments.

Abstract: Deep models trained on a single source domain often fail catastrophically under distribution shifts, a critical challenge in Single Domain Generalization (SDG). While existing methods focus on augmenting source data or learning invariant features, they neglect a readily available resource: textual descriptions of the target deployment environment. We propose Target-Oriented Single Domain Generalization (TO-SDG), a novel problem setup that leverages the textual description of the target domain, without requiring any target data, to guide model generalization. To address TO-SDG, we introduce Spectral TARget Alignment (STAR), a lightweight module that injects target semantics into source features by exploiting visual-language models (VLMs) such as CLIP. STAR uses a target-anchored subspace derived from the text embedding of the target description to recenter image features toward the deployment domain, then utilizes spectral projection to retain directions aligned with target cues while discarding source-specific noise. Moreover, we use a vision-language distillation to align backbone features with VLM’s semantic geometry. STAR further employs feature-space Mixup to ensure smooth transitions between source and target-oriented representations. Experiments across various image classification and object detection benchmarks demonstrate STAR’s superiority. This work establishes that minimal textual metadata, which is a practical and often overlooked resource, significantly enhances generalization under severe data constraints, opening new avenues for deploying robust models in target environments with unseen data.

[261] MSC: A Marine Wildlife Video Dataset with Grounded Segmentation and Clip-Level Captioning

Quang-Trung Truong, Yuk-Kwan Wong, Vo Hoang Kim Tuyen Dang, Rinaldi Gotama, Duc Thanh Nguyen, Sai-Kit Yeung

Main category: cs.CV

TL;DR: A two-stage marine video captioning pipeline with video-text-segmentation triplets for improved marine video understanding and generation, addressing challenges of marine environments.

DetailsMotivation: Existing video captioning datasets fail to handle marine environment complexities, camera motion, and marine object dynamics, limiting insights about marine life.

Method: Two-stage marine object-oriented video captioning pipeline using video-text-segmentation triplets with video splitting for detecting salient object transitions in scene changes.

Result: Improved marine video understanding, analysis, and generation capabilities with enriched captioning semantics through scene change detection.

Conclusion: The proposed benchmark and pipeline effectively address marine video challenges, providing better tools for marine life analysis and video generation.

Abstract: Marine videos present significant challenges for video understanding due to the dynamics of marine objects and the surrounding environment, camera motion, and the complexity of underwater scenes. Existing video captioning datasets, typically focused on generic or human-centric domains, often fail to generalize to the complexities of the marine environment and gain insights about marine life. To address these limitations, we propose a two-stage marine object-oriented video captioning pipeline. We introduce a comprehensive video understanding benchmark that leverages the triplets of video, text, and segmentation masks to facilitate visual grounding and captioning, leading to improved marine video understanding and analysis, and marine video generation. Additionally, we highlight the effectiveness of video splitting in order to detect salient object transitions in scene changes, which significantly enrich the semantics of captioning content. Our dataset and code have been released at https://msc.hkustvgd.com.

[262] AQFusionNet: Multimodal Deep Learning for Air Quality Index Prediction with Imagery and Sensor Data

Koushik Ahmed Kushal, Abdullah Al Mamun

Main category: cs.CV

TL;DR: AQFusionNet is a multimodal deep learning framework that combines atmospheric imagery with sensor data for accurate AQI prediction in resource-constrained regions, achieving 92.02% accuracy with low computational overhead.

DetailsMotivation: Air pollution monitoring in resource-constrained regions is challenging due to sparse sensor deployment and limited infrastructure, requiring efficient solutions that can work with limited data availability.

Method: Multimodal framework integrating ground-level atmospheric imagery with pollutant concentration data using lightweight CNN backbones (MobileNetV2, ResNet18, EfficientNet-B0) through semantically aligned embedding spaces.

Result: Outperforms unimodal baselines with 92.02% classification accuracy and RMSE of 7.70 using EfficientNet-B0 backbone, delivering 18.5% improvement over single-modality approaches while maintaining low computational overhead.

Conclusion: AQFusionNet provides a scalable and practical solution for AQI monitoring in infrastructure-limited environments, offering robust predictive capability even under partial sensor availability and suitable for edge device deployment.

Abstract: Air pollution monitoring in resource-constrained regions remains challenging due to sparse sensor deployment and limited infrastructure. This work introduces AQFusionNet, a multimodal deep learning framework for robust Air Quality Index (AQI) prediction. The framework integrates ground-level atmospheric imagery with pollutant concentration data using lightweight CNN backbones (MobileNetV2, ResNet18, EfficientNet-B0). Visual and sensor features are combined through semantically aligned embedding spaces, enabling accurate and efficient prediction. Experiments on more than 8,000 samples from India and Nepal demonstrate that AQFusionNet consistently outperforms unimodal baselines, achieving up to 92.02% classification accuracy and an RMSE of 7.70 with the EfficientNet-B0 backbone. The model delivers an 18.5% improvement over single-modality approaches while maintaining low computational overhead, making it suitable for deployment on edge devices. AQFusionNet provides a scalable and practical solution for AQI monitoring in infrastructure-limited environments, offering robust predictive capability even under partial sensor availability.

[263] Iterative Low-rank Network for Hyperspectral Image Denoising

Jin Ye, Fengchao Xiong, Jun Zhou, Yuntao Qian

Main category: cs.CV

TL;DR: ILRNet is a novel iterative low-rank network that combines model-driven and data-driven approaches for hyperspectral image denoising, achieving state-of-the-art performance by integrating rank minimization within U-Net architecture with adaptive learning and iterative refinement.

DetailsMotivation: Hyperspectral images have clean data residing in low-dimensional subspaces with low-rank and sparse properties, but effectively leveraging these physical priors for denoising while preserving image details remains challenging.

Method: ILRNet embeds a rank minimization module (RMM) within U-Net architecture, transforming features to wavelet domain and applying singular value thresholding to low-frequency components. It adaptively learns thresholding parameters and uses iterative refinement to combine intermediate results with noisy inputs.

Result: Experimental results show ILRNet achieves state-of-the-art performance in both synthetic and real-world noise removal tasks.

Conclusion: ILRNet successfully integrates physical priors with deep learning, providing effective hyperspectral image denoising with superior detail preservation through its iterative low-rank network approach.

Abstract: Hyperspectral image (HSI) denoising is a crucial preprocessing step for subsequent tasks. The clean HSI usually reside in a low-dimensional subspace, which can be captured by low-rank and sparse representation, known as the physical prior of HSI. It is generally challenging to adequately use such physical properties for effective denoising while preserving image details. This paper introduces a novel iterative low-rank network (ILRNet) to address these challenges. ILRNet integrates the strengths of model-driven and data-driven approaches by embedding a rank minimization module (RMM) within a U-Net architecture. This module transforms feature maps into the wavelet domain and applies singular value thresholding (SVT) to the low-frequency components during the forward pass, leveraging the spectral low-rankness of HSIs in the feature domain. The parameter, closely related to the hyperparameter of the singular vector thresholding algorithm, is adaptively learned from the data, allowing for flexible and effective capture of low-rankness across different scenarios. Additionally, ILRNet features an iterative refinement process that adaptively combines intermediate denoised HSIs with noisy inputs. This manner ensures progressive enhancement and superior preservation of image details. Experimental results demonstrate that ILRNet achieves state-of-the-art performance in both synthetic and real-world noise removal tasks.

[264] SurgLLM: A Versatile Large Multimodal Model with Spatial Focus and Temporal Awareness for Surgical Video Understanding

Zhen Chen, Xingjian Luo, Kun Yuan, Jinlin Wu, Danny T. M. Chan, Nassir Navab, Hongbin Liu, Zhen Lei, Jiebo Luo

Main category: cs.CV

TL;DR: SurgLLM is a large multimodal model for surgical video understanding that enhances spatial focus through instrument-centric pretraining and improves temporal awareness via multimodal tuning, achieving state-of-the-art results on various surgical video tasks.

DetailsMotivation: Existing surgical video understanding systems suffer from inadequate visual content perception and insufficient temporal awareness, limiting the development of versatile Computer-Assisted Surgery solutions.

Method: Proposed SurgLLM framework with Surgical Context-aware Multimodal Pretraining (instrument-centric Masked Video Reconstruction), Temporal-aware Multimodal Tuning for temporal reasoning, and Surgical Task Dynamic Ensemble for efficient task triaging.

Result: Extensive experiments on diverse surgical video understanding tasks (captioning, general VQA, temporal VQA) demonstrate significant improvements over state-of-the-art approaches.

Conclusion: SurgLLM effectively addresses spatial and temporal challenges in surgical video understanding, providing a versatile solution for Computer-Assisted Surgery systems with enhanced performance across multiple tasks.

Abstract: Surgical video understanding is crucial for facilitating Computer-Assisted Surgery (CAS) systems. Despite significant progress in existing studies, two major limitations persist, including inadequate visual content perception and insufficient temporal awareness in surgical videos, and hinder the development of versatile CAS solutions. In this work, we propose the SurgLLM framework, an effective large multimodal model tailored for versatile surgical video understanding tasks with enhanced spatial focus and temporal awareness. Specifically, to empower the spatial focus of surgical videos, we first devise Surgical Context-aware Multimodal Pretraining (Surg-Pretrain) for the video encoder of SurgLLM, by performing instrument-centric Masked Video Reconstruction (MV-Recon) and subsequent multimodal alignment. To incorporate surgical temporal knowledge into SurgLLM, we further propose Temporal-aware Multimodal Tuning (TM-Tuning) to enhance temporal reasoning with interleaved multimodal embeddings. Moreover, to accommodate various understanding tasks of surgical videos without conflicts, we devise a Surgical Task Dynamic Ensemble to efficiently triage a query with optimal learnable parameters in our SurgLLM. Extensive experiments performed on diverse surgical video understanding tasks, including captioning, general VQA, and temporal VQA, demonstrate significant improvements over the state-of-the-art approaches, validating the effectiveness of our SurgLLM in versatile surgical video understanding. The source code is available at https://github.com/franciszchen/SurgLLM.

[265] A Multimodal Head and Neck Cancer Dataset for AI-Driven Precision Oncology

Numan Saeed, Salma Hassan, Shahad Hardan, Ahmed Aly, Darya Taratynova, Umair Nawaz, Ufaq Khan, Muhammad Ridzuan, Thomas Eugene, Rapha"el Metz, M’elanie Dore, Gregory Delpon, Vijay Ram Kumar Papineni, Kareem Wahid, Cem Dede, Alaa Mohamed Shawky Ali, Carlos Sjogreen, Mohamed Naser, Clifton D. Fuller, Valentin Oreiller, Mario Jreige, John O. Prior, Catherine Cheze Le Rest, Olena Tankyevych, Pierre Decazes, Su Ruan, Stephanie Tanadini-Lang, Martin Valli`eres, Hesham Elhalawani, Ronan Abgral, Romain Floch, Kevin Kerleguer, Ulrike Schick, Maelle Mauguen, Vincent Andrearczyk, Adrien Depeursinge, Mathieu Hatt, Arman Rahmim, Mohammad Yaqub

Main category: cs.CV

TL;DR: A large multimodal PET/CT dataset for head and neck cancer research with 1123 studies from 10 international centers, featuring expert annotations, clinical metadata, and benchmark results for segmentation, survival prediction, and HPV classification tasks.

DetailsMotivation: To provide a comprehensive, publicly available multimodal dataset that reflects real-world clinical diversity to advance head and neck cancer research, particularly for developing and validating AI models in medical imaging analysis.

Method: Collection of 1123 FDG-PET/CT studies from 10 international medical centers with varying acquisition protocols. Expert radiation oncologists and radiologists manually segmented tumor volumes following standardized guidelines. The dataset includes NifTi files, segmentation masks, radiotherapy dose data, and comprehensive clinical metadata.

Result: Created a large-scale annotated dataset with expert segmentations and rich clinical information. Demonstrated utility through benchmark results using state-of-the-art deep learning models (UNet, SegResNet, multimodal frameworks) for three clinical tasks: automated tumor segmentation, recurrence-free survival prediction, and HPV status classification.

Conclusion: This publicly available dataset provides a valuable resource for the research community to develop and validate AI models for head and neck cancer analysis, addressing key clinical challenges with standardized annotations and real-world diversity across multiple institutions.

Abstract: We describe a publicly available multimodal dataset of annotated Positron Emission Tomography/Computed Tomography (PET/CT) studies for head and neck cancer research. The dataset includes 1123 FDG-PET/CT studies from patients with histologically confirmed head and neck cancer, acquired from 10 international medical centers. All examinations consisted of co-registered PET/CT scans with varying acquisition protocols, reflecting real-world clinical diversity across institutions. Primary gross tumor volumes (GTVp) and involved lymph nodes (GTVn) were manually segmented by experienced radiation oncologists and radiologists following standardized guidelines and quality control measures. We provide anonymized NifTi files of all studies, along with expert-annotated segmentation masks, radiotherapy dose distribution for a subset of patients, and comprehensive clinical metadata. This metadata includes TNM staging, HPV status, demographics (age and gender), long-term follow-up outcomes, survival times, censoring indicators, and treatment information. We demonstrate how this dataset can be used for three key clinical tasks: automated tumor segmentation, recurrence-free survival prediction, and HPV status classification, providing benchmark results using state-of-the-art deep learning models, including UNet, SegResNet, and multimodal prognostic frameworks.

[266] Two Causes, Not One: Rethinking Omission and Fabrication Hallucinations in MLLMs

Guangzong Si, Hao Yin, Xianfei Li, Qing Ding, Wenlong Liao, Tao He, Pai Peng

Main category: cs.CV

TL;DR: This paper challenges the common assumption that omission and fabrication hallucinations in MLLMs share the same cause, proposing instead that they have distinct origins and introducing VPFC to reduce omissions without increasing fabrications.

DetailsMotivation: Existing methods for addressing object hallucination in Multimodal Large Language Models are flawed because they assume omission and fabrication hallucinations share a common cause, leading to solutions that reduce omissions but trigger more fabrications.

Method: The authors propose Visual Potential Field Calibration (VPFC), a plug-and-play hallucination mitigation method based on their conceptual framework called Visual-Semantic Attention Potential Field, which reveals how models construct visual evidence for object presence/absence.

Result: VPFC effectively reduces omission hallucinations without introducing additional fabrication hallucinations, demonstrating that the two types of hallucinations have distinct causes that require different mitigation strategies.

Conclusion: The research reveals a critical oversight in current object hallucination research and provides new directions for developing more robust and balanced hallucination mitigation strategies by addressing omission and fabrication hallucinations as distinct phenomena.

Abstract: Multimodal Large Language Models (MLLMs) have achieved impressive advances, yet object hallucination remains a persistent challenge. Existing methods, based on the flawed assumption that omission and fabrication hallucinations share a common cause, often reduce omissions only to trigger more fabrications. In this work, we overturn this view by demonstrating that omission hallucinations arise from insufficient confidence when mapping perceived visual features to linguistic expressions, whereas fabrication hallucinations result from spurious associations within the cross-modal representation space due to statistical biases in the training corpus. Building on findings from visual attention intervention experiments, we propose the Visual-Semantic Attention Potential Field, a conceptual framework that reveals how the model constructs visual evidence to infer the presence or absence of objects. Leveraging this insight, we introduce Visual Potential Field Calibration (VPFC), a plug-and-play hallucination mitigation method that effectively reduces omission hallucinations without introducing additional fabrication hallucinations. Our findings reveal a critical oversight in current object hallucination research and chart new directions for developing more robust and balanced hallucination mitigation strategies.

[267] Activation Steering Meets Preference Optimization: Defense Against Jailbreaks in Vision Language Models

Sihao Wu, Gaojie Jin, Wei Huang, Jianhong Wang, Xiaowei Huang

Main category: cs.CV

TL;DR: SPO-VLM is a two-stage defense framework that combines activation steering with policy optimization to protect Vision Language Models from adversarial attacks while maintaining visual understanding capabilities.

DetailsMotivation: VLMs are vulnerable to adversarial attacks, and existing defense methods using task-specific contrastive prompts have suboptimal performance and degrade visual grounding.

Method: Two-stage approach: Stage I computes adaptive layer-specific steering vectors from diverse data sources to suppress harmful behaviors. Stage II refines these vectors through sequence-level preference optimization with toxicity assessment and visual-consistency rewards.

Result: SPO-VLM enhances safety against attacks while maintaining strong performance on benign tasks without compromising visual understanding capabilities.

Conclusion: The proposed two-stage framework effectively balances efficiency and effectiveness for VLM defense, combining lightweight mitigation with deeper policy refinement for robust and safe text generation.

Abstract: Vision Language Models (VLMs) have demonstrated impressive capabilities in integrating visual and textual information for understanding and reasoning, but remain highly vulnerable to adversarial attacks. While activation steering has emerged as a promising defence, existing approaches often rely on task-specific contrastive prompts to extract harmful directions, which exhibit suboptimal performance and can degrade visual grounding performance. To address these limitations, we propose \textit{Sequence-Level Preference Optimization} for VLM (\textit{SPO-VLM}), a novel two-stage defense framework that combines activation-level intervention with policy-level optimization to enhance model robustness. In \textit{Stage I}, we compute adaptive layer-specific steering vectors from diverse data sources, enabling generalized suppression of harmful behaviors during inference. In \textit{Stage II}, we refine these steering vectors through a sequence-level preference optimization process. This stage integrates automated toxicity assessment, as well as visual-consistency rewards based on caption-image alignment, to achieve safe and semantically grounded text generation. The two-stage structure of SPO-VLM balances efficiency and effectiveness by combining a lightweight mitigation foundation in Stage I with deeper policy refinement in Stage II. Extensive experiments shown SPO-VLM enhances safety against attacks via activation steering and preference optimization, while maintaining strong performance on benign tasks without compromising visual understanding capabilities. We will release our code, model weights, and evaluation toolkit to support reproducibility and future research. \textcolor{red}{Warning: This paper may contain examples of offensive or harmful text and images.}

[268] MultiverSeg: Scalable Interactive Segmentation of Biomedical Imaging Datasets with In-Context Guidance

Hallee E. Wong, Jose Javier Gonzalez Ortiz, John Guttag, Adrian V. Dalca

Main category: cs.CV

TL;DR: MultiverSeg is a system that enables rapid segmentation of new medical image datasets without requiring pre-labeled data, using user interactions that become context for subsequent segmentations, reducing interaction effort by 36% for clicks and 25% for scribbles.

DetailsMotivation: Medical researchers need to segment new image datasets but existing methods require either substantial human effort per image or pre-existing labeled datasets, creating barriers for novel segmentation tasks.

Method: The system takes user interactions (clicks, bounding boxes, scribbles) along with the image to segment, and uses previously segmented images as context. As more images are segmented, the context grows and reduces interaction requirements for new images.

Result: MultiverSeg reduced total clicks by 36% and scribble steps by 25% to achieve 90% Dice score on unseen tasks compared to state-of-the-art interactive segmentation methods.

Conclusion: The system efficiently amortizes user interactions across multiple images, enabling rapid segmentation of new medical datasets without requiring pre-existing labeled data from the target domain.

Abstract: Medical researchers and clinicians often need to perform novel segmentation tasks on a set of related images. Existing methods for segmenting a new dataset are either interactive, requiring substantial human effort for each image, or require an existing set of previously labeled images. We introduce a system, MultiverSeg, that enables practitioners to rapidly segment an entire new dataset without requiring access to any existing labeled data from that task or domain. Along with the image to segment, the model takes user interactions such as clicks, bounding boxes or scribbles as input, and predicts a segmentation. As the user segments more images, those images and segmentations become additional inputs to the model, providing context. As the context set of labeled images grows, the number of interactions required to segment each new image decreases. We demonstrate that MultiverSeg enables users to interactively segment new datasets efficiently, by amortizing the number of interactions per image to achieve an accurate segmentation. Compared to using a state-of-the-art interactive segmentation method, MultiverSeg reduced the total number of clicks by 36% and scribble steps by 25% to achieve 90% Dice on sets of images from unseen tasks. We release code and model weights at https://multiverseg.csail.mit.edu

[269] Adaptive Point-Prompt Tuning: Fine-Tuning Heterogeneous Foundation Models for 3D Point Cloud Analysis

Mengke Li, Lihao Chen, Peng Zhang, Yiu-ming Cheung, Hui Huang

Main category: cs.CV

TL;DR: APPT method enables efficient 3D point cloud analysis by directly leveraging point features to adapt pre-trained models from any modality without heterogeneous mappings, using parameter-efficient fine-tuning with point embeddings and dynamic prompts.

DetailsMotivation: Address the challenge of applying pre-trained foundation models to 3D point cloud analysis due to data scarcity and the limitations of existing "high-to-low" mapping approaches that lose spatial geometries.

Method: Propose Adaptive Point-Prompt Tuning (APPT) which converts point clouds into embeddings with local geometry aggregation, uses permutation-invariant features for relative positions, and employs a weight-sharing prompt generator to dynamically produce point-prompts for self-attention calibration.

Result: Enables direct point cloud processing without heterogeneous mappings, maintains spatial geometries, and reduces computational overhead while adapting any modality foundation model to 3D analysis.

Conclusion: APPT provides a generalizable framework for efficient 3D point cloud analysis by directly leveraging point features and dynamic prompt generation, overcoming limitations of traditional cross-modal adaptation methods.

Abstract: Parameter-efficient fine-tuning strategies for foundation models in 1D textual and 2D visual analysis have demonstrated remarkable efficacy. However, due to the scarcity of point cloud data, pre-training large 3D models remains a challenging task. While many efforts have been made to apply pre-trained visual models to 3D domains through “high-to-low” mapping, these approaches often lead to the loss of spatial geometries and lack a generalizable framework for adapting any modality to 3D. This paper, therefore, attempts to directly leverage point features to calibrate the heterogeneous foundation model of any modality for 3D point cloud analysis. Specifically, we propose the Adaptive Point-Prompt Tuning (APPT) method, which fine-tunes pre-trained models with a modest number of parameters, enabling direct point cloud processing without heterogeneous mappings. We convert raw point clouds into point embeddings by aggregating local geometry to capture spatial features followed by linear layers to ensure seamless utilization of frozen pre-trained models. Given the inherent disorder of point clouds, in contrast to the structured nature of images and language, we employ a permutation-invariant feature to capture the relative positions of point embeddings, thereby obtaining point tokens enriched with location information to optimize self-attention mechanisms. To calibrate self-attention across source domains of any modality to 3D and reduce computational overhead, we introduce a prompt generator that shares weights with the point embedding module, dynamically producing point-prompts without adding additional parameters. These prompts are then concatenated into a frozen foundation model, providing rich global structural information and compensating for the lack of structural context in the heterogeneous data.

[270] Full-Head Segmentation of MRI with Abnormal Brain Anatomy: Model and Data Release

Andrew M Birnbaum, Adam Buchwald, Peter Turkeltaub, Adam Jacks, George Carra, Shreya Kannana, Yu Huang, Abhisheck Datta, Lucas C Parra, Lukas A Hirsch

Main category: cs.CV

TL;DR: Developed MultiAxial network - three 2D U-Nets for sagittal, axial, coronal planes combined for 3D segmentation. Achieved state-of-the-art performance (Dice 0.88) on whole-head MRI segmentation including abnormal anatomy, outperforming existing tools.

DetailsMotivation: To develop a deep network for whole-head segmentation that works with clinical MRIs containing abnormal anatomy, and create the first public benchmark dataset for this purpose to address limitations of existing tools.

Method: Collected 98 MRIs with manual corrections of automated segmentations. Developed MultiAxial network with three independent 2D U-Nets operating in sagittal, axial and coronal planes, then combined to produce 3D segmentation without requiring atlas coregistration.

Result: MultiAxial achieved Dice score of 0.88±0.04, outperforming Multipriors (0.86±0.04) and SPM12 (0.79±0.10). Showed robustness with abnormal anatomy and de-identified images. Enables more accurate current flow modeling in transcranial electric stimulation.

Conclusion: Released state-of-the-art tool for whole-head MRI segmentation in abnormal anatomy along with largest labeled clinical head MRI dataset. Serves as benchmark for future efforts in this domain.

Abstract: Purpose: The goal of this work was to develop a deep network for whole-head segmentation including clinical MRIs with abnormal anatomy, and compile the first public benchmark dataset for this purpose. We collected 98 MRIs with volumetric segmentation labels for a diverse set of human subjects including normal, as well as abnormal anatomy in clinical cases of stroke and disorders of consciousness. Approach: Training labels were generated by manually correcting initial automated segmentations for skin/scalp, skull, CSF, gray matter, white matter, air cavity and extracephalic air. We developed a MultiAxial network consisting of three 2D U-Net that operate independently in sagittal, axial and coronal planes and are then combined to produce a single 3D segmentation. Results: The MultiAxial network achieved a test-set Dice scores of 0.88+-0.04 (median +- interquartile range) on whole head segmentation including gray and white matter. This compared to 0.86 +- 0.04 for Multipriors and 0.79 +- 0.10 for SPM12, two standard tools currently available for this task. The MultiAxial network gains in robustness by avoiding the need for coregistration with an atlas. It performed well in regions with abnormal anatomy and on images that have been de-identified. It enables more accurate and robust current flow modeling when incorporated into ROAST, a widely-used modeling toolbox for transcranial electric stimulation.Conclusions: We are releasing a new state-of-the-art tool for whole-head MRI segmentation in abnormal anatomy, along with the largest volume of labeled clinical head MRIs including labels for non-brain structures. Together the model and data may serve as a benchmark for future efforts.

[271] NoiseCutMix: A Novel Data Augmentation Approach by Mixing Estimated Noise in Diffusion Models

Shumpei Takezaki, Ryoma Bise, Shinnosuke Matsuo

Main category: cs.CV

TL;DR: NoiseCutMix is a novel data augmentation method that combines CutMix with diffusion models to generate natural, high-resolution images with fused characteristics from two classes by partially combining their estimated noise.

DetailsMotivation: Traditional image combination methods like CutMix often produce unnatural boundaries due to contextual differences between images. The authors aim to leverage diffusion models' ability to generate natural, high-resolution images while maintaining the diverse augmentation benefits of CutMix.

Method: The proposed NoiseCutMix method partially combines the estimated noise corresponding to two different classes in a diffusion model, enabling natural fusion of characteristics from both classes without unnatural boundaries.

Result: The method was validated through classification experiments, showing effectiveness compared to conventional data augmentation techniques, random image generation using Stable Diffusion, and combinations of these methods.

Conclusion: NoiseCutMix successfully achieves natural, high-resolution image generation with fused characteristics from two classes, addressing the limitations of traditional CutMix while maintaining data diversity benefits.

Abstract: In this study, we propose a novel data augmentation method that introduces the concept of CutMix into the generation process of diffusion models, thereby exploiting both the ability of diffusion models to generate natural and high-resolution images and the characteristic of CutMix, which combines features from two classes to create diverse augmented data. Representative data augmentation methods for combining images from multiple classes include CutMix and MixUp. However, techniques like CutMix often result in unnatural boundaries between the two images due to contextual differences. Therefore, in this study, we propose a method, called NoiseCutMix, to achieve natural, high-resolution image generation featuring the fused characteristics of two classes by partially combining the estimated noise corresponding to two different classes in a diffusion model. In the classification experiments, we verified the effectiveness of the proposed method by comparing it with conventional data augmentation techniques that combine multiple classes, random image generation using Stable Diffusion, and combinations of these methods. Our codes are available at: https://github.com/shumpei-takezaki/NoiseCutMix

[272] Part Segmentation of Human Meshes via Multi-View Human Parsing

James Dickens, Kamyar Hamad

Main category: cs.CV

TL;DR: A method for semantic segmentation of human meshes using point cloud deep learning, with pseudo-ground truth generation and memory-efficient sampling.

DetailsMotivation: To bridge point cloud deep learning and human parsing by enabling per-vertex semantic segmentation of large-scale human meshes without relying on texture information.

Method: Developed pseudo-ground truth labeling pipeline for Thuman2.1 dataset: mesh alignment, multi-viewpoint segmentation, backprojection. Introduced windowed iterative FPS with space-filling curve serialization for efficient downsampling, followed by geometric segmentation using PointTransformer.

Result: Experimental results confirm the effectiveness and accuracy of the proposed approach for semantic parsing of human meshes.

Conclusion: The method successfully enables semantic segmentation of human meshes using only geometric information, providing an effective solution that bridges computer graphics and human parsing domains.

Abstract: Recent advances in point cloud deep learning have led to models that achieve high per-part labeling accuracy on large-scale point clouds, using only the raw geometry of unordered point sets. In parallel, the field of human parsing focuses on predicting body part and clothing/accessory labels from images. This work aims to bridge these two domains by enabling per-vertex semantic segmentation of large-scale human meshes. To achieve this, a pseudo-ground truth labeling pipeline is developed for the Thuman2.1 dataset: meshes are first aligned to a canonical pose, segmented from multiple viewpoints, and the resulting point-level labels are then backprojected onto the original mesh to produce per-point pseudo ground truth annotations. Subsequently, a novel, memory-efficient sampling strategy is introduced, a windowed iterative farthest point sampling (FPS) with space-filling curve-based serialization to effectively downsample the point clouds. This is followed by a purely geometric segmentation using PointTransformer, enabling semantic parsing of human meshes without relying on texture information. Experimental results confirm the effectiveness and accuracy of the proposed approach. Project code and pre-processed data is available at https://github.com/JamesMcCullochDickens/Human3DParsing/tree/master.

[273] Domain Adaptation-Based Crossmodal Knowledge Distillation for 3D Semantic Segmentation

Jialiang Kang, Jiawen Wang, Dingsheng Luo

Main category: cs.CV

TL;DR: Proposes two crossmodal knowledge distillation methods (UDAKD and FSKD) to transfer knowledge from 2D image models to 3D LiDAR segmentation without requiring 3D annotations, leveraging synchronized camera-LiDAR data in autonomous driving.

DetailsMotivation: Traditional 3D LiDAR segmentation requires extensive annotated point cloud data which is costly and time-consuming to obtain, while 2D image datasets are more abundant and readily available.

Method: Uses crossmodal knowledge distillation with known 2D-3D correspondence to align 3D network outputs with 2D network outputs. Employs self-calibrated convolution on 3D point clouds for domain adaptation, preserving modality-general information while filtering modality-specific details.

Result: The proposed methods consistently surpass state-of-the-art performance in 3D LiDAR semantic segmentation without requiring 3D annotations.

Conclusion: Crossmodal knowledge distillation effectively transfers knowledge from 2D image models to 3D LiDAR segmentation, eliminating the need for costly 3D annotations while achieving superior performance compared to existing methods.

Abstract: Semantic segmentation of 3D LiDAR data plays a pivotal role in autonomous driving. Traditional approaches rely on extensive annotated data for point cloud analysis, incurring high costs and time investments. In contrast, realworld image datasets offer abundant availability and substantial scale. To mitigate the burden of annotating 3D LiDAR point clouds, we propose two crossmodal knowledge distillation methods: Unsupervised Domain Adaptation Knowledge Distillation (UDAKD) and Feature and Semantic-based Knowledge Distillation (FSKD). Leveraging readily available spatio-temporally synchronized data from cameras and LiDARs in autonomous driving scenarios, we directly apply a pretrained 2D image model to unlabeled 2D data. Through crossmodal knowledge distillation with known 2D-3D correspondence, we actively align the output of the 3D network with the corresponding points of the 2D network, thereby obviating the necessity for 3D annotations. Our focus is on preserving modality-general information while filtering out modality-specific details during crossmodal distillation. To achieve this, we deploy self-calibrated convolution on 3D point clouds as the foundation of our domain adaptation module. Rigorous experimentation validates the effectiveness of our proposed methods, consistently surpassing the performance of state-of-the-art approaches in the field.

[274] Visually Grounded Narratives: Reducing Cognitive Burden in Researcher-Participant Interaction

Runtong Wu, Jiayao Song, Fei Teng, Xianhao Ren, Yuyan Gao, Kailun Yang

Main category: cs.CV

TL;DR: NAME is a new paradigm that transforms research documents into coherent story images to reduce cognitive burden in narrative inquiry, achieving state-of-the-art performance with minimal data requirements.

DetailsMotivation: To address the immense burden of data analysis in narrative inquiry where researchers must transform various data into hand-drafted narratives and participants face heavy member checking processes, requiring more efficient and participant-friendly approaches.

Method: Proposed NAME paradigm with actor location and shape module for plausible image generation, and developed robust evaluation metrics across three key dimensions to measure perceptual quality and narrative consistency.

Result: Achieves SOTA performance across different data splits: reduces FID from 195 to 152 using only 0.96% data, improves from 175 to 152 (70:30 split), and nearly halves from 96 to 49 (95:5 split). Scores 3.62 vs baseline 2.66 on new metric.

Conclusion: NAME successfully reduces cognitive burden in narrative analysis by generating coherent story images from research documents, demonstrating significant efficiency improvements with minimal data requirements while maintaining high quality.

Abstract: Narrative inquiry has been one of the prominent application domains for the analysis of human experience, aiming to know more about the complexity of human society. However, researchers are often required to transform various forms of data into coherent hand-drafted narratives in storied form throughout narrative analysis, which brings an immense burden of data analysis. Participants, too, are expected to engage in member checking and presentation of these narrative products, which involves reviewing and responding to large volumes of documents. Given the dual burden and the need for more efficient and participant-friendly approaches to narrative making and representation, we made a first attempt: (i) a new paradigm is proposed, NAME, as the initial attempt to push the field of narrative inquiry. Name is able to transfer research documents into coherent story images, alleviating the cognitive burden of interpreting extensive text-based materials during member checking for both researchers and participants. (ii) We develop an actor location and shape module to facilitate plausible image generation. (iii) We have designed a set of robust evaluation metrics comprising three key dimensions to objectively measure the perceptual quality and narrative consistency of generated characters. Our approach consistently demonstrates state-of-the-art performance across different data partitioning schemes. Remarkably, while the baseline relies on the full 100% of the available data, our method requires only 0.96% yet still reduces the FID score from 195 to 152. Under identical data volumes, our method delivers substantial improvements: for the 70:30 split, the FID score decreases from 175 to 152, and for the 95:5 split, it is nearly halved from 96 to 49. Furthermore, the proposed model achieves a score of 3.62 on the newly introduced metric, surpassing the baseline score of 2.66.

[275] HERO-VQL: Hierarchical, Egocentric and Robust Visual Query Localization

Joohyun Chang, Soyeon Hong, Hyogun Lee, Seong Jong Ha, Dongho Lee, Seong Tae Kim, Jinwoo Choi

Main category: cs.CV

TL;DR: HERO-VQL is a novel method for egocentric visual query localization that handles viewpoint changes and occlusions through top-down attention guidance and egocentric augmentation with consistency training, achieving state-of-the-art performance.

DetailsMotivation: Egocentric videos present challenges like frequent viewpoint changes, object appearance variations, and partial occlusions that make accurate object localization difficult for existing methods.

Method: Proposes HERO-VQL with two key components: 1) Top-down Attention Guidance (TAG) using class token for context and principal component scores for fine-grained localization, and 2) Egocentric Augmentation with Consistency Training (EgoACT) that enhances query diversity and simulates extreme viewpoint changes.

Result: Extensive experiments on VQ2D dataset show HERO-VQL significantly outperforms baseline methods, effectively handling egocentric challenges.

Conclusion: The method successfully addresses egocentric visual query localization challenges by mimicking human cognitive processes and employing robust training strategies with consistency constraints.

Abstract: In this work, we tackle the egocentric visual query localization (VQL), where a model should localize the query object in a long-form egocentric video. Frequent and abrupt viewpoint changes in egocentric videos cause significant object appearance variations and partial occlusions, making it difficult for existing methods to achieve accurate localization. To tackle these challenges, we introduce Hierarchical, Egocentric and RObust Visual Query Localization (HERO-VQL), a novel method inspired by human cognitive process in object recognition. We propose i) Top-down Attention Guidance (TAG) and ii) Egocentric Augmentation based Consistency Training (EgoACT). Top-down Attention Guidance refines the attention mechanism by leveraging the class token for high-level context and principal component score maps for fine-grained localization. To enhance learning in diverse and challenging matching scenarios, EgoAug enhances query diversity by replacing the query with a randomly selected corresponding object from groundtruth annotations and simulates extreme viewpoint changes by reordering video frames. Additionally, CT loss enforces stable object localization across different augmentation scenarios. Extensive experiments on VQ2D dataset validate that HERO-VQL effectively handles egocentric challenges, significantly outperforming baselines.

[276] Double-Constraint Diffusion Model with Nuclear Regularization for Ultra-low-dose PET Reconstruction

Mengxiao Geng, Ran Hong, Bingxuan Li, Qiegen Liu

Main category: cs.CV

TL;DR: DCDM is a parameter-efficient diffusion model for ultra-low-dose PET reconstruction that uses frozen pre-trained weights with trainable double-constraint controllers to handle different dose levels without full retraining.

DetailsMotivation: Ultra-low-dose PET reduces radiation exposure but causes increased noise and reduced image quality. Need for flexible reconstruction methods that can adapt to various dose levels without complete retraining.

Method: Double-Constraint Diffusion Model (DCDM) with frozen pre-trained diffusion model weights and two trainable constraint modules: Nuclear Transformer Constraint (NTC) for low-rank feature compression and Encoding Nexus Constraint (ENC) for controlling the diffusion model using compressed features.

Result: Outperforms state-of-the-art methods on known dose reduction factors and generalizes well to unknown DRF scenarios, even at ultra-low doses (1% of full dose) on UDPET and Clinical datasets.

Conclusion: DCDM provides efficient, flexible ultra-low-dose PET reconstruction with minimal trainable parameters, adapting to various dose levels without full retraining while maintaining high image quality.

Abstract: Ultra-low-dose positron emission tomography (PET) reconstruction holds significant potential for reducing patient radiation exposure and shortening examination times. However, it may also lead to increased noise and reduced imaging detail, which could decrease the image quality. In this study, we present a Double-Constraint Diffusion Model (DCDM), which freezes the weights of a pre-trained diffusion model and injects a trainable double-constraint controller into the encoding architecture, greatly reducing the number of trainable parameters for ultra-low-dose PET reconstruction. Unlike full fine-tuning models, DCDM can adapt to different dose levels without retraining all model parameters, thereby improving reconstruction flexibility. Specifically, the two constraint modules, named the Nuclear Transformer Constraint (NTC) and the Encoding Nexus Constraint (ENC), serve to refine the pre-trained diffusion model. The NTC leverages the nuclear norm as an approximation for matrix rank minimization, integrates the low-rank property into the Transformer architecture, and enables efficient information extraction from low-dose images and conversion into compressed feature representations in the latent space. Subsequently, the ENC utilizes these compressed feature representations to encode and control the pre-trained diffusion model, ultimately obtaining reconstructed PET images in the pixel space. In clinical reconstruction, the compressed feature representations from NTC help select the most suitable ENC for efficient unknown low-dose PET reconstruction. Experiments conducted on the UDPET public dataset and the Clinical dataset demonstrated that DCDM outperforms state-of-the-art methods on known dose reduction factors (DRF) and generalizes well to unknown DRF scenarios, proving valuable even at ultra-low dose levels, such as 1% of the full dose.

[277] DAOVI: Distortion-Aware Omnidirectional Video Inpainting

Ryosuke Seshimo, Mariko Isogawa

Main category: cs.CV

TL;DR: Proposes DAOVI, a deep learning model for omnidirectional video inpainting that handles distortion in equirectangular projection using geodesic distance and depth-aware feature propagation.

DetailsMotivation: Omnidirectional videos capture full surroundings but often contain unwanted objects. Existing video inpainting methods don't handle the distortion in equirectangular projection of 360-degree videos.

Method: Introduces Distortion-Aware Omnidirectional Video Inpainting (DAOVI) with two modules: temporal motion evaluation using geodesic distance in image space, and depth-aware feature propagation in feature space to address geometric distortion.

Result: Experimental results show DAOVI outperforms existing methods both quantitatively and qualitatively.

Conclusion: The proposed method successfully addresses the unique challenges of omnidirectional video inpainting by considering distortion and geometric properties specific to 360-degree content.

Abstract: Omnidirectional videos that capture the entire surroundings are employed in a variety of fields such as VR applications and remote sensing. However, their wide field of view often causes unwanted objects to appear in the videos. This problem can be addressed by video inpainting, which enables the natural removal of such objects while preserving both spatial and temporal consistency. Nevertheless, most existing methods assume processing ordinary videos with a narrow field of view and do not tackle the distortion in equirectangular projection of omnidirectional videos. To address this issue, this paper proposes a novel deep learning model for omnidirectional video inpainting, called Distortion-Aware Omnidirectional Video Inpainting (DAOVI). DAOVI introduces a module that evaluates temporal motion information in the image space considering geodesic distance, as well as a depth-aware feature propagation module in the feature space that is designed to address the geometric distortion inherent to omnidirectional videos. The experimental results demonstrate that our proposed method outperforms existing methods both quantitatively and qualitatively.

[278] DevilSight: Augmenting Monocular Human Avatar Reconstruction through a Virtual Perspective

Yushuo Chen, Ruizhi Shao, Youxin Pang, Hongwen Zhang, Xinyi Wu, Rihui Wu, Yebin Liu

Main category: cs.CV

TL;DR: A novel framework for human avatar reconstruction from monocular videos using video generative models to generate additional supervision from alternative viewpoints, improving detail capture and novel view synthesis.

DetailsMotivation: Current approaches struggle with capturing fine-grained dynamic details and generating plausible details at novel viewpoints due to limited model capacity and insufficient observational data from monocular videos.

Method: Leverages Human4DiT video generative model to generate human motions from alternative perspectives as additional supervision. Uses video fine-tuning to inject physical identity for consistent motion reproduction, and employs patch-based denoising for higher-resolution outputs with finer details.

Result: Outperforms recent state-of-the-art approaches in human avatar reconstruction from monocular videos, demonstrating effectiveness in enriching details in unseen regions and mitigating artifacts.

Conclusion: The proposed framework successfully addresses limitations of current monocular video-based avatar reconstruction by using generative models for additional supervision and complementary strategies for motion consistency and detail enhancement.

Abstract: We present a novel framework to reconstruct human avatars from monocular videos. Recent approaches have struggled either to capture the fine-grained dynamic details from the input or to generate plausible details at novel viewpoints, which mainly stem from the limited representational capacity of the avatar model and insufficient observational data. To overcome these challenges, we propose to leverage the advanced video generative model, Human4DiT, to generate the human motions from alternative perspective as an additional supervision signal. This approach not only enriches the details in previously unseen regions but also effectively regularizes the avatar representation to mitigate artifacts. Furthermore, we introduce two complementary strategies to enhance video generation: To ensure consistent reproduction of human motion, we inject the physical identity into the model through video fine-tuning. For higher-resolution outputs with finer details, a patch-based denoising algorithm is employed. Experimental results demonstrate that our method outperforms recent state-of-the-art approaches and validate the effectiveness of our proposed strategies.

[279] LightVLM: Acceleraing Large Multimodal Models with Pyramid Token Merging and KV Cache Compression

Lianyu Hu, Fanhua Shang, Wei Feng, Liang Wan

Main category: cs.CV

TL;DR: LightVLM is a training-free method that accelerates Vision-Language Models by reducing tokens during encoding and compressing KV Cache during decoding, achieving significant speed improvements with minimal performance loss.

DetailsMotivation: Vision-Language Models suffer from high inference latency, especially when processing images and generating long text sequences, which hinders their real-world deployment.

Method: Two-stage approach: 1) Pyramid token merging during encoding to hierarchically reduce image tokens, 2) KV Cache compression during decoding to remove unnecessary caches and increase throughput.

Result: Maintains 100% performance with 35% tokens, 98% performance with 3% tokens. Achieves 2.02× throughput, 3.65× prefilling time reduction, and 3.21× inference time reduction for long sequences.

Conclusion: LightVLM enables efficient deployment of large VLMs by significantly accelerating inference while preserving performance, making heavy models faster than smaller counterparts.

Abstract: In this paper, we introduce LightVLM, a simple but effective method that can be seamlessly deployed upon existing Vision-Language Models (VLMs) to greatly accelerate the inference process in a training-free manner. We divide the inference procedure of VLMs into two stages, i.e., encoding and decoding, and propose to simultaneously accelerate VLMs in both stages to largely improve model efficiency. During encoding, we propose pyramid token merging to reduce tokens of different LLM layers in a hierarchical manner by finally only keeping a few dominant tokens to achieve high efficiency. During decoding, aimed at reducing the high latency of outputting long sequences, we propose KV Cache compression to remove unnecessary caches to increase the network throughput. Experimental results show that LightVLM successfully retains 100% performance when only preserving 35% image tokens, and maintains around 98% performance when keeping only 3% image tokens. LightVLM could 2.02$\times$ the network throughput and reduce the prefilling time by 3.65$\times$. LightVLM also makes large VLMs faster again by enabling a heavy model (e.g., InternVL2.5 26B) to infer faster than significantly smaller models (e.g., InternVL2.5 8B), hopefully facilitating the real-world deployment. When generating long text sequences (e.g., 4096 tokens), LightVLM could reduce the inference time by 3.21$\times$, largely outperforming existing methods.

[280] Mixture of Global and Local Experts with Diffusion Transformer for Controllable Face Generation

Xuechao Zou, Shun Zhang, Xing Fu, Yue Li, Kai Li, Yushe Cao, Congyan Lang, Pin Tao, Junliang Xing

Main category: cs.CV

TL;DR: Face-MoGLE is a novel diffusion transformer framework for controllable face generation that uses semantic-decoupled latent modeling, mixture of global/local experts, and dynamic gating for precise attribute manipulation and high photorealism.

DetailsMotivation: Existing approaches struggle with disentangling semantic controls from generation pipelines while maintaining photorealism in face generation tasks.

Method: Uses mask-conditioned space factorization for semantic-decoupled latent modeling, mixture of global and local experts for structure/semantics capture, and dynamic gating network with time-dependent coefficients.

Result: Demonstrates effectiveness in multimodal and monomodal face generation settings with robust zero-shot generalization capability.

Conclusion: Provides a powerful and flexible solution for high-quality, controllable face generation with strong potential in generative modeling and security applications.

Abstract: Controllable face generation poses critical challenges in generative modeling due to the intricate balance required between semantic controllability and photorealism. While existing approaches struggle with disentangling semantic controls from generation pipelines, we revisit the architectural potential of Diffusion Transformers (DiTs) through the lens of expert specialization. This paper introduces Face-MoGLE, a novel framework featuring: (1) Semantic-decoupled latent modeling through mask-conditioned space factorization, enabling precise attribute manipulation; (2) A mixture of global and local experts that captures holistic structure and region-level semantics for fine-grained controllability; (3) A dynamic gating network producing time-dependent coefficients that evolve with diffusion steps and spatial locations. Face-MoGLE provides a powerful and flexible solution for high-quality, controllable face generation, with strong potential in generative modeling and security applications. Extensive experiments demonstrate its effectiveness in multimodal and monomodal face generation settings and its robust zero-shot generalization capability. Project page is available at https://github.com/XavierJiezou/Face-MoGLE.

[281] SemaMIL: Semantic Reordering with Retrieval-Guided State Space Modeling for Whole Slide Image Classification

Lubin Gan, Xiaoman Wu, Jing Zhang, Zhifeng Wang, Linhao Qu, Siying Wu, Xiaoyan Sun

Main category: cs.CV

TL;DR: SemaMIL introduces Semantic Reordering and Semantic-guided Retrieval State Space Module to improve WSI analysis with linear complexity while maintaining interpretability.

DetailsMotivation: Existing MIL methods for whole slide images either overlook contextual relationships (attention-based) or have quadratic computational cost (Transformers), while state space models lose interpretability when patch order is shuffled.

Method: Integrates Semantic Reordering (clusters and arranges semantically similar patches through reversible permutation) with Semantic-guided Retrieval State Space Module (selects representative query subset to adjust state space parameters for better global modeling).

Result: Achieves state-of-the-art accuracy on four WSI subtype datasets with fewer FLOPs and parameters compared to strong baselines.

Conclusion: SemaMIL effectively addresses computational complexity and interpretability issues in WSI analysis while maintaining high performance.

Abstract: Multiple instance learning (MIL) has become the leading approach for extracting discriminative features from whole slide images (WSIs) in computational pathology. Attention-based MIL methods can identify key patches but tend to overlook contextual relationships. Transformer models are able to model interactions but require quadratic computational cost and are prone to overfitting. State space models (SSMs) offer linear complexity, yet shuffling patch order disrupts histological meaning and reduces interpretability. In this work, we introduce SemaMIL, which integrates Semantic Reordering (SR), an adaptive method that clusters and arranges semantically similar patches in sequence through a reversible permutation, with a Semantic-guided Retrieval State Space Module (SRSM) that chooses a representative subset of queries to adjust state space parameters for improved global modeling. Evaluation on four WSI subtype datasets shows that, compared to strong baselines, SemaMIL achieves state-of-the-art accuracy with fewer FLOPs and parameters.

[282] Stage-wise Adaptive Label Distribution for Facial Age Estimation

Bo Wu, Zhiqi Ai, Jun Jiang, Congcong Zhu, Shugong Xu

Main category: cs.CV

TL;DR: SA-LDL algorithm addresses label ambiguity in age estimation through stage-wise adaptive variance modeling and weighted loss, achieving MAE of 1.74-2.15 on benchmark datasets.

DetailsMotivation: Existing methods overlook varying degrees of label ambiguity across different age stages, which poses significant challenges in age estimation tasks.

Method: Stage-wise Adaptive Label Distribution Learning (SA-LDL) that uses stage-wise adaptive variance modeling and weighted loss function based on analysis of embedding similarities between age groups.

Result: Achieves competitive performance with MAE of 1.74 on MORPH-II and 2.15 on FG-NET datasets, demonstrating effectiveness in capturing structured label ambiguity.

Conclusion: SA-LDL effectively addresses the stage-wise patterns of label ambiguity in age estimation, leading to more accurate and robust performance compared to existing methods.

Abstract: Label ambiguity poses a significant challenge in age estimation tasks. Most existing methods address this issue by modeling correlations between adjacent age groups through label distribution learning. However, they often overlook the varying degrees of ambiguity present across different age stages. In this paper, we propose a Stage-wise Adaptive Label Distribution Learning (SA-LDL) algorithm, which leverages the observation – revealed through our analysis of embedding similarities between an anchor and all other ages – that label ambiguity exhibits clear stage-wise patterns. By jointly employing stage-wise adaptive variance modeling and weighted loss function, SA-LDL effectively captures the complex and structured nature of label ambiguity, leading to more accurate and robust age estimation. Extensive experiments demonstrate that SA-LDL achieves competitive performance, with MAE of 1.74 and 2.15 on the MORPH-II and FG-NET datasets.

[283] Encoder-Only Image Registration

Xiang Chen, Renjiu Hu, Jinwei Zhang, Yuxi Zhang, Xinyao Yue, Min Liu, Yaonan Wang, Hang Zhang

Main category: cs.CV

TL;DR: EOIR framework uses 3-layer ConvNet for feature extraction and separate flow estimators to achieve better accuracy-efficiency trade-off in deformable image registration.

DetailsMotivation: To address challenges in learning-based deformable image registration including computational complexity and handling large deformations, by understanding how ConvNets influence registration performance.

Method: Encoder-Only Image Registration (EOIR) framework that separates feature learning from flow estimation, using a 3-layer ConvNet for feature extraction and 3-layer flow estimators to build a Laplacian feature pyramid for progressive diffeomorphic deformations.

Result: EOIR achieves superior accuracy-efficiency and accuracy-smoothness trade-offs across five datasets of different modalities and anatomical regions, providing better efficiency and smoothness with comparable accuracy.

Conclusion: The proposed EOIR framework effectively addresses registration challenges by leveraging insights about ConvNet roles in linearizing intensities and harmonizing contrast variations, offering a practical solution with public code availability.

Abstract: Learning-based techniques have significantly improved the accuracy and speed of deformable image registration. However, challenges such as reducing computational complexity and handling large deformations persist. To address these challenges, we analyze how convolutional neural networks (ConvNets) influence registration performance using the Horn-Schunck optical flow equation. Supported by prior studies and our empirical experiments, we observe that ConvNets play two key roles in registration: linearizing local intensities and harmonizing global contrast variations. Based on these insights, we propose the Encoder-Only Image Registration (EOIR) framework, designed to achieve a better accuracy-efficiency trade-off. EOIR separates feature learning from flow estimation, employing only a 3-layer ConvNet for feature extraction and a set of 3-layer flow estimators to construct a Laplacian feature pyramid, progressively composing diffeomorphic deformations under a large-deformation model. Results on five datasets across different modalities and anatomical regions demonstrate EOIR’s effectiveness, achieving superior accuracy-efficiency and accuracy-smoothness trade-offs. With comparable accuracy, EOIR provides better efficiency and smoothness, and vice versa. The source code of EOIR will be publicly available on https://github.com/XiangChen1994/EOIR.

[284] Exploring Decision-Making Capabilities of LLM Agents: An Experimental Study on Jump-Jump Game

Juwu Li

Main category: cs.CV

TL;DR: The Jump-Jump game serves as an ideal testbed for evaluating LLM decision-making capabilities through its requirement for precise force control based on spatial reasoning and strategic planning.

DetailsMotivation: To study LLM decision-making capabilities using a simple yet challenging casual game that involves multiple cognitive aspects including spatial reasoning, physical modeling, and strategic planning.

Method: Using the Jump-Jump game as a testing environment where players control a red circle character to jump across platforms with appropriate force to maximize score.

Result: The paper establishes the Jump-Jump game as a suitable platform for evaluating cognitive aspects of LLM decision-making, though specific experimental results are not provided in this abstract.

Conclusion: The Jump-Jump game provides an effective framework for studying and testing LLM decision-making capabilities through its gameplay mechanics that require complex cognitive processing.

Abstract: The Jump-Jump game, as a simple yet challenging casual game, provides an ideal testing environment for studying LLM decision-making capabilities. The game requires players to precisely control jumping force based on current position and target platform distance, involving multiple cognitive aspects including spatial reasoning, physical modeling, and strategic planning. It illustrates the basic gameplay mechanics of the Jump-Jump game, where the player character (red circle) must jump across platforms with appropriate force to maximize score.

[285] VideoRewardBench: Comprehensive Evaluation of Multimodal Reward Models for Video Understanding

Zhihong Zhang, Xiaojian Huang, Jin Xu, Zhuodong Luo, Xinzhi Wang, Jiansheng Wei, Xuejin Chen

Main category: cs.CV

TL;DR: VideoRewardBench is a comprehensive benchmark for evaluating multimodal reward models in video understanding, covering perception, knowledge, reasoning, and safety with 1,563 annotated samples, showing current models achieve only 53-57% accuracy.

DetailsMotivation: Existing benchmarks for multimodal reward models in video domain suffer from limited question diversity, lack of comprehensive evaluation dimensions, and inadequate coverage of different MRM types.

Method: Created VideoRewardBench through AI-assisted data pipeline with 1,563 annotated samples (video-text prompts with chosen/rejected responses), covering 4 video understanding aspects. Evaluated 28 MRMs across generative, discriminative, and semi-scalar categories.

Result: Top-performing GPT-4o achieved only 57.0% accuracy, best open-source model Qwen2.5-VL-72B reached 53.3%. Key findings: RL-trained MRMs don’t necessarily have better cross-modal generalization; most MRMs benefit from inference-time scaling; video frame count affects different MRM types differently.

Conclusion: VideoRewardBench provides a challenging benchmark that reveals significant performance gaps in current MRMs and offers valuable insights for advancing multimodal reward model development in video domain.

Abstract: Multimodal reward models (MRMs) play a crucial role in the training, inference, and evaluation of Large Vision Language Models (LVLMs) by assessing response quality. However, existing benchmarks for evaluating MRMs in the video domain suffer from a limited number and diversity of questions, a lack of comprehensive evaluation dimensions, and inadequate evaluation of diverse types of MRMs. To address these gaps, we introduce VideoRewardBench, the first comprehensive benchmark covering four core aspects of video understanding: perception, knowledge, reasoning, and safety. Through our AI-assisted data pipeline, we curate a high-quality preference dataset of 1,563 annotated samples, including 1,482 unique videos and 1,559 distinct questions–15 times the number found in the most question-rich prior benchmark. Each sample is a triplet consisting of a video-text prompt, a chosen response, and a rejected response. We also conduct a comprehensive evaluation across 28 multimodal reward models spanning three categories: generative, discriminative, and semi-scalar. Results show that even the top-performing model GPT-4o achieves only 57.0% overall accuracy, and the state-of-the-art open-source model Qwen2.5-VL-72B reaches merely 53.3%. Our analysis further reveals three key insights: (i) MRMs trained with reinforcement learning (RL) do not necessarily exhibit stronger cross-modal generalization than those trained without RL; (ii) except for discriminative MRMs, other types of MRMs across varying model capacities can benefit from inference-time scaling; and (iii) variations in input video frame count have different effects on different types of MRMs. We believe VideoRewardBench offers a challenging and valuable benchmark for advancing the evaluation and development of MRMs in the video domain.

[286] Multi-Focused Video Group Activities Hashing

Zhongmiao Qi, Yan Jiang, Bolin Zhang, Lijun Guo, Chong Wang, Qiangbo Qian

Main category: cs.CV

TL;DR: Proposes STVH and M-STVH techniques for fine-grained video activity retrieval, capturing both group interactions and individual object dynamics through spatiotemporal modeling.

DetailsMotivation: Address the need for granular activity retrieval in videos rather than just whole-video retrieval, especially in complex scenarios with explosive video data growth.

Method: STVH: Unified framework modeling individual object dynamics and group interactions with spatiotemporal evolution. M-STVH: Enhanced version with hierarchical feature integration and multi-focused representation learning to handle both activity semantics and object visual features.

Result: Both STVH and M-STVH achieve excellent results on publicly available datasets.

Conclusion: The proposed techniques successfully solve the problem of fine-grained group activity retrieval in videos through spatiotemporal modeling and multi-focused representation learning.

Abstract: With the explosive growth of video data in various complex scenarios, quickly retrieving group activities has become an urgent problem. However, many tasks can only retrieve videos focusing on an entire video, not the activity granularity. To solve this problem, we propose a new STVH (spatiotemporal interleaved video hashing) technique for the first time. Through a unified framework, the STVH simultaneously models individual object dynamics and group interactions, capturing the spatiotemporal evolution on both group visual features and positional features. Moreover, in real-life video retrieval scenarios, it may sometimes require activity features, while at other times, it may require visual features of objects. We then further propose a novel M-STVH (multi-focused spatiotemporal video hashing) as an enhanced version to handle this difficult task. The advanced method incorporates hierarchical feature integration through multi-focused representation learning, allowing the model to jointly focus on activity semantics features and object visual features. We conducted comparative experiments on publicly available datasets, and both STVH and M-STVH can achieve excellent results.

[287] TRUST: Token-dRiven Ultrasound Style Transfer for Cross-Device Adaptation

Nhat-Tuong Do-Tran, Ngoc-Hoang-Lam Le, Ian Chiu, Po-Tsun Paul Kuo, Ching-Chun Huang

Main category: cs.CV

TL;DR: TRUST is a token-driven dual-stream framework for ultrasound image style transfer that preserves source content while transferring target domain style, improving downstream task performance by filtering relevant style features.

DetailsMotivation: Existing unpaired image-to-image translation methods don't adequately filter style features for downstream tasks, causing translated images to be misaligned with task needs when ultrasound devices have different styles.

Method: Proposes TRUST framework with Token-dRiven module that selects suitable target tokens from data and model perspectives, uses behavior mirror loss, and injects auxiliary prompts to match content representation with downstream behavior.

Result: Experimental results on ultrasound datasets show TRUST outperforms existing UI2I methods in both visual quality and downstream task performance.

Conclusion: TRUST effectively bridges the style gap in ultrasound images while preserving content integrity, making it suitable for improving downstream medical imaging tasks across different devices.

Abstract: Ultrasound images acquired from different devices exhibit diverse styles, resulting in decreased performance of downstream tasks. To mitigate the style gap, unpaired image-to-image (UI2I) translation methods aim to transfer images from a source domain, corresponding to new device acquisitions, to a target domain where a frozen task model has been trained for downstream applications. However, existing UI2I methods have not explicitly considered filtering the most relevant style features, which may result in translated images misaligned with the needs of downstream tasks. In this work, we propose TRUST, a token-driven dual-stream framework that preserves source content while transferring the common style of the target domain, ensuring that content and style remain unblended. Given multiple styles in the target domain, we introduce a Token-dRiven (TR) module that operates from two perspectives: (1) a data view–selecting “suitable” target tokens corresponding to each source token, and (2) a model view–identifying ``optimal” target tokens for the downstream model, guided by a behavior mirror loss. Additionally, we inject auxiliary prompts into the source encoder to match content representation with downstream behavior. Experimental results on ultrasound datasets demonstrate that TRUST outperforms existing UI2I methods in both visual quality and downstream task performance.

[288] Make me an Expert: Distilling from Generalist Black-Box Models into Specialized Models for Semantic Segmentation

Yasser Benigmim, Subhankar Roy, Khalid Oublal, Imad Eddine Marouf, Slim Essid, Vicky Kalogeiton, Stéphane Lathuilière

Main category: cs.CV

TL;DR: ATGC method enables effective local model training using black-box AI APIs by dynamically selecting optimal input scales through attention-guided analysis, overcoming resolution sensitivity limitations in open-vocabulary models.

DetailsMotivation: To address the challenge of training local models using black-box AIaaS APIs that don't expose weights, training data, or logits, which makes current domain adaptation methods impractical.

Method: ATtention-Guided sCaler (ATGC) leverages DINOv2 attention maps to dynamically select optimal scales for black-box model inference, scoring attention maps with entropy to identify informative scales for pseudo-labelling.

Result: Experiments show substantial improvements under black-box supervision across multiple datasets while requiring only one-hot API predictions.

Conclusion: ATGC effectively enables local model adaptation using black-box APIs by addressing the ‘curse of resolution’ through attention-guided scale selection, making black-box distillation practical with minimal API access requirements.

Abstract: The rise of Artificial Intelligence as a Service (AIaaS) democratizes access to pre-trained models via Application Programming Interfaces (APIs), but also raises a fundamental question: how can local models be effectively trained using black-box models that do not expose their weights, training data, or logits, a constraint in which current domain adaptation paradigms are impractical ? To address this challenge, we introduce the Black-Box Distillation (B2D) setting, which enables local model adaptation under realistic constraints: (1) the API model is open-vocabulary and trained on large-scale general-purpose data, and (2) access is limited to one-hot predictions only. We identify that open-vocabulary models exhibit significant sensitivity to input resolution, with different object classes being segmented optimally at different scales, a limitation termed the “curse of resolution”. Our method, ATtention-Guided sCaler (ATGC), addresses this challenge by leveraging DINOv2 attention maps to dynamically select optimal scales for black-box model inference. ATGC scores the attention maps with entropy to identify informative scales for pseudo-labelling, enabling effective distillation. Experiments demonstrate substantial improvements under black-box supervision across multiple datasets while requiring only one-hot API predictions. Our code is available at https://github.com/yasserben/ATGC.

[289] Learning Yourself: Class-Incremental Semantic Segmentation with Language-Inspired Bootstrapped Disentanglement

Ruitao Wu, Yifan Zhao, Jia Li

Main category: cs.CV

TL;DR: A novel framework (LBD) that uses language semantics from pre-trained models like CLIP to address catastrophic semantic entanglement in class-incremental semantic segmentation, achieving state-of-the-art performance.

DetailsMotivation: Existing CISS methods suffer from catastrophic semantic entanglement - prototype-feature entanglement due to semantic misalignment and background-increment entanglement from dynamic data evolution, which introduces noise and errors.

Method: Language-inspired Bootstrapped Disentanglement framework using CLIP’s prior class semantics. Includes Language-guided Prototypical Disentanglement (using text features as topological templates) and Manifold Mutual Background Disentanglement (multiple learnable prototypes with mask-pooling supervision).

Result: Achieves state-of-the-art performance on Pascal VOC and ADE20k datasets, particularly in multi-step scenarios.

Conclusion: LBD effectively addresses semantic entanglement in CISS by leveraging language semantics for autonomous feature disentanglement, bridging the capability gap between dense and sparse tasks through soft prompt tuning and encoder adaptation.

Abstract: Class-Incremental Semantic Segmentation (CISS) requires continuous learning of newly introduced classes while retaining knowledge of past classes. By abstracting mainstream methods into two stages (visual feature extraction and prototype-feature matching), we identify a more fundamental challenge termed catastrophic semantic entanglement. This phenomenon involves Prototype-Feature Entanglement caused by semantic misalignment during the incremental process, and Background-Increment Entanglement due to dynamic data evolution. Existing techniques, which rely on visual feature learning without sufficient cues to distinguish targets, introduce significant noise and errors. To address these issues, we introduce a Language-inspired Bootstrapped Disentanglement framework (LBD). We leverage the prior class semantics of pre-trained visual-language models (e.g., CLIP) to guide the model in autonomously disentangling features through Language-guided Prototypical Disentanglement and Manifold Mutual Background Disentanglement. The former guides the disentangling of new prototypes by treating hand-crafted text features as topological templates, while the latter employs multiple learnable prototypes and mask-pooling-based supervision for background-incremental class disentanglement. By incorporating soft prompt tuning and encoder adaptation modifications, we further bridge the capability gap of CLIP between dense and sparse tasks, achieving state-of-the-art performance on both Pascal VOC and ADE20k, particularly in multi-step scenarios.

[290] A Modality-agnostic Multi-task Foundation Model for Human Brain Imaging

Peirong Liu, Oula Puonti, Xiaoling Hu, Karthik Gopinath, Annabel Sorby-Adams, Daniel C. Alexander, W. Taylor Kimberly, Juan E. Iglesias

Main category: cs.CV

TL;DR: BrainFM is a modality-agnostic vision foundation model for brain imaging that handles diverse MR contrasts and protocols through novel training strategies, enabling robust performance across five fundamental tasks without modality-specific tuning.

DetailsMotivation: Current learning-based methods work well for calibrated imaging like CT but fail to generalize across uncalibrated MR modalities due to sensitivity to contrast, resolution, and orientation differences, limiting clinical applicability.

Method: Uses ‘mild-to-severe’ intra-subject generation and ‘real-synth’ mix-up training strategy to build resilience against appearance variations (modality, contrast, deformation, resolution, artifacts).

Result: Demonstrates robust performance across eleven public datasets for five tasks: image synthesis (CT/T1w/T2w/FLAIR MRI), anatomy segmentation, scalp-to-cortical distance, bias field estimation, and registration.

Conclusion: BrainFM provides a versatile foundation model that can be directly applied to diverse brain imaging tasks across multiple modalities without task-specific tuning, enabling broad clinical protocol applicability.

Abstract: Recent learning-based approaches have made astonishing advances in calibrated medical imaging like computerized tomography (CT), yet they struggle to generalize in uncalibrated modalities – notably magnetic resonance (MR) imaging, where performance is highly sensitive to the differences in MR contrast, resolution, and orientation. This prevents broad applicability to diverse real-world clinical protocols. Here we introduce BrainFM, a modality-agnostic, multi-task vision foundation model for human brain imaging. With the proposed “mild-to-severe” intra-subject generation and “real-synth” mix-up training strategy, BrainFM is resilient to the appearance of acquired images (e.g., modality, contrast, deformation, resolution, artifacts), and can be directly applied to five fundamental brain imaging tasks, including image synthesis for CT and T1w/T2w/FLAIR MRI, anatomy segmentation, scalp-to-cortical distance, bias field estimation, and registration. We evaluate the efficacy of BrainFM on eleven public datasets, and demonstrate its robustness and effectiveness across all tasks and input modalities. Code is available at https://github.com/jhuldr/BrainFM.

[291] C-DiffDet+: Fusing Global Scene Context with Generative Denoising for High-Fidelity Object Detection

Abdellah Zakaria Sellam, Ilyes Benaissa, Salah Eddine Bekhouche, Abdenour Hadid, Vito Renó, Cosimo Distante

Main category: cs.CV

TL;DR: Introduces Context-Aware Fusion (CAF) to enhance DiffusionDet by integrating global scene context with local features using cross-attention, achieving state-of-the-art performance on vehicle damage detection.

DetailsMotivation: Fine-grained object detection in challenging domains like vehicle damage assessment is difficult even for human experts. DiffusionDet has limitations in context-dependent scenarios due to local feature conditioning.

Method: Proposes Context-Aware Fusion (CAF) using cross-attention mechanisms to integrate global scene context (from a separate encoder) with local proposal features, enabling each object proposal to attend to comprehensive environmental information.

Result: Experimental results show improvement over state-of-the-art models on the CarDD benchmark, establishing new performance benchmarks for context-aware object detection in fine-grained domains.

Conclusion: The framework significantly enhances the generative detection paradigm by enabling better context integration, making it particularly effective for challenging fine-grained object detection tasks.

Abstract: Fine-grained object detection in challenging visual domains, such as vehicle damage assessment, presents a formidable challenge even for human experts to resolve reliably. While DiffusionDet has advanced the state-of-the-art through conditional denoising diffusion, its performance remains limited by local feature conditioning in context-dependent scenarios. We address this fundamental limitation by introducing Context-Aware Fusion (CAF), which leverages cross-attention mechanisms to integrate global scene context with local proposal features directly. The global context is generated using a separate dedicated encoder that captures comprehensive environmental information, enabling each object proposal to attend to scene-level understanding. Our framework significantly enhances the generative detection paradigm by enabling each object proposal to attend to comprehensive environmental information. Experimental results demonstrate an improvement over state-of-the-art models on the CarDD benchmark, establishing new performance benchmarks for context-aware object detection in fine-grained domains

[292] DGL-RSIS: Decoupling Global Spatial Context and Local Class Semantics for Training-Free Remote Sensing Image Segmentation

Boyi Li, Ce Zhang, Richard M. Timmerman, Wenxuan Bao

Main category: cs.CV

TL;DR: DGL-RSIS is a training-free framework that decouples visual and textual inputs for remote sensing image segmentation, enabling open-vocabulary semantic segmentation and referring expression segmentation through global-local alignment strategies.

DetailsMotivation: Transferring vision language models from natural images to remote sensing segmentation is challenging due to limited category diversity in RS datasets and the domain gap between natural and RS imagery.

Method: Proposes a training-free framework with: 1) Global-local decoupling module separating text into class nouns/modifiers and images into mask proposals, 2) Local-scale alignment with context-aware cropping and RS-specific knowledge enrichment, 3) Global-scale alignment using Cross-Scale Grad-CAM and mask selection modules.

Result: Enables mask classification for open-vocabulary semantic segmentation and achieves accurate, interpretable alignment across global and local dimensions for referring expression segmentation.

Conclusion: DGL-RSIS effectively bridges the domain gap for remote sensing segmentation without requiring additional training, providing a novel approach for visual-language alignment in specialized domains.

Abstract: The emergence of vision language models (VLMs) has bridged vision and language, enabling joint multimodal understanding beyond traditional visual-only deep learning models. However, transferring VLMs from the natural image domain to remote sensing (RS) segmentation remains challenging due to the limited category diversity in RS datasets and the domain gap between natural and RS imagery. Here, we propose a training-free framework, DGL-RSIS, that decouples visual and textual inputs, performing visual-language alignment at both the local semantic and global contextual levels through tailored strategies. Specifically, we first introduce a global-local decoupling (GLD) module, where text inputs are divided into local class nouns and global modifiers using natural language processing (NLP) techniques; image inputs are partitioned into a set of class-agnostic mask proposals via unsupervised mask proposal networks. Second, visual and textual features are aligned at local scale, through a novel context-aware cropping strategy for extracting image patches with proper boundaries and introducing RS-specific knowledge to enrich the text inputs. By matching the enhanced text features with mask-guided visual features, we enable the mask classification, supporting open-vocabulary semantic segmentation (OVSS). Third, at the global scale, we propose a Cross-Scale Grad-CAM module to refine Grad-CAM maps using contextual information from global modifiers. A subsequent mask selection module integrates pixel-level Grad-CAM activations into the mask-level segmentation output, such that accurate and interpretable alignment can be realized across global and local dimensions for referring expression segmentation (RES).

[293] Towards Methane Detection Onboard Satellites

Maggie Chen, Hala Lambdouar, Luca Marini, Laura Martínez-Ferrer, Chris Bridges, Giacomo Acciarini

Main category: cs.CV

TL;DR: ML models trained on unorthorectified satellite data achieve comparable methane detection performance to orthorectified data, bypassing preprocessing steps while outperforming traditional matched filter methods.

DetailsMotivation: Methane is a potent greenhouse gas requiring rapid detection for climate mitigation. Current methods rely on computationally intensive preprocessing like orthorectification, which adds complexity and delays.

Method: Proposed UnorthoDOS approach using unorthorectified hyperspectral data from EMIT sensor. Trained ML models on both orthorectified and unorthorectified datasets, comparing against traditional matched filter baseline (mag1c).

Result: ML models on unorthorectified data achieved performance comparable to orthorectified data. Models on orthorectified data outperformed the matched filter baseline. Released datasets and code publicly available.

Conclusion: Bypassing orthorectification preprocessing is feasible for methane detection, enabling faster onboard satellite ML deployment with reduced computational costs while maintaining detection accuracy.

Abstract: Methane is a potent greenhouse gas and a major driver of climate change, making its timely detection critical for effective mitigation. Machine learning (ML) deployed onboard satellites can enable rapid detection while reducing downlink costs, supporting faster response systems. Conventional methane detection methods often rely on image processing techniques, such as orthorectification to correct geometric distortions and matched filters to enhance plume signals. We introduce a novel approach that bypasses these preprocessing steps by using \textit{unorthorectified} data (UnorthoDOS). We find that ML models trained on this dataset achieve performance comparable to those trained on orthorectified data. Moreover, we also train models on an orthorectified dataset, showing that they can outperform the matched filter baseline (mag1c). We release model checkpoints and two ML-ready datasets comprising orthorectified and unorthorectified hyperspectral images from the Earth Surface Mineral Dust Source Investigation (EMIT) sensor at https://huggingface.co/datasets/SpaceML/UnorthoDOS , along with code at https://github.com/spaceml-org/plume-hunter.

[294] MV-SSM: Multi-View State Space Modeling for 3D Human Pose Estimation

Aviral Chharia, Wenbo Gou, Haoye Dong

Main category: cs.CV

TL;DR: MV-SSM is a novel Multi-View State Space Modeling framework that improves 3D human pose estimation by better handling occlusions and generalizing to new camera configurations through state space modeling and grid token-guided bidirectional scanning.

DetailsMotivation: Existing attention-based transformers struggle with modeling spatial arrangements of keypoints in occluded scenarios and overfit to specific camera arrangements, leading to poor generalization to new settings.

Method: Proposes a Multi-View State Space Modeling framework with Projective State Space (PSS) blocks that model joint spatial sequences at feature and keypoint levels, using Grid Token-guided Bidirectional Scanning (GTBS) instead of traditional Mamba scanning.

Result: Achieves significant improvements: +10.8 AP25 (+24%) on CMU Panoptic three-camera setting, +7.0 AP25 (+13%) on varying camera arrangements, and +15.3 PCP (+38%) on Campus A1 in cross-dataset evaluations.

Conclusion: MV-SSM demonstrates strong generalization capabilities and outperforms state-of-the-art methods in multi-view 3D human pose estimation, particularly in challenging scenarios with occlusions and varying camera configurations.

Abstract: While significant progress has been made in single-view 3D human pose estimation, multi-view 3D human pose estimation remains challenging, particularly in terms of generalizing to new camera configurations. Existing attention-based transformers often struggle to accurately model the spatial arrangement of keypoints, especially in occluded scenarios. Additionally, they tend to overfit specific camera arrangements and visual scenes from training data, resulting in substantial performance drops in new settings. In this study, we introduce a novel Multi-View State Space Modeling framework, named MV-SSM, for robustly estimating 3D human keypoints. We explicitly model the joint spatial sequence at two distinct levels: the feature level from multi-view images and the person keypoint level. We propose a Projective State Space (PSS) block to learn a generalized representation of joint spatial arrangements using state space modeling. Moreover, we modify Mamba’s traditional scanning into an effective Grid Token-guided Bidirectional Scanning (GTBS), which is integral to the PSS block. Multiple experiments demonstrate that MV-SSM achieves strong generalization, outperforming state-of-the-art methods: +10.8 on AP25 (+24%) on the challenging three-camera setting in CMU Panoptic, +7.0 on AP25 (+13%) on varying camera arrangements, and +15.3 PCP (+38%) on Campus A1 in cross-dataset evaluations. Project Website: https://aviralchharia.github.io/MV-SSM

[295] Face4FairShifts: A Large Image Benchmark for Fairness and Robust Learning across Visual Domains

Yumeng Lin, Dong Li, Xintao Wu, Minglai Shao, Xujiang Zhao, Zhong Chen, Chen Zhao

Main category: cs.CV

TL;DR: Face4FairShifts is a large-scale facial image benchmark with 100,000 images across 4 domains and 39 annotations, designed to evaluate fairness-aware learning and domain generalization under distribution shifts.

DetailsMotivation: To address the challenge of ensuring fairness and robustness in machine learning models under domain shifts, and overcome limitations of existing datasets for fairness evaluation.

Method: Created a comprehensive benchmark dataset with 100,000 facial images across four visually distinct domains, annotated with 39 attributes within 14 categories covering demographic and facial features.

Result: Extensive experiments revealed significant performance gaps under distribution shifts, highlighting the limitations of current fairness-aware approaches and the need for better domain adaptation techniques.

Conclusion: Face4FairShifts provides a valuable testbed for advancing equitable and reliable AI systems, demonstrating the critical need for more effective fairness-aware domain adaptation methods in facial analysis applications.

Abstract: Ensuring fairness and robustness in machine learning models remains a challenge, particularly under domain shifts. We present Face4FairShifts, a large-scale facial image benchmark designed to systematically evaluate fairness-aware learning and domain generalization. The dataset includes 100,000 images across four visually distinct domains with 39 annotations within 14 attributes covering demographic and facial features. Through extensive experiments, we analyze model performance under distribution shifts and identify significant gaps. Our findings emphasize the limitations of existing related datasets and the need for more effective fairness-aware domain adaptation techniques. Face4FairShifts provides a comprehensive testbed for advancing equitable and reliable AI systems. The dataset is available online at https://meviuslab.github.io/Face4FairShifts/.

[296] Automatic Identification and Description of Jewelry Through Computer Vision and Neural Networks for Translators and Interpreters

Jose Manuel Alcalde-Llergo, Aurora Ruiz-Mezcua, Rocio Avila-Ramirez, Andrea Zingoni, Juri Taborri, Enrique Yeguas-Bolivar

Main category: cs.CV

TL;DR: Neural network approach for automatic jewelry identification and description using computer vision and image captioning across three hierarchical levels, achieving over 90% accuracy.

DetailsMotivation: Translators and interpreters need comprehensive jewelry understanding but currently rely on industry experts for precise descriptions. The challenge lies in the wide range of jewelry styles and designs that require expert knowledge.

Method: Employed computer vision techniques and image captioning with neural networks to detect jewels in images and generate natural language descriptions across three hierarchical levels. Used different image captioning architectures, particularly focusing on encoder-decoder models, and assembled a comprehensive jewelry image database for training and evaluation.

Result: The final model achieved captioning accuracy exceeding 90%, demonstrating effective recognition of diverse jewelry types and generation of accurate descriptions at varying detail levels.

Conclusion: The innovative approach successfully enables automatic jewelry identification and description, providing translators and interpreters with quick access to accurate information without relying on industry experts, effectively addressing the challenge of jewelry recognition across varied styles.

Abstract: Identifying jewelry pieces presents a significant challenge due to the wide range of styles and designs. Currently, precise descriptions are typically limited to industry experts. However, translators and interpreters often require a comprehensive understanding of these items. In this study, we introduce an innovative approach to automatically identify and describe jewelry using neural networks. This method enables translators and interpreters to quickly access accurate information, aiding in resolving queries and gaining essential knowledge about jewelry. Our model operates at three distinct levels of description, employing computer vision techniques and image captioning to emulate expert analysis of accessories. The key innovation involves generating natural language descriptions of jewelry across three hierarchical levels, capturing nuanced details of each piece. Different image captioning architectures are utilized to detect jewels in images and generate descriptions with varying levels of detail. To demonstrate the effectiveness of our approach in recognizing diverse types of jewelry, we assembled a comprehensive database of accessory images. The evaluation process involved comparing various image captioning architectures, focusing particularly on the encoder decoder model, crucial for generating descriptive captions. After thorough evaluation, our final model achieved a captioning accuracy exceeding 90 per cent.

[297] Fusion to Enhance: Fusion Visual Encoder to Enhance Multimodal Language Model

Yifei She, Huangxuan Wu

Main category: cs.CV

TL;DR: FtZ introduces a dual-encoder framework combining semantic and perception-focused encoders to address MLLMs’ fine-grained visual perception limitations, achieving superior performance on challenging benchmarks.

DetailsMotivation: MLLMs excel at high-level semantic understanding but fail at basic visual tasks requiring precise detail perception due to single-encoder architectures that sacrifice fine-grained visual information capture.

Method: FtZ uses a novel vision tower framework with a semantically powerful anchor encoder and perception-rich augmenting encoder connected via lightweight Multi-Head Cross-Attention mechanism.

Result: Significantly outperforms single-encoder baselines and existing feature fusion methods on TextVQA, POPE, MMMU, MME and MM-Vet benchmarks requiring fine-grained visual understanding.

Conclusion: Composing heterogeneous expert encoders is an efficient and effective approach to overcome visual perception bottlenecks in MLLMs, offering a new design paradigm for next-generation AI systems with stronger perceptual capabilities.

Abstract: Multimodal Large Language Models (MLLMs) have made significant progress in bridging visual perception with high-level textual reasoning. However, they face a fundamental contradiction: while excelling at complex semantic understanding, these models often fail at basic visual tasks that require precise detail perception. This deficiency primarily stems from the prevalent architectural reliance on a single vision encoder optimized for high-level semantic alignment, which inherently sacrifices the ability to capture fine-grained visual information. To address this issue, we introduce Fusion to Enhance (FtZ), a novel vision tower framework. FtZ moves beyond the single-encoder design by innovatively composing a semantically powerful anchor encoder with a perception-rich augmenting encoder via a lightweight Multi-Head Cross-Attention mechanism. Experimental results demonstrate that on several challenging benchmarks demanding fine-grained visual understanding, such as TextVQA, POPE, MMMU, MME and MM-Vet, our FtZ model significantly outperforms baselines that use only a single encoder or existing feature fusion methods. This work proves that composing heterogeneous expert encoders is an efficient and effective path to overcoming the visual perception bottleneck in current MLLMs, offering a new design paradigm for building next-generation AI systems with stronger perceptual capabilities.

[298] ER-LoRA: Effective-Rank Guided Adaptation for Weather-Generalized Depth Estimation

Weilong Yan, Xin Zhang, Robby T. Tan

Main category: cs.CV

TL;DR: Proposes STM strategy for weather-generalized monocular depth estimation using parameter-efficient fine-tuning of vision foundation models with minimal normal data

DetailsMotivation: Monocular depth estimation under adverse weather conditions is challenging due to lack of reliable ground truth and domain gaps in synthetic data. Self-supervised methods fail due to violated photometric assumptions in adverse scenarios.

Method: Introduces Selecting-Tuning-Maintaining (STM) strategy that structurally decomposes pretrained VFM weights using entropy-rank and stable-rank. Adaptively selects rank numbers and task-aware singular directions for initialization, with principal direction regularization to preserve generalization.

Result: Extensive experiments on four real-world benchmarks across diverse weather conditions show STM outperforms existing PEFT methods, full fine-tuning, methods trained with adverse synthetic data, and even depth foundation models.

Conclusion: STM enables effective weather-generalized depth estimation through parameter-efficient fine-tuning while preserving pretrained knowledge, demonstrating superior performance across adverse weather conditions with minimal normal data.

Abstract: Monocular depth estimation under adverse weather conditions (e.g.\ rain, fog, snow, and nighttime) remains highly challenging due to the lack of reliable ground truth and the difficulty of learning from unlabeled real-world data. Existing methods often rely on synthetic adverse data with pseudo-labels, which suffer from domain gaps, or employ self-supervised learning, which violates photometric assumptions in adverse scenarios. In this work, we propose to achieve weather–generalized depth estimation by Parameter–Efficient Fine–Tuning (PEFT) of Vision Foundation Models (VFMs), using only a small amount of high–visibility (normal) data. While PEFT has shown strong performance in semantic tasks such as segmentation, it remains underexplored for geometry–centric tasks like depth estimation – especially in terms of balancing effective adaptation with the preservation of pretrained knowledge. To this end, we introduce the Selecting–Tuning–Maintaining (STM) strategy, which structurally decomposes the pretrained weights of VFMs based on two kinds of effective ranks (entropy–rank and stable–rank). In the tuning phase, we adaptively select the proper rank number as well as the task–aware singular directions for initialization, based on the entropy–rank and full–tuned weight; while in the maintaining stage, we enforce a principal direction regularization based on the stable–rank. This design guarantees flexible task adaptation while preserving the strong generalization capability of the pretrained VFM. Extensive experiments on four real–world benchmarks across diverse weather conditions demonstrate that STM not only outperforms existing PEFT methods and full fine–tuning but also surpasses methods trained with adverse synthetic data, and even the depth foundation model

[299] LLaVA-Critic-R1: Your Critic Model is Secretly a Strong Policy Model

Xiyao Wang, Chunyuan Li, Jianwei Yang, Kai Zhang, Bo Liu, Tianyi Xiong, Furong Huang

Main category: cs.CV

TL;DR: Training multimodal critic models with RL on preference data produces unified models that excel at both evaluation and generation, outperforming specialized reasoning VLMs across multiple benchmarks.

DetailsMotivation: Challenge the conventional separation between critic models (for evaluation) and policy models (for generation) in vision-language modeling, exploring whether critics can also serve as effective policy models.

Method: Reorganize preference-labeled critic datasets into verifiable training signals and perform reinforcement learning directly on a base generative model to create LLaVA-Critic models that optimize preference judgments while retaining generation ability.

Result: LLaVA-Critic-R1 emerges as both top-performing critic and competitive policy model, achieving +5.7% average gain over base model across 26 benchmarks. LLaVA-Critic-R1+ achieves 71.9 SoTA on MMMU at 7B scale. Self-critique at test time yields +13.8% improvement on reasoning tasks.

Conclusion: RL training on critic data produces unified models excelling at both evaluation and generation, offering a simple path toward scalable, self-improving multimodal systems that break the traditional critic-policy separation.

Abstract: In vision-language modeling, critic models are typically trained to evaluate outputs – assigning scalar scores or pairwise preferences – rather than to generate responses. This separation from policy models, which produce the responses, is so entrenched that critics are rarely considered for direct policy use. In this work, we challenge this convention. We propose to reorganize preference-labeled critic datasets into verifiable training signals and perform reinforcement learning directly on a base generative model, producing LLaVA-Critic-R1, a multimodal critic trained to optimize preference judgments while retaining full generation ability. Surprisingly, LLaVA-Critic-R1 emerges not only as a top-performing critic but also as a competitive policy model – matching or surpassing specialized reasoning VLMs trained with in-domain data across 26 visual reasoning and understanding benchmarks, with an average gain of +5.7% over its base model (Qwen-2.5-VL-7B). Extending this approach to existing strong reasoning VLMs yields LLaVA-Critic-R1+, which further advances policy performance without sacrificing critic quality, achieving a SoTA performance of 71.9 on MMMU at the 7B scale. Finally, we show that the enhanced critic ability benefits inference: applying self-critique at test time yields an average +13.8% improvement on five representative reasoning tasks without additional training. Our results reveal that RL training on critic data can produce a unified model excelling at both evaluation and generation, offering a simple path toward scalable, self-improving multimodal systems.

[300] CSFMamba: Cross State Fusion Mamba Operator for Multimodal Remote Sensing Image Classification

Qingyu Wang, Xue Jiang, Guozheng Xu

Main category: cs.CV

TL;DR: CSFMamba Network combines Mamba’s efficiency with CNN for multimodal remote sensing image fusion, achieving better performance than Transformer with lower computational burden.

DetailsMotivation: Existing multimodal fusion methods suffer from quadratic computational complexity, making long-range dependency modeling burdensome. Mamba offers efficient computation but cannot perform direct feature fusion.

Method: Proposed Cross State Fusion Mamba (CSFMamba) Network with: 1) Preprocessing module for Mamba structure combined with CNN for multi-layer feature extraction, 2) Cross-state module based on Mamba operator for multimodal feature fusion, 3) Combined Mamba-CNN backbone for stronger full-image understanding.

Result: Experimental results on MUUFL and Houston2018 datasets show superior performance compared to Transformer while reducing network training burden.

Conclusion: CSFMamba effectively leverages Mamba’s computational efficiency and CNN’s feature extraction capabilities for improved multimodal remote sensing image classification with reduced computational complexity.

Abstract: Multimodal fusion has made great progress in the field of remote sensing image classification due to its ability to exploit the complementary spatial-spectral information. Deep learning methods such as CNN and Transformer have been widely used in these domains. State Space Models recently highlighted that prior methods suffer from quadratic computational complexity. As a result, modeling longer-range dependencies of spatial-spectral features imposes an overwhelming burden on the network. Mamba solves this problem by incorporating time-varying parameters into ordinary SSM and performing hardware optimization, but it cannot perform feature fusion directly. In order to make full use of Mamba’s low computational burden and explore the potential of internal structure in multimodal feature fusion, we propose Cross State Fusion Mamba (CSFMamba) Network. Specifically, we first design the preprocessing module of remote sensing image information for the needs of Mamba structure, and combine it with CNN to extract multi-layer features. Secondly, a cross-state module based on Mamba operator is creatively designed to fully fuse the feature of the two modalities. The advantages of Mamba and CNN are combined by designing a more powerful backbone. We capture the fusion relationship between HSI and LiDAR modalities with stronger full-image understanding. The experimental results on two datasets of MUUFL and Houston2018 show that the proposed method outperforms the experimental results of Transformer under the premise of reducing the network training burden.

[301] CascadeFormer: A Family of Two-stage Cascading Transformers for Skeleton-based Human Action Recognition

Yusen Peng, Alper Yilmaz

Main category: cs.CV

TL;DR: CascadeFormer: a two-stage transformer framework with masked pretraining and cascading fine-tuning for skeleton-based human action recognition, achieving competitive results on benchmark datasets.

DetailsMotivation: Recent advances in transformer models and masked pretraining frameworks provide new opportunities for representation learning in skeleton-based action recognition, moving beyond the dominant Graph Convolutional Networks (GCNs) approach.

Method: Proposes CascadeFormer - a family of two-stage cascading transformers: 1) masked pretraining stage to learn generalizable skeleton representations, 2) cascading fine-tuning stage tailored for discriminative action classification.

Result: Achieves competitive performance across three benchmark datasets: Penn Action, N-UCLA, and NTU RGB+D 60.

Conclusion: Transformer-based approaches with masked pretraining offer a promising alternative to GCNs for skeleton-based human action recognition, with code and model checkpoints released for reproducibility.

Abstract: Skeleton-based human action recognition leverages sequences of human joint coordinates to identify actions performed in videos. Owing to the intrinsic spatiotemporal structure of skeleton data, Graph Convolutional Networks (GCNs) have been the dominant architecture in this field. However, recent advances in transformer models and masked pretraining frameworks open new avenues for representation learning. In this work, we propose CascadeFormer, a family of two-stage cascading transformers for skeleton-based human action recognition. Our framework consists of a masked pretraining stage to learn generalizable skeleton representations, followed by a cascading fine-tuning stage tailored for discriminative action classification. We evaluate CascadeFormer across three benchmark datasets (Penn Action N-UCLA, and NTU RGB+D 60), achieving competitive performance on all tasks. To promote reproducibility, we release our code and model checkpoints.

[302] Prompt the Unseen: Evaluating Visual-Language Alignment Beyond Supervision

Raehyuk Jung, Seungjun Yu, Hyunjung Shim

Main category: cs.CV

TL;DR: The paper proposes a benchmark to evaluate projection layer generalization in Vision-Language Models, finding it retains 79-88% performance on unseen concepts and functions like a key-value memory.

DetailsMotivation: Despite the importance of projection layers in VLMs that map visual features to language embeddings, their ability to generalize to unseen visual concepts has not been systematically evaluated.

Method: Adapt object detection datasets with fine-grained annotations into prompting format, design train/test splits with disjoint label sets to control seen/unseen concept separation, and analyze through mechanistic interpretability.

Result: Projection layer retains about 79-88% of performance on unseen classes compared to seen ones across various settings, showing non-trivial generalization without explicit alignment supervision.

Conclusion: The study introduces a new evaluation framework for alignment generalization and highlights potential for efficient VLM training with limited aligned data, with the projection layer functioning like a key-value memory.

Abstract: Vision-Language Models (VLMs) combine a vision encoder and a large language model (LLM) through alignment training, showing strong performance on multimodal tasks. A central component in this architecture is the projection layer, which maps visual features into the LLM’s embedding space. Despite its importance, its ability to generalize to unseen visual concepts has not been systematically evaluated. To address this, we propose a benchmark for evaluating projection-layer generalization. We adapt object detection datasets (rich in fine-grained annotations) into a prompting format and design train/test splits with disjoint label sets, enabling precise control over seen and unseen concept separation. Experimental results show that the projection layer retains about 79 to 88 percent of the performance on unseen classes compared to seen ones across various settings, suggesting a non-trivial level of generalization even without explicit alignment supervision on those concepts. We further analyze this behavior through a mechanistic interpretability lens. Our findings indicate that the feed-forward network in the projection layer functions like a key-value memory, processing seen and unseen tokens in similar ways. This study introduces a new evaluation framework for alignment generalization and highlights the potential for efficient VLM training with limited aligned data.

[303] Enhancing Fairness in Skin Lesion Classification for Medical Diagnosis Using Prune Learning

Kuniko Paxton, Koorosh Aslansefat, Dhavalkumar Thakker, Yiannis Papadopoulos, Tanaya Maslekar

Main category: cs.CV

TL;DR: Proposes a fairness algorithm for skin lesion classification that reduces skin tone bias by analyzing feature map skewness in VGG and Vision Transformer networks, focusing on lesion areas instead of skin color.

DetailsMotivation: Address concerns about potential biases related to skin color in deep learning models for skin lesion classification, which can impact diagnostic outcomes and healthcare equity.

Method: Calculates skewness of feature maps in VGG convolution layers and Vision Transformer patches/heads to identify and reduce unnecessary channels related to skin tone, focusing attention on lesion areas rather than skin color.

Result: The approach lowers computational costs, mitigates bias without conventional statistical methods, potentially reduces model size while maintaining fairness, making it practical for real-world applications.

Conclusion: The proposed fairness algorithm effectively addresses skin tone bias in medical imaging models by focusing on lesion features rather than skin color characteristics, offering a computationally efficient solution for equitable healthcare diagnostics.

Abstract: Recent advances in deep learning have significantly improved the accuracy of skin lesion classification models, supporting medical diagnoses and promoting equitable healthcare. However, concerns remain about potential biases related to skin color, which can impact diagnostic outcomes. Ensuring fairness is challenging due to difficulties in classifying skin tones, high computational demands, and the complexity of objectively verifying fairness. To address these challenges, we propose a fairness algorithm for skin lesion classification that overcomes the challenges associated with achieving diagnostic fairness across varying skin tones. By calculating the skewness of the feature map in the convolution layer of the VGG (Visual Geometry Group) network and the patches and the heads of the Vision Transformer, our method reduces unnecessary channels related to skin tone, focusing instead on the lesion area. This approach lowers computational costs and mitigates bias without relying on conventional statistical methods. It potentially reduces model size while maintaining fairness, making it more practical for real-world applications.

[304] Causal Interpretation of Sparse Autoencoder Features in Vision

Sangyu Han, Yearim Kim, Nojun Kwak

Main category: cs.CV

TL;DR: CaFE method uses input-attribution to identify causal patches driving SAE feature activations in vision transformers, revealing that activation locations alone can be misleading and that features often require contextual co-occurrence.

DetailsMotivation: Traditional inspection of SAE features by looking at highest activation patches is flawed because self-attention mixes information across images, causing activated patches to co-occur with but not necessarily cause feature firing.

Method: Proposes Causal Feature Explanation (CaFE) which leverages Effective Receptive Field (ERF) and applies input-attribution methods to identify image patches that causally drive specific SAE feature activations.

Result: ERF maps frequently diverge from naive activation maps, revealing hidden context dependencies. Patch insertion tests confirm CaFE more effectively recovers/suppresses feature activations than activation-ranked patches.

Conclusion: CaFE provides more faithful and semantically precise explanations of vision-SAE features, highlighting the risk of misinterpretation when relying solely on activation location analysis.

Abstract: Understanding what sparse auto-encoder (SAE) features in vision transformers truly represent is usually done by inspecting the patches where a feature’s activation is highest. However, self-attention mixes information across the entire image, so an activated patch often co-occurs with-but does not cause-the feature’s firing. We propose Causal Feature Explanation (CaFE), which leverages Effective Receptive Field (ERF). We consider each activation of an SAE feature to be a target and apply input-attribution methods to identify the image patches that causally drive that activation. Across CLIP-ViT features, ERF maps frequently diverge from naive activation maps, revealing hidden context dependencies (e.g., a “roaring face” feature that requires the co-occurrence of eyes and nose, rather than merely an open mouth). Patch insertion tests confirm that CaFE more effectively recovers or suppresses feature activations than activation-ranked patches. Our results show that CaFE yields more faithful and semantically precise explanations of vision-SAE features, highlighting the risk of misinterpretation when relying solely on activation location.

[305] EVENT-Retriever: Event-Aware Multimodal Image Retrieval for Realistic Captions

Dinh-Khoi Vo, Van-Loc Nguyen, Minh-Triet Tran, Trung-Nghia Le

Main category: cs.CV

TL;DR: A multi-stage retrieval framework combining dense article retrieval, event-aware language model reranking, and caption-guided semantic matching achieves top performance in event-based image retrieval from free-form captions.

DetailsMotivation: Conventional vision-language retrieval approaches struggle with abstract events, implicit causality, temporal context, and complex narratives in free-form captions.

Method: Multi-stage framework using Qwen3 for article search, Qwen3-Reranker for contextual alignment, Qwen2-VL for image scoring, and Reciprocal Rank Fusion for output fusion.

Result: Achieved top-1 score on the private test set of Track 2 in the EVENTA 2025 Grand Challenge.

Conclusion: Combining language-based reasoning and multimodal retrieval is effective for complex, real-world image understanding tasks.

Abstract: Event-based image retrieval from free-form captions presents a significant challenge: models must understand not only visual features but also latent event semantics, context, and real-world knowledge. Conventional vision-language retrieval approaches often fall short when captions describe abstract events, implicit causality, temporal context, or contain long, complex narratives. To tackle these issues, we introduce a multi-stage retrieval framework combining dense article retrieval, event-aware language model reranking, and efficient image collection, followed by caption-guided semantic matching and rank-aware selection. We leverage Qwen3 for article search, Qwen3-Reranker for contextual alignment, and Qwen2-VL for precise image scoring. To further enhance performance and robustness, we fuse outputs from multiple configurations using Reciprocal Rank Fusion (RRF). Our system achieves the top-1 score on the private test set of Track 2 in the EVENTA 2025 Grand Challenge, demonstrating the effectiveness of combining language-based reasoning and multimodal retrieval for complex, real-world image understanding. The code is available at https://github.com/vdkhoi20/EVENT-Retriever.

[306] Multi-Level CLS Token Fusion for Contrastive Learning in Endoscopy Image Classification

Y Hop Nguyen, Doan Anh Phan Huu, Trung Thai Tran, Nhat Nam Mai, Van Toi Giap, Thao Thi Phuong Dao, Trung-Nghia Le

Main category: cs.CV

TL;DR: A unified vision-language framework for ENT endoscopy analysis that achieves state-of-the-art performance on classification and retrieval tasks using CLIP backbone enhanced with LoRA, multi-level token aggregation, and natural language prompts.

DetailsMotivation: Conventional CNN-based pipelines struggle to capture cross-modal semantics in medical imaging, particularly for ENT endoscopy where limited data and the need for multimodal understanding (visual inputs + diagnostic context) present challenges.

Method: Leverages CLIP ViT-B/16 backbone enhanced with Low-Rank Adaptation (LoRA), multi-level CLS token aggregation, spherical feature interpolation, and class-specific natural language prompts. Uses joint training combining supervised classification with contrastive learning.

Result: Achieved 95% accuracy and F1-score in classification, Recall@1 of 0.93 (image-to-image) and 0.92 (text-to-image), MRR scores of 0.97 and 0.96. Won ACM MM'25 ENTRep Grand Challenge.

Conclusion: The framework effectively addresses multimodal medical understanding in low-resource clinical settings, with ablation studies confirming the incremental benefits of each architectural component for robust ENT endoscopy analysis.

Abstract: We present a unified vision-language framework tailored for ENT endoscopy image analysis that simultaneously tackles three clinically-relevant tasks: image classification, image-to-image retrieval, and text-to-image retrieval. Unlike conventional CNN-based pipelines that struggle to capture cross-modal semantics, our approach leverages the CLIP ViT-B/16 backbone and enhances it through Low-Rank Adaptation, multi-level CLS token aggregation, and spherical feature interpolation. These components collectively enable efficient fine-tuning on limited medical data while improving representation diversity and semantic alignment across modalities. To bridge the gap between visual inputs and textual diagnostic context, we introduce class-specific natural language prompts that guide the image encoder through a joint training objective combining supervised classification with contrastive learning. We validated our framework through participation in the ACM MM'25 ENTRep Grand Challenge, achieving 95% accuracy and F1-score in classification, Recall@1 of 0.93 and 0.92 for image-to-image and text-to-image retrieval respectively, and MRR scores of 0.97 and 0.96. Ablation studies demonstrated the incremental benefits of each architectural component, validating the effectiveness of our design for robust multimodal medical understanding in low-resource clinical settings.

[307] MarkSplatter: Generalizable Watermarking for 3D Gaussian Splatting Model via Splatter Image Structure

Xiufeng Huang, Ziyuan Luo, Qi Song, Ruofei Wang, Renjie Wan

Main category: cs.CV

TL;DR: First generalizable watermarking framework for 3D Gaussian Splatting models that enables efficient copyright protection through single forward pass without fine-tuning.

DetailsMotivation: Growing need for effective copyright protection in 3DGS models, as current methods require computationally expensive fine-tuning for each predefined message.

Method: Introduces GaussianBridge to transform unstructured 3D Gaussians into Splatter Image format for neural processing, uses Gaussian-Uncertainty-Perceptual heatmap for imperceptibility, and dense segmentation-based extraction for robust message recovery.

Result: Enables arbitrary message embedding with preserved visual quality and reliable extraction even when watermarked objects occupy minimal regions in rendered views.

Conclusion: Proposes an efficient and robust watermarking solution for 3DGS models that eliminates the need for per-message fine-tuning while maintaining imperceptibility and extraction reliability.

Abstract: The growing popularity of 3D Gaussian Splatting (3DGS) has intensified the need for effective copyright protection. Current 3DGS watermarking methods rely on computationally expensive fine-tuning procedures for each predefined message. We propose the first generalizable watermarking framework that enables efficient protection of Splatter Image-based 3DGS models through a single forward pass. We introduce GaussianBridge that transforms unstructured 3D Gaussians into Splatter Image format, enabling direct neural processing for arbitrary message embedding. To ensure imperceptibility, we design a Gaussian-Uncertainty-Perceptual heatmap prediction strategy for preserving visual quality. For robust message recovery, we develop a dense segmentation-based extraction mechanism that maintains reliable extraction even when watermarked objects occupy minimal regions in rendered views. Project page: https://kevinhuangxf.github.io/marksplatter.

[308] No More Sibling Rivalry: Debiasing Human-Object Interaction Detection

Bin Yang, Yulin Zhang, Hong-Yu Zhou, Sibei Yang

Main category: cs.CV

TL;DR: The paper identifies ‘Toxic Siblings’ bias in HOI detection transformers where similar triplets interfere with each other, and proposes two debiasing objectives that significantly improve performance.

DetailsMotivation: Detection transformers for human-object interaction detection suffer from 'Toxic Siblings' bias where similar HOI triplets compete and interfere with each other, reducing precision despite high similarity.

Method: Proposes two debiasing objectives: 1) ‘contrastive-then-calibration’ for input perspective - samples incorrect sibling triplets and reconstructs them correctly using positional priors; 2) ‘merge-then-split’ for output perspective - learns shared features among siblings then refines intra-group differentiation.

Result: Significant performance improvements: +9.18% mAP over baseline and +3.59% mAP over state-of-the-art on HICO-Det dataset across various settings.

Conclusion: The proposed debiasing methods effectively address the Toxic Siblings bias in HOI detection transformers, demonstrating substantial performance gains over existing approaches.

Abstract: Detection transformers have been applied to human-object interaction (HOI) detection, enhancing the localization and recognition of human-action-object triplets in images. Despite remarkable progress, this study identifies a critical issue-“Toxic Siblings” bias-which hinders the interaction decoder’s learning, as numerous similar yet distinct HOI triplets interfere with and even compete against each other both input side and output side to the interaction decoder. This bias arises from high confusion among sibling triplets/categories, where increased similarity paradoxically reduces precision, as one’s gain comes at the expense of its toxic sibling’s decline. To address this, we propose two novel debiasing learning objectives-“contrastive-then-calibration” and “merge-then-split”-targeting the input and output perspectives, respectively. The former samples sibling-like incorrect HOI triplets and reconstructs them into correct ones, guided by strong positional priors. The latter first learns shared features among sibling categories to distinguish them from other groups, then explicitly refines intra-group differentiation to preserve uniqueness. Experiments show that we significantly outperform both the baseline (+9.18% mAP on HICO-Det) and the state-of-the-art (+3.59% mAP) across various settings.

[309] InterPose: Learning to Generate Human-Object Interactions from Large-Scale Web Videos

Yangsong Zhang, Abdul Ahad Butt, Gül Varol, Ivan Laptev

Main category: cs.CV

TL;DR: InterPose: A large-scale dataset of 73.8K human-object interaction motions extracted automatically from videos, enabling improved motion generation and zero-shot animation of people interacting with diverse objects.

DetailsMotivation: Existing motion generation works focus on isolated people in empty scenes, lacking realistic human-object interactions in complex 3D environments due to the absence of large-scale datasets with diverse object manipulations.

Method: Proposed an automatic motion extraction pipeline to collect interaction-rich human motions from 45.8K videos, creating InterPose dataset with 73.8K sequences of 3D human motions and corresponding text captions.

Result: InterPose brings significant improvements to state-of-the-art methods for human motion generation and enables development of an LLM-based agent for zero-shot animation of people interacting with diverse objects and scenes.

Conclusion: The InterPose dataset successfully addresses the lack of large-scale human-object interaction data and enables more realistic and versatile human motion generation in complex 3D scenes.

Abstract: Human motion generation has shown great advances thanks to the recent diffusion models trained on large-scale motion capture data. Most of existing works, however, currently target animation of isolated people in empty scenes. Meanwhile, synthesizing realistic human-object interactions in complex 3D scenes remains a critical challenge in computer graphics and robotics. One obstacle towards generating versatile high-fidelity human-object interactions is the lack of large-scale datasets with diverse object manipulations. Indeed, existing motion capture data is typically restricted to single people and manipulations of limited sets of objects. To address this issue, we propose an automatic motion extraction pipeline and use it to collect interaction-rich human motions. Our new dataset InterPose contains 73.8K sequences of 3D human motions and corresponding text captions automatically obtained from 45.8K videos with human-object interactions. We perform extensive experiments and demonstrate InterPose to bring significant improvements to state-of-the-art methods for human motion generation. Moreover, using InterPose we develop an LLM-based agent enabling zero-shot animation of people interacting with diverse objects and scenes.

[310] Secure and Scalable Face Retrieval via Cancelable Product Quantization

Haomiao Tang, Wenjie Li, Yixiang Qiu, Genping Wang, Shu-Tao Xia

Main category: cs.CV

TL;DR: A secure face retrieval framework using cancelable product quantization to protect user privacy while maintaining efficiency, addressing limitations of homomorphic encryption for real-time applications.

DetailsMotivation: Modern face retrieval systems often outsource retrieval to third parties, creating privacy risks for user portraits. Homomorphic encryption provides strong security but is computationally inefficient for real-time use.

Method: Proposes a two-stage framework: 1) cancelable PQ indexing for fast candidate filtering, and 2) cipher-space retrieval module for precise face ranking. Includes tailored protection mechanism for cancelable biometric authentication.

Result: Experiments on benchmark datasets show the method achieves a good balance between effectiveness, efficiency and security.

Conclusion: The proposed cancelable product quantization framework provides an efficient solution for secure face representation retrieval that addresses privacy concerns while maintaining practical performance.

Abstract: Despite the ubiquity of modern face retrieval systems, their retrieval stage is often outsourced to third-party entities, posing significant risks to user portrait privacy. Although homomorphic encryption (HE) offers strong security guarantees by enabling arithmetic computations in the cipher space, its high computational inefficiency makes it unsuitable for real-time, real-world applications. To address this issue, we propose Cancelable Product Quantization, a highly efficient framework for secure face representation retrieval. Our hierarchical two-stage framework comprises: (i) a high-throughput cancelable PQ indexing module for fast candidate filtering, and (ii) a fine-grained cipher-space retrieval module for final precise face ranking. A tailored protection mechanism is designed to secure the indexing module for cancelable biometric authentication while ensuring efficiency. Experiments on benchmark datasets demonstrate that our method achieves an decent balance between effectiveness, efficiency and security.

[311] Aligned Anchor Groups Guided Line Segment Detector

Zeyu Li, Annan Shu

Main category: cs.CV

TL;DR: AAGLSD is a novel line segment detector that uses aligned anchor groups and hierarchical pixel extraction to achieve high precision and completeness in line detection without complex refinement.

DetailsMotivation: To develop a more effective line segment detector that can extract complete line segments from images with high precision, addressing limitations of existing methods.

Method: Uses hierarchical approach with regular anchors and aligned anchor groups, sequentially linking anchors while updating predicted line segments, followed by simple validation and merging of adjacent segments.

Result: Quantitative experiments show AAGLSD effectively extracts complete line segments compared to other advanced detectors across various datasets.

Conclusion: AAGLSD provides an effective solution for precise and complete line segment detection without requiring complex refinement strategies, with code publicly available.

Abstract: This paper introduces a novel line segment detector, the Aligned Anchor Groups guided Line Segment Detector (AAGLSD), designed to detect line segments from images with high precision and completeness. The algorithm employs a hierarchical approach to extract candidate pixels with different saliency levels, including regular anchors and aligned anchor groups. AAGLSD initiates from these aligned anchor groups, sequentially linking anchors and updating the currently predicted line segment simultaneously. The final predictions are derived through straightforward validation and merging of adjacent line segments, avoiding complex refinement strategies. AAGLSD is evaluated on various datasets and quantitative experiments demonstrate that the proposed method can effectively extract complete line segments from input images compared to other advanced line segment detectors. The implementation is available at https://github.com/LLiDaBao/AAGLSD.

[312] Diffusion-Based Image-to-Brain Signal Generation with Cross-Attention Mechanisms for Visual Prostheses

Ganxi Xu, Jinyi Long, Jia Zhang

Main category: cs.CV

TL;DR: A novel image-to-brain framework using DDPMs with cross-attention to generate biologically plausible brain signals for visual prostheses, addressing the lack of supervised validation in existing approaches.

DetailsMotivation: Visual prostheses struggle to generate brain signals with sufficient biological similarity, as existing approaches lack supervised signals from real brain responses to validate biological plausibility of predicted stimuli.

Method: Proposes a DDPM-based framework with cross-attention mechanisms, consisting of a pre-trained CLIP visual encoder and cross-attention enhanced U-Net diffusion model that learns to reconstruct brain signals through iterative denoising.

Result: Evaluated on THINGS-EEG2 and THINGS-MEG datasets, demonstrating effectiveness in generating biologically plausible brain signals. Visualized training and test M/EEG topographies to show intra-subject and inter-subject variations.

Conclusion: The cross-attention enhanced diffusion framework successfully addresses the biological plausibility issue in visual prostheses by enabling dynamic interaction between visual features and brain signal representations.

Abstract: Visual prostheses have shown great potential in restoring vision for blind individuals. On the one hand, researchers have been continuously improving the brain decoding framework of visual prostheses by leveraging the powerful image generation capabilities of diffusion models. On the other hand, the brain encoding stage of visual prostheses struggles to generate brain signals with sufficient biological similarity. Although existing works have recognized this problem, the quality of predicted stimuli still remains a critical issue, as existing approaches typically lack supervised signals from real brain responses to validate the biological plausibility of predicted stimuli. To address this issue, we propose a novel image-to-brain framework based on denoising diffusion probabilistic models (DDPMs) enhanced with cross-attention mechanisms. Our framework consists of two key architectural components: a pre-trained CLIP visual encoder that extracts rich semantic representations from input images, and a cross-attention enhanced U-Net diffusion model that learns to reconstruct biologically plausible brain signals through iterative denoising. Unlike conventional generative models that rely on simple concatenation for conditioning, our cross-attention modules enable dynamic interaction between visual features and brain signal representations, facilitating fine-grained alignment during the generation process. We evaluate our framework on two multimodal datasets (THINGS-EEG2 and THINGS-MEG) to demonstrate its effectiveness in generating biologically plausible brain signals. Moreover, we visualize the training and test M/EEG topographies for all subjects on both datasets to intuitively demonstrate the intra-subject variations and inter-subject variations in M/EEG signals.

[313] OmniReason: A Temporal-Guided Vision-Language-Action Framework for Autonomous Driving

Pei Liu, Qingtian Ning, Xinyan Lu, Haipeng Liu, Weiliang Ma, Dangen She, Peng Jia, Xianpeng Lang, Jun Ma

Main category: cs.CV

TL;DR: OmniReason framework introduces spatiotemporal reasoning for autonomous driving with large-scale VLA datasets and a novel agent architecture that achieves state-of-the-art performance in planning and VQA tasks.

DetailsMotivation: Existing vision-language models focus on static scene understanding but neglect the temporal dimension essential for real-world driving scenarios, creating a critical limitation in autonomous vehicle systems.

Method: Developed OmniReason-Data (large-scale VLA datasets with dense spatiotemporal annotations) using hallucination-mitigated auto-labeling, and OmniReason-Agent architecture with sparse temporal memory and explanation generator using spatiotemporal knowledge distillation.

Result: Achieved state-of-the-art performance with significant improvements in open-loop planning tasks and visual question answering benchmarks, establishing new capabilities for interpretable, temporally-aware autonomous vehicles.

Conclusion: OmniReason framework successfully addresses the temporal reasoning gap in autonomous driving by providing robust spatiotemporal modeling and human-interpretable decision rationales for complex dynamic environments.

Abstract: Recent advances in vision-language models (VLMs) have demonstrated impressive spatial reasoning capabilities for autonomous driving, yet existing methods predominantly focus on static scene understanding while neglecting the essential temporal dimension of real-world driving scenarios. To address this critical limitation, we propose the OmniReason framework, which establishes robust spatiotemporal reasoning by jointly modeling dynamic 3D environments and their underlying decision-making processes. Our work makes two fundamental advances: (1) We introduce OmniReason-Data, two large-scale vision-language-action (VLA) datasets with dense spatiotemporal annotations and natural language explanations, generated through a novel hallucination-mitigated auto-labeling pipeline that ensures both physical plausibility and temporal coherence; (2) We develop the OmniReason-Agent architecture, which integrates a sparse temporal memory module for persistent scene context modeling and an explanation generator that produces human-interpretable decision rationales, facilitated by our spatiotemporal knowledge distillation approach that effectively captures spatiotemporal causal reasoning patterns. Comprehensive experiments demonstrate state-of-the-art performance, where OmniReason-Agent achieves significant improvements in both open-loop planning tasks and visual question answering (VQA) benchmarks, while establishing new capabilities for interpretable, temporally-aware autonomous vehicles operating in complex, dynamic environments.

[314] Multimodal Iterative RAG for Knowledge Visual Question Answering

Changin Choi, Wonseok Lee, Jungmin Ko, Wonjong Rhee

Main category: cs.CV

TL;DR: MI-RAG is a multimodal iterative RAG framework that enhances knowledge retrieval and reasoning for visual question answering by using accumulated reasoning to dynamically formulate multi-queries across heterogeneous knowledge bases.

DetailsMotivation: Multimodal LLMs struggle with knowledge-intensive visual questions requiring external knowledge beyond images. Conventional single-pass RAG frameworks often fail to gather sufficient knowledge for these complex tasks.

Method: Proposes MI-RAG framework that iteratively leverages accumulated reasoning records to formulate multi-queries, performs joint search across visually-grounded and textual knowledge bases, and synthesizes new knowledge into reasoning records across multiple iterations.

Result: Significant improvements in both retrieval recall and answer accuracy on challenging benchmarks including Encyclopedic VQA, InfoSeek, and OK-VQA.

Conclusion: MI-RAG establishes a scalable approach for compositional reasoning in knowledge-intensive VQA by progressively refining understanding through iterative retrieval and reasoning across modalities.

Abstract: While Multimodal Large Language Models (MLLMs) have significantly advanced multimodal understanding, their performance remains limited on knowledge-intensive visual questions that require external knowledge beyond the image. Retrieval-Augmented Generation (RAG) has become a promising solution for providing models with external knowledge, its conventional single-pass framework often fails to gather sufficient knowledge. To overcome this limitation, we propose MI-RAG, a Multimodal Iterative RAG framework that leverages reasoning to enhance retrieval and update reasoning over newly retrieved knowledge across modalities. At each iteration, MI-RAG leverages an accumulated reasoning record to dynamically formulate a multi-query. These queries then drive a joint search across heterogeneous knowledge bases containing both visually-grounded and textual knowledge. The newly acquired knowledge is synthesized into the reasoning record, progressively refining understanding across iterations. Experiments on challenging benchmarks, including Encyclopedic VQA, InfoSeek, and OK-VQA, show that MI-RAG significantly improves both retrieval recall and answer accuracy, establishing a scalable approach for compositional reasoning in knowledge-intensive VQA.

[315] SWAGSplatting: Semantic-guided Water-scene Augmented Gaussian Splatting

Zhuodong Jiang, Haoran Wang, Guoxi Huang, Brett Seymour, Nantheera Anantrasirichai

Main category: cs.CV

TL;DR: Novel semantic-guided 3D Gaussian Splatting framework for underwater 3D reconstruction using multimodal cross-knowledge and CLIP semantic features, achieving state-of-the-art results.

DetailsMotivation: Underwater 3D reconstruction faces challenges like light distortion and turbidity. Existing AI methods don't fully exploit language models with visual processing for semantic understanding.

Method: Embed semantic features into Gaussian primitives supervised by CLIP, use semantic consistency loss, and implement stage-wise training with coarse-to-fine learning and late-stage parameter refinement.

Result: Outperforms state-of-the-art methods on SeaThru-NeRF and Submerged3D datasets with up to 3.09 dB PSNR improvement on average.

Conclusion: The proposed framework provides robust and high-fidelity deep-sea scene reconstruction, making it suitable for underwater exploration and marine perception applications.

Abstract: Accurate 3D reconstruction in underwater environments remains a complex challenge due to issues such as light distortion, turbidity, and limited visibility. AI-based techniques have been applied to address these issues, however, existing methods have yet to fully exploit the potential of AI, particularly in integrating language models with visual processing. In this paper, we propose a novel framework that leverages multimodal cross-knowledge to create semantic-guided 3D Gaussian Splatting for robust and high-fidelity deep-sea scene reconstruction. By embedding an extra semantic feature into each Gaussian primitive and supervised by the CLIP extracted semantic feature, our method enforces semantic and structural awareness throughout the training. The dedicated semantic consistency loss ensures alignment with high-level scene understanding. Besides, we propose a novel stage-wise training strategy, combining coarse-to-fine learning with late-stage parameter refinement, to further enhance both stability and reconstruction quality. Extensive results show that our approach consistently outperforms state-of-the-art methods on SeaThru-NeRF and Submerged3D datasets across three metrics, with an improvement of up to 3.09 dB on average in terms of PSNR, making it a strong candidate for applications in underwater exploration and marine perception.

[316] Adaptive Contrast Adjustment Module: A Clinically-Inspired Plug-and-Play Approach for Enhanced Fetal Plane Classification

Yang Chen, Sanglin Zhao, Baoyu Chen, Mans Gustaf

Main category: cs.CV

TL;DR: A plug-and-play adaptive contrast adjustment module (ACAM) improves fetal ultrasound classification by mimicking clinical contrast adjustment practices, enhancing various models’ accuracy by 1-2% on multi-center data.

DetailsMotivation: Fetal ultrasound classification faces challenges like low tissue contrast, boundary ambiguity, and operator-dependent image quality variations that limit diagnostic reliability.

Method: ACAM uses a shallow texture-sensitive network to predict clinical contrast parameters, creates multiple contrast-enhanced views through differentiable mapping, and fuses them within downstream classifiers.

Result: Validated on 12,400 multi-center images across six anatomical categories, ACAM improved accuracy: lightweight models +2.02%, traditional models +1.29%, state-of-the-art models +1.15%.

Conclusion: The module bridges low-level image features with high-level semantics through physics-informed transformations aligned with clinical workflows, establishing a new paradigm for robust medical image analysis.

Abstract: Fetal ultrasound standard plane classification is essential for reliable prenatal diagnosis but faces inherent challenges, including low tissue contrast, boundary ambiguity, and operator-dependent image quality variations. To overcome these limitations, we propose a plug-and-play adaptive contrast adjustment module (ACAM), whose core design is inspired by the clinical practice of doctors adjusting image contrast to obtain clearer and more discriminative structural information. The module employs a shallow texture-sensitive network to predict clinically plausible contrast parameters, transforms input images into multiple contrast-enhanced views through differentiable mapping, and fuses them within downstream classifiers. Validated on a multi-center dataset of 12,400 images across six anatomical categories, the module consistently improves performance across diverse models, with accuracy of lightweight models increasing by 2.02 percent, accuracy of traditional models increasing by 1.29 percent, and accuracy of state-of-the-art models increasing by 1.15 percent. The innovation of the module lies in its content-aware adaptation capability, replacing random preprocessing with physics-informed transformations that align with sonographer workflows while improving robustness to imaging heterogeneity through multi-view fusion. This approach effectively bridges low-level image features with high-level semantics, establishing a new paradigm for medical image analysis under real-world image quality variations.

[317] IPG: Incremental Patch Generation for Generalized Adversarial Patch Training

Wonho Lee, Hyunsik Na, Jisu Lee, Daeseon Choi

Main category: cs.CV

TL;DR: IPG method generates adversarial patches 11.1x more efficiently than existing approaches while maintaining comparable attack performance, with applications in AI security and real-world systems.

DetailsMotivation: Adversarial patches pose significant threats to AI model robustness in computer vision tasks like object detection, requiring more efficient generation methods to address model vulnerabilities.

Method: Incremental Patch Generation (IPG) - a novel approach that efficiently generates adversarial patches through systematic incremental generation process.

Result: IPG achieves 11.1x efficiency improvement over existing methods while maintaining comparable attack performance, producing well-generalized patches that cover broader model vulnerabilities.

Conclusion: IPG has significant potential for adversarial patch defense and real-world applications including autonomous vehicles, security systems, and medical imaging where AI resilience is critical.

Abstract: The advent of adversarial patches poses a significant challenge to the robustness of AI models, particularly in the domain of computer vision tasks such as object detection. In contradistinction to traditional adversarial examples, these patches target specific regions of an image, resulting in the malfunction of AI models. This paper proposes Incremental Patch Generation (IPG), a method that generates adversarial patches up to 11.1 times more efficiently than existing approaches while maintaining comparable attack performance. The efficacy of IPG is demonstrated by experiments and ablation studies including YOLO’s feature distribution visualization and adversarial training results, which show that it produces well-generalized patches that effectively cover a broader range of model vulnerabilities. Furthermore, IPG-generated datasets can serve as a robust knowledge foundation for constructing a robust model, enabling structured representation, advanced reasoning, and proactive defenses in AI security ecosystems. The findings of this study suggest that IPG has considerable potential for future utilization not only in adversarial patch defense but also in real-world applications such as autonomous vehicles, security systems, and medical imaging, where AI models must remain resilient to adversarial attacks in dynamic and high-stakes environments.

[318] Sequential Difference Maximization: Generating Adversarial Examples via Multi-Stage Optimization

Xinlei Liu, Tao Hu, Peng Yi, Weitao Han, Jichao Xie, Baolin Li

Main category: cs.CV

TL;DR: Proposes SDM, a gradient-based adversarial attack method that maximizes the difference between non-true labels’ probability upper bound and true label’s probability through a three-layer optimization framework.

DetailsMotivation: To develop more efficient adversarial attack methods for assessing computer vision model robustness by reconstructing the optimization objective for generating adversarial examples.

Method: Sequential Difference Maximization (SDM) with three-layer ‘cycle-stage-step’ optimization framework. Uses negative probability of true label initially to compress solution space, then introduces Directional Probability Difference Ratio (DPDR) loss to gradually increase non-true labels’ probability upper bound.

Result: SDM demonstrates stronger attack performance and higher attack cost-effectiveness compared to previous SOTA methods. Can be combined with adversarial training to enhance defensive effects.

Conclusion: SDM provides an effective gradient-based adversarial attack method that improves both attack performance and efficiency while being compatible with adversarial training defenses.

Abstract: Efficient adversarial attack methods are critical for assessing the robustness of computer vision models. In this paper, we reconstruct the optimization objective for generating adversarial examples as “maximizing the difference between the non-true labels’ probability upper bound and the true label’s probability,” and propose a gradient-based attack method termed Sequential Difference Maximization (SDM). SDM establishes a three-layer optimization framework of “cycle-stage-step.” The processes between cycles and between iterative steps are respectively identical, while optimization stages differ in terms of loss functions: in the initial stage, the negative probability of the true label is used as the loss function to compress the solution space; in subsequent stages, we introduce the Directional Probability Difference Ratio (DPDR) loss function to gradually increase the non-true labels’ probability upper bound by compressing the irrelevant labels’ probabilities. Experiments demonstrate that compared with previous SOTA methods, SDM not only exhibits stronger attack performance but also achieves higher attack cost-effectiveness. Additionally, SDM can be combined with adversarial training methods to enhance their defensive effects. The code is available at https://github.com/X-L-Liu/SDM.

[319] Surface Defect Detection with Gabor Filter Using Reconstruction-Based Blurring U-Net-ViT

Jongwook Si, Sungyoung Kim

Main category: cs.CV

TL;DR: Novel texture defect detection using Gabor filters and blurring U-Net-ViT model with Gaussian loss function and SP masking, achieving 0.939 AUC on multiple datasets.

DetailsMotivation: To enhance accuracy and reliability of texture-based surface defect detection by combining local and global feature processing while handling noisy environments.

Method: Combines U-Net’s local feature training with Vision Transformer’s global processing, uses Gaussian filter-based loss for noise removal, SP masking for boundary reinforcement, and Gabor filters in post-processing for defect orientation/frequency emphasis.

Result: Achieved average AUC of 0.939 across MVTec-AD, Surface Crack Detection, and Marble Surface Anomaly datasets, with ablation studies confirming optimal filter size and noise probability enhance performance.

Conclusion: The proposed approach effectively detects surface defects across various textures with robust performance in noisy environments through optimized parameter settings and combined local-global processing.

Abstract: This paper proposes a novel approach to enhance the accuracy and reliability of texture-based surface defect detection using Gabor filters and a blurring U-Net-ViT model. By combining the local feature training of U-Net with the global processing of the Vision Transformer(ViT), the model effectively detects defects across various textures. A Gaussian filter-based loss function removes background noise and highlights defect patterns, while Salt-and-Pepper(SP) masking in the training process reinforces texture-defect boundaries, ensuring robust performance in noisy environments. Gabor filters are applied in post-processing to emphasize defect orientation and frequency characteristics. Parameter optimization, including filter size, sigma, wavelength, gamma, and orientation, maximizes performance across datasets like MVTec-AD, Surface Crack Detection, and Marble Surface Anomaly Dataset, achieving an average Area Under the Curve(AUC) of 0.939. The ablation studies validate that the optimal filter size and noise probability significantly enhance defect detection performance.

[320] UPGS: Unified Pose-aware Gaussian Splatting for Dynamic Scene Deblurring

Zhijing Wu, Longguang Wang

Main category: cs.CV

TL;DR: A unified optimization framework that jointly optimizes camera poses and 3D Gaussian attributes to handle motion blur in dynamic 3D scene reconstruction from monocular video.

DetailsMotivation: Existing methods fail with motion blur because they use a two-step pipeline where pose estimation errors accumulate and degrade 3D reconstruction quality.

Method: Proposes end-to-end optimization with camera poses as learnable parameters, models motion as per-primitive SE(3) transformations on 3D Gaussians, and uses a three-stage training schedule alternating between pose and Gaussian optimization.

Result: Achieves significant improvements in reconstruction quality and pose estimation accuracy on Stereo Blur dataset and real-world sequences compared to prior methods.

Conclusion: Joint optimization of camera poses and 3DGS attributes effectively handles motion blur and produces superior dynamic 3D scene reconstruction.

Abstract: Reconstructing dynamic 3D scenes from monocular video has broad applications in AR/VR, robotics, and autonomous navigation, but often fails due to severe motion blur caused by camera and object motion. Existing methods commonly follow a two-step pipeline, where camera poses are first estimated and then 3D Gaussians are optimized. Since blurring artifacts usually undermine pose estimation, pose errors could be accumulated to produce inferior reconstruction results. To address this issue, we introduce a unified optimization framework by incorporating camera poses as learnable parameters complementary to 3DGS attributes for end-to-end optimization. Specifically, we recast camera and object motion as per-primitive SE(3) affine transformations on 3D Gaussians and formulate a unified optimization objective. For stable optimization, we introduce a three-stage training schedule that optimizes camera poses and Gaussians alternatively. Particularly, 3D Gaussians are first trained with poses being fixed, and then poses are optimized with 3D Gaussians being untouched. Finally, all learnable parameters are optimized together. Extensive experiments on the Stereo Blur dataset and challenging real-world sequences demonstrate that our method achieves significant gains in reconstruction quality and pose estimation accuracy over prior dynamic deblurring methods.

[321] SegDINO: An Efficient Design for Medical and Natural Image Segmentation with DINO-V3

Sicheng Yang, Hongqiu Wang, Zhaohu Xing, Sixiang Chen, Lei Zhu

Main category: cs.CV

TL;DR: SegDINO is an efficient segmentation framework that combines a frozen DINOv3 backbone with a lightweight MLP decoder, achieving state-of-the-art performance across multiple benchmarks with minimal parameters.

DetailsMotivation: Existing approaches for adapting DINO self-supervised vision models to segmentation tasks rely on heavy decoders with multi-scale fusion or complex upsampling, which introduce substantial parameter overhead and computational cost.

Method: SegDINO extracts multi-level features from a frozen pretrained DINOv3 encoder, aligns them to a common resolution and channel width, and uses a lightweight MLP head to directly predict segmentation masks.

Result: Extensive experiments across six benchmarks (three medical datasets: TN3K, Kvasir-SEG, ISIC; three natural image datasets: MSD, VMD-D, ViSha) demonstrate that SegDINO consistently achieves state-of-the-art performance compared to existing methods.

Conclusion: SegDINO provides an efficient segmentation framework that minimizes trainable parameters while preserving the representational power of foundation features from DINOv3, offering strong performance across diverse segmentation tasks.

Abstract: The DINO family of self-supervised vision models has shown remarkable transferability, yet effectively adapting their representations for segmentation remains challenging. Existing approaches often rely on heavy decoders with multi-scale fusion or complex upsampling, which introduce substantial parameter overhead and computational cost. In this work, we propose SegDINO, an efficient segmentation framework that couples a frozen DINOv3 backbone with a lightweight decoder. SegDINO extracts multi-level features from the pretrained encoder, aligns them to a common resolution and channel width, and utilizes a lightweight MLP head to directly predict segmentation masks. This design minimizes trainable parameters while preserving the representational power of foundation features. Extensive experiments across six benchmarks, including three medical datasets (TN3K, Kvasir-SEG, ISIC) and three natural image datasets (MSD, VMD-D, ViSha), demonstrate that SegDINO consistently achieves state-of-the-art performance compared to existing methods. Code is available at https://github.com/script-Yang/SegDINO.

[322] Satellite Image Utilization for Dehazing with Swin Transformer-Hybrid U-Net and Watershed loss

Jongwook Si, Sungyoung Kim

Main category: cs.CV

TL;DR: Proposes SUFERNOBWA, a hybrid dehazing framework combining Swin Transformer and U-Net for satellite image restoration, achieving state-of-the-art results on benchmark datasets.

DetailsMotivation: Atmospheric interference and haze degrade satellite image clarity and reduce information extraction accuracy, requiring effective dehazing solutions.

Method: Hybrid framework integrating Swin Transformer and U-Net with SwinRRDB blocks for global context and local detail learning, plus composite loss function (L2 + guided + watershed loss) for structural preservation.

Result: Outperforms state-of-the-art models on RICE and SateHaze1K datasets, achieving PSNR of 33.24 dB and SSIM of 0.967 on RICE dataset.

Conclusion: Provides effective solution for mitigating atmospheric interference in satellite imagery with potential applicability across diverse remote sensing applications.

Abstract: Satellite imagery plays a crucial role in various fields; however, atmospheric interference and haze significantly degrade image clarity and reduce the accuracy of information extraction. To address these challenges, this paper proposes a hybrid dehazing framework that integrates Swin Transformer and U-Net to balance global context learning and local detail restoration, called SUFERNOBWA. The proposed network employs SwinRRDB, a Swin Transformer-based Residual-in-Residual Dense Block, in both the encoder and decoder to effectively extract features. This module enables the joint learning of global contextual information and fine spatial structures, which is crucial for structural preservation in satellite image. Furthermore, we introduce a composite loss function that combines L2 loss, guided loss, and a novel watershed loss, which enhances structural boundary preservation and ensures pixel-level accuracy. This architecture enables robust dehazing under diverse atmospheric conditions while maintaining structural consistency across restored images. Experimental results demonstrate that the proposed method outperforms state-of-the-art models on both the RICE and SateHaze1K datasets. Specifically, on the RICE dataset, the proposed approach achieved a PSNR of 33.24 dB and an SSIM of 0.967, which is a significant improvement over existing method. This study provides an effective solution for mitigating atmospheric interference in satellite imagery and highlights its potential applicability across diverse remote sensing applications.

[323] Look Beyond: Two-Stage Scene View Generation via Panorama and Video Diffusion

Xueyang Kang, Zhengkang Xiang, Zezheng Zhang, Kourosh Khoshelham

Main category: cs.CV

TL;DR: A novel view synthesis method that decomposes single-view synthesis into 360-degree scene extrapolation followed by view interpolation using panoramic representations and spatial noise diffusion.

DetailsMotivation: Existing single-view novel view synthesis methods fail to maintain coherence and correct view alignment across long-range or looped trajectories, especially for views significantly deviating from the input.

Method: Two-stage approach: 1) 360-degree scene extrapolation using panorama diffusion model, 2) novel view interpolation using perspective keyframes warped from panorama and spatial noise diffusion in a video diffusion model.

Result: Outperforms existing methods in generating coherent views along user-defined trajectories, maintains global consistency even in loop closure scenarios, and enables flexible camera control.

Conclusion: The proposed decomposition approach effectively addresses long-term view and scene consistency challenges in single-view novel view synthesis, producing superior results compared to prior work.

Abstract: Novel view synthesis (NVS) from a single image is highly ill-posed due to large unobserved regions, especially for views that deviate significantly from the input. While existing methods focus on consistency between the source and generated views, they often fail to maintain coherence and correct view alignment across long-range or looped trajectories. We propose a model that addresses this by decomposing single-view NVS into a 360-degree scene extrapolation followed by novel view interpolation. This design ensures long-term view and scene consistency by conditioning on keyframes extracted and warped from a generated panoramic representation. In the first stage, a panorama diffusion model learns the scene prior from the input perspective image. Perspective keyframes are then sampled and warped from the panorama and used as anchor frames in a pre-trained video diffusion model, which generates novel views through a proposed spatial noise diffusion process. Compared to prior work, our method produces globally consistent novel views – even in loop closure scenarios – while enabling flexible camera control. Experiments on diverse scene datasets demonstrate that our approach outperforms existing methods in generating coherent views along user-defined trajectories. Our implementation is available at https://github.com/YiGuYT/LookBeyond.

[324] Quantization Meets OOD: Generalizable Quantization-aware Training from a Flatness Perspective

Jiacheng Jiang, Yuan Meng, Chen Tang, Han Yu, Qun Li, Zhi Wang, Wenwu Zhu

Main category: cs.CV

TL;DR: FQAT is a flatness-oriented quantization-aware training method that addresses OOD generalization degradation in quantized models through layer-wise freezing and adaptive optimization.

DetailsMotivation: Current QAT methods focus only on in-distribution performance but cause significant out-of-distribution performance degradation, creating a need for methods that maintain OOD generalization.

Method: Proposes FQAT with layer-wise freezing mechanism to mitigate gradient conflicts and disorder-guided adaptive freezing algorithm with gradient disorder metric to dynamically identify and freeze unstable layers.

Result: Extensive experiments show FQAT outperforms state-of-the-art baselines on both I.D and OOD image classification tasks across influential benchmarks.

Conclusion: FQAT successfully addresses the OOD generalization problem in QAT by incorporating flatness optimization through intelligent layer freezing strategies, achieving superior performance on both distribution types.

Abstract: Current quantization-aware training (QAT) methods primarily focus on enhancing the performance of quantized models on in-distribution (I.D) data, while overlooking the potential performance degradation on out-of-distribution (OOD) data. In this paper, we first substantiate this problem through rigorous experiment, showing that QAT can lead to a significant OOD generalization performance degradation. Further, we find the contradiction between the perspective that flatness of loss landscape gives rise to superior OOD generalization and the phenomenon that QAT lead to a sharp loss landscape, can cause the above problem. Therefore, we propose a flatness-oriented QAT method, FQAT, to achieve generalizable QAT. Specifically, i) FQAT introduces a layer-wise freezing mechanism to mitigate the gradient conflict issue between dual optimization objectives (i.e., vanilla QAT and flatness). ii) FQAT proposes an disorder-guided adaptive freezing algorithm to dynamically determines which layers to freeze at each training step, effectively addressing the challenges caused by interference between layers. A gradient disorder metric is designed to help the algorithm identify unstable layers during training. Extensive experiments on influential OOD benchmark demonstrate the superiority of our method over state-of-the-art baselines under both I.D and OOD image classification tasks.

[325] Pose as Clinical Prior: Learning Dual Representations for Scoliosis Screening

Zirui Zhou, Zizhao Peng, Dongyang Jin, Chao Fan, Fengwei An, Shiqi Yu

Main category: cs.CV

TL;DR: Proposes Scoliosis1K-Pose dataset and Dual Representation Framework (DRF) for pose-based scoliosis screening, addressing data scarcity and noise sensitivity in raw pose coordinates through continuous skeleton maps and clinical asymmetry encoding.

DetailsMotivation: Current AI scoliosis screening relies on silhouette data, missing clinically relevant postural asymmetries. Pose data offers better clinical interpretability but faces challenges with dataset scarcity and noise sensitivity in raw coordinates.

Method: Introduces Scoliosis1K-Pose dataset with 447,900 pose frames from 1,050 adolescents. Develops Dual Representation Framework (DRF) combining continuous skeleton maps with discrete Postural Asymmetry Vector (PAV), enhanced by PAV-Guided Attention module for clinical prior integration.

Result: DRF achieves state-of-the-art performance. Visualizations confirm the model effectively uses clinical asymmetry cues to guide feature extraction and creates synergy between dual representations.

Conclusion: The proposed framework successfully addresses pose-based scoliosis screening limitations by integrating clinical knowledge with deep learning, providing improved interpretability and performance. Dataset and code are publicly available.

Abstract: Recent AI-based scoliosis screening methods primarily rely on large-scale silhouette datasets, often neglecting clinically relevant postural asymmetries-key indicators in traditional screening. In contrast, pose data provide an intuitive skeletal representation, enhancing clinical interpretability across various medical applications. However, pose-based scoliosis screening remains underexplored due to two main challenges: (1) the scarcity of large-scale, annotated pose datasets; and (2) the discrete and noise-sensitive nature of raw pose coordinates, which hinders the modeling of subtle asymmetries. To address these limitations, we introduce Scoliosis1K-Pose, a 2D human pose annotation set that extends the original Scoliosis1K dataset, comprising 447,900 frames of 2D keypoints from 1,050 adolescents. Building on this dataset, we introduce the Dual Representation Framework (DRF), which integrates a continuous skeleton map to preserve spatial structure with a discrete Postural Asymmetry Vector (PAV) that encodes clinically relevant asymmetry descriptors. A novel PAV-Guided Attention (PGA) module further uses the PAV as clinical prior to direct feature extraction from the skeleton map, focusing on clinically meaningful asymmetries. Extensive experiments demonstrate that DRF achieves state-of-the-art performance. Visualizations further confirm that the model leverages clinical asymmetry cues to guide feature extraction and promote synergy between its dual representations. The dataset and code are publicly available at https://zhouzi180.github.io/Scoliosis1K/.

[326] Spotlighter: Revisiting Prompt Tuning from a Representative Mining View

Yutong Gao, Maoyuan Shao, Xinyang Huang, Chuang Zhu, Lijuan Sun, Yu Weng, Xuan Liu, Guoshun Nan

Main category: cs.CV

TL;DR: Spotlighter is a lightweight token-selection framework that improves prompt tuning efficiency and accuracy by selecting only the most relevant visual tokens using semantic prototypes and a two-level ranking mechanism.

DetailsMotivation: CLIP's prompt tuning achieves cross-modal alignment but suffers from redundant features that introduce noise and computational costs. There's a need for more efficient and accurate token selection methods.

Method: Evaluates visual tokens from sample-wise and semantic-wise perspectives, retains top-scoring tokens, uses class-specific semantic memory bank with learned prototypes, and implements two-level ranking mechanism for dynamic token-prototype weighting.

Result: Outperforms CLIP by up to 11.19% in harmonic mean accuracy across 11 few-shot benchmarks, achieves up to 0.8K additional FPS, with only 21 extra parameters.

Conclusion: Spotlighter establishes an effective and scalable baseline for prompt tuning, demonstrating significant improvements in both accuracy and efficiency with minimal parameter overhead.

Abstract: CLIP’s success has demonstrated that prompt tuning can achieve robust cross-modal semantic alignment for tasks ranging from open-domain recognition to fine-grained classification. However, redundant or weakly relevant feature components introduce noise and incur unnecessary computational costs. In this work, we propose Spotlighter, a lightweight token-selection framework that simultaneously enhances accuracy and efficiency in prompt tuning. Spotlighter evaluates each visual token’s activation from both sample-wise and semantic-wise perspectives and retains only the top-scoring tokens for downstream prediction. A class-specific semantic memory bank of learned prototypes refines this selection, ensuring semantic representativeness and compensating for discarded features. To further prioritize informative signals, we introduce a two-level ranking mechanism that dynamically weights token–prototype interactions. Across 11 few-shot benchmarks, Spotlighter outperforms CLIP by up to 11.19% in harmonic mean accuracy and achieves up to 0.8K additional FPS, with only 21 extra parameters. These results establish Spotlighter as an effective and scalable baseline for prompt tuning. Code for our method will be available at https://github.com/greatest-gourmet/Spotlighter.

[327] DarkVRAI: Capture-Condition Conditioning and Burst-Order Selective Scan for Low-light RAW Video Denoising

Youngjin Oh, Junhyeong Kwon, Junyoung Park, Nam Ik Cho

Main category: cs.CV

TL;DR: DarkVRAI is a state-of-the-art low-light RAW video denoising framework that won the AIM 2025 challenge, featuring metadata conditioning and a novel temporal modeling mechanism.

DetailsMotivation: Low-light RAW video denoising is challenging due to severe signal degradation from high sensor gain and short exposure times required by video frame rates.

Method: Uses a conditioning scheme leveraging capture metadata to guide alignment/denoising, and a Burst-Order Selective Scan (BOSS) mechanism for long-range temporal dependency modeling.

Result: Achieved first place in AIM 2025 Low-light RAW Video Denoising Challenge and demonstrates state-of-the-art performance on rigorous benchmark datasets.

Conclusion: DarkVRAI sets a new standard for low-light video denoising through synergistic combination of metadata conditioning and advanced temporal modeling.

Abstract: Low-light RAW video denoising is a fundamentally challenging task due to severe signal degradation caused by high sensor gain and short exposure times, which are inherently limited by video frame rate requirements. To address this, we propose DarkVRAI, a novel framework that achieved first place in the AIM 2025 Low-light RAW Video Denoising Challenge. Our method introduces two primary contributions: (1) a successful application of a conditioning scheme for image denoising, which explicitly leverages capture metadata, to video denoising to guide the alignment and denoising processes, and (2) a Burst-Order Selective Scan (BOSS) mechanism that effectively models long-range temporal dependencies within the noisy video sequence. By synergistically combining these components, DarkVRAI demonstrates state-of-the-art performance on a rigorous and realistic benchmark dataset, setting a new standard for low-light video denoising.

[328] Seeing More, Saying More: Lightweight Language Experts are Dynamic Video Token Compressors

Xiangchen Wang, Jinrui Zhang, Teng Wang, Haigang Zhang, Feng Zheng

Main category: cs.CV

TL;DR: LangDC is a language-aware dynamic token compression method that adaptively adjusts compression ratios based on video semantic density, reducing FLOPs by 49% while maintaining performance.

DetailsMotivation: Fixed token compression ratios in video-language models inefficiently process videos with varying semantic density, causing inadequate representation of information-rich clips and wasted computation on static content.

Method: Uses lightweight language model to generate soft caption tokens as visual representations, trained with semantic density-aware supervision to dynamically adjust compression based on scene richness (description length).

Result: Achieves 49% reduction in FLOPs compared to VideoGPT+ while maintaining competitive performance, with adaptive compression based on video segment richness.

Conclusion: LangDC successfully mimics human visual processing by dynamically adjusting token compression based on scene complexity, providing efficient and effective video representation.

Abstract: Recent advancements in large video-language models have revolutionized video understanding tasks. However, their efficiency is significantly constrained by processing high volumes of visual tokens. Existing token compression strategies apply a fixed compression ratio, ignoring the variability in semantic density among different video clips. Consequently, this lead to inadequate representation of information-rich clips due to insufficient tokens and unnecessary computation on static or content-poor ones. To address this, we propose LangDC, a Language-aware Dynamic Token Compressor. LangDC leverages a lightweight language model to describe video clips, converting them into soft caption tokens as visual representations. Trained with our proposed semantic density-aware supervision, LangDC aims to 1) cover key visual cues necessary for downstream task reasoning and 2) dynamically adjust compression ratios based on scene richness, reflected by descriptions length. Our design mimics how humans dynamically express what they see: complex scenes (seeing more) elicit more detailed language to convey nuances (saying more), whereas simpler scenes are described with fewer words. Experimental results show that our method reduces FLOPs by 49% compared to VideoGPT+ while maintaining competitive performance. Furthermore, qualitative results demonstrate our approach adaptively adjusts the token compression ratio based on video segment richness.

[329] Towards Integrating Multi-Spectral Imaging with Gaussian Splatting

Josef Grün, Lukas Meyer, Maximilian Weiherer, Bernhard Egger, Marc Stamminger, Linus Franke

Main category: cs.CV

TL;DR: Integration of RGB and multi-spectral imagery into 3D Gaussian Splatting framework for improved 3D reconstruction, with joint optimization strategy showing best results.

DetailsMotivation: 3DGS works well on RGB data but naive per-band optimization of additional spectral bands yields poor reconstructions due to inconsistent geometry appearance across different spectra, despite the actual geometry being the same.

Method: Evaluated three strategies: 1) Separate per-band reconstruction, 2) Splitting optimization (first optimize RGB geometry then fit new bands), 3) Joint optimization (modalities optimized together, optionally with initial RGB-only phase). Integrated multi-spectral data into spherical harmonics color components.

Result: Joint optimization strategy significantly improved overall spectral reconstruction and enhanced RGB results through spectral cross-talk. Analysis revealed key trade-offs in when and how to introduce spectral bands during optimization.

Conclusion: Integrating multi-spectral data directly into spherical harmonics color components effectively models each Gaussian’s multi-spectral reflectance. Joint optimization approach provides robust multi-modal 3DGS reconstruction with practical insights for spectral band introduction timing.

Abstract: We present a study of how to integrate color (RGB) and multi-spectral imagery (red, green, red-edge, and near-infrared) into the 3D Gaussian Splatting (3DGS) framework, a state-of-the-art explicit radiance-field-based method for fast and high-fidelity 3D reconstruction from multi-view images. While 3DGS excels on RGB data, naive per-band optimization of additional spectra yields poor reconstructions due to inconsistently appearing geometry in the spectral domain. This problem is prominent, even though the actual geometry is the same, regardless of spectral modality. To investigate this, we evaluate three strategies: 1) Separate per-band reconstruction with no shared structure. 2) Splitting optimization, in which we first optimize RGB geometry, copy it, and then fit each new band to the model by optimizing both geometry and band representation. 3) Joint, in which the modalities are jointly optimized, optionally with an initial RGB-only phase. We showcase through quantitative metrics and qualitative novel-view renderings on multi-spectral datasets the effectiveness of our dedicated optimized Joint strategy, increasing overall spectral reconstruction as well as enhancing RGB results through spectral cross-talk. We therefore suggest integrating multi-spectral data directly into the spherical harmonics color components to compactly model each Gaussian’s multi-spectral reflectance. Moreover, our analysis reveals several key trade-offs in when and how to introduce spectral bands during optimization, offering practical insights for robust multi-modal 3DGS reconstruction.

[330] Weather-Dependent Variations in Driver Gaze Behavior: A Case Study in Rainy Conditions

Ghazal Farhani, Taufiq Rahman, Dominique Charlebois

Main category: cs.CV

TL;DR: Rainy conditions cause drivers to exhibit different gaze patterns including more dashboard glances, longer fixations, and higher gaze elevation, indicating increased cognitive focus compared to clear weather driving.

DetailsMotivation: Rainy weather increases road accident risks due to reduced visibility and traction. Understanding how experienced drivers adapt their visual perception through gaze behavior is critical for designing better driver monitoring systems (DMS) and advanced driver assistance systems (ADAS).

Method: Case study analyzing eye gaze behavior on the same highway route under clear vs rainy conditions using a two-step clustering approach: clustering gaze points within 10-second intervals, then aggregating cluster centroids into meta-clusters. Also used Markov transition matrices, fixation duration, gaze elevation, and azimuth distribution metrics.

Result: While overall gaze behavior (road focus with occasional mirror checks) remains consistent, rainy conditions lead to: more frequent dashboard glances, longer fixation durations, and higher gaze elevation. These changes indicate increased cognitive focus during adverse weather.

Conclusion: The findings provide valuable insights into visual attention patterns under adverse conditions and demonstrate the potential of leveraging gaze modeling to design more robust ADAS and DMS systems that account for weather-related behavioral adaptations.

Abstract: Rainy weather significantly increases the risk of road accidents due to reduced visibility and vehicle traction. Understanding how experienced drivers adapt their visual perception through gaze behavior under such conditions is critical for designing robust driver monitoring systems (DMS) and for informing advanced driver assistance systems (ADAS). This case study investigates the eye gaze behavior of a driver operating the same highway route under both clear and rainy conditions. To this end, gaze behavior was analyzed by a two-step clustering approach: first, clustering gaze points within 10-second intervals, and then aggregating cluster centroids into meta-clusters. This, along with Markov transition matrices and metrics such as fixation duration, gaze elevation, and azimuth distributions, reveals meaningful behavioral shifts. While the overall gaze behavior focused on the road with occasional mirror checks remains consistent, rainy conditions lead to more frequent dashboard glances, longer fixation durations, and higher gaze elevation, indicating increased cognitive focus. These findings offer valuable insight into visual attention patterns under adverse conditions and highlight the potential of leveraging gaze modeling to aid in the design of more robust ADAS and DMS.

[331] AI-driven Dispensing of Coral Reseeding Devices for Broad-scale Restoration of the Great Barrier Reef

Scarlett Raine, Benjamin Moshirian, Tobias Fischer

Main category: cs.CV

TL;DR: AI-powered automated coral reseeding system using computer vision and robotics for substrate classification, achieving 77.8% deployment accuracy and 89.1% classification accuracy on Great Barrier Reef.

DetailsMotivation: Coral reefs face 70-90% species loss within a decade due to climate change, acidification, and pollution. Restoration requires automation to scale up efforts and reduce human expert reliance.

Method: Developed automated coral reseeding devices using AI, computer vision, and robotics for substrate classification to detect suitable seafloor areas for coral growth.

Result: Achieved 77.8% deployment accuracy, 89.1% sub-image patch classification accuracy, and real-time inference at 5.5 frames per second on Great Barrier Reef. Contributed large annotated substrate image dataset publicly.

Conclusion: Automated AI-powered coral restoration system successfully demonstrates scalable reef rehabilitation with high accuracy, reducing human dependency and enabling efficient large-scale deployment.

Abstract: Coral reefs are on the brink of collapse, with climate change, ocean acidification, and pollution leading to a projected 70-90% loss of coral species within the next decade. Restoration efforts are crucial, but their success hinges on introducing automation to upscale efforts. We present automated deployment of coral re-seeding devices powered by artificial intelligence, computer vision, and robotics. Specifically, we perform automated substrate classification, enabling detection of areas of the seafloor suitable for coral growth, thus significantly reducing reliance on human experts and increasing the range and efficiency of restoration. Real-world testing of the algorithms on the Great Barrier Reef leads to deployment accuracy of 77.8%, sub-image patch classification of 89.1%, and real-time model inference at 5.5 frames per second. Further, we present and publicly contribute a large collection of annotated substrate image data to foster future research in this area.

[332] CompSlider: Compositional Slider for Disentangled Multiple-Attribute Image Generation

Zixin Zhu, Kevin Duarte, Mamshad Nayeem Rizve, Chengyuan Xu, Ratheesh Kalarot, Junsong Yuan

Main category: cs.CV

TL;DR: CompSlider enables simultaneous control of multiple image attributes in text-to-image generation using disentangled sliders, avoiding attribute interference while maintaining structural consistency.

DetailsMotivation: Existing slider-based methods train individual adapters per attribute, causing interference between attributes and preventing precise multi-attribute control due to attribute entanglement.

Method: CompSlider generates a conditional prior for T2I foundation models, using novel disentanglement and structure losses to compose multiple attribute changes while maintaining image structural consistency. It operates in latent space without retraining the foundation model.

Result: The approach enables reliable and independent manipulation of multiple attributes simultaneously, reduces computational burden for training and inference, and demonstrates generality by extending to video generation.

Conclusion: CompSlider successfully addresses attribute entanglement in slider-based generation, providing precise multi-attribute control while maintaining structural consistency and computational efficiency.

Abstract: In text-to-image (T2I) generation, achieving fine-grained control over attributes - such as age or smile - remains challenging, even with detailed text prompts. Slider-based methods offer a solution for precise control of image attributes. Existing approaches typically train individual adapter for each attribute separately, overlooking the entanglement among multiple attributes. As a result, interference occurs among different attributes, preventing precise control of multiple attributes together. To address this challenge, we aim to disentangle multiple attributes in slider-based generation to enbale more reliable and independent attribute manipulation. Our approach, CompSlider, can generate a conditional prior for the T2I foundation model to control multiple attributes simultaneously. Furthermore, we introduce novel disentanglement and structure losses to compose multiple attribute changes while maintaining structural consistency within the image. Since CompSlider operates in the latent space of the conditional prior and does not require retraining the foundation model, it reduces the computational burden for both training and inference. We evaluate our approach on a variety of image attributes and highlight its generality by extending to video generation.

[333] Seeing through Unclear Glass: Occlusion Removal with One Shot

Qiang Li, Yuanming Cao

Main category: cs.CV

TL;DR: Proposes an all-in-one model for restoring images taken through contaminated glass using real paired data and test-time adaptation for various contaminants like mud, dirt, and particles.

DetailsMotivation: Existing methods rely on synthetic data or only handle specific contaminants like raindrops, but real-world glass contamination involves diverse occluders that degrade image quality through light attenuation and scattering.

Method: Uses real paired images of clean/contaminated glass and proposes an all-in-one model with self-supervised test-time adaptation to handle different contamination types by updating for each test image’s unique occlusion.

Result: Outperforms state-of-the-art methods quantitatively and qualitatively, especially on unseen contamination types, demonstrating effective restoration of realistic contaminated images.

Conclusion: The proposed approach successfully handles diverse real-world glass contaminants through real data collection and adaptive test-time learning, showing superior performance over existing methods.

Abstract: Images taken through window glass are often degraded by contaminants adhered to the glass surfaces. Such contaminants cause occlusions that attenuate the incoming light and scatter stray light towards the camera. Most of existing deep learning methods for neutralizing the effects of contaminated glasses relied on synthetic training data. Few researchers used real degraded and clean image pairs, but they only considered removing or alleviating the effects of rain drops on glasses. This paper is concerned with the more challenging task of learning the restoration of images taken through glasses contaminated by a wide range of occluders, including muddy water, dirt and other small foreign particles found in reality. To facilitate the learning task we have gone to a great length to acquire real paired images with and without glass contaminants. More importantly, we propose an all-in-one model to neutralize contaminants of different types by utilizing the one-shot test-time adaptation mechanism. It involves a self-supervised auxiliary learning task to update the trained model for the unique occlusion type of each test image. Experimental results show that the proposed method outperforms the state-of-the-art methods quantitatively and qualitatively in cleaning realistic contaminated images, especially the unseen ones.

[334] A Unified Low-level Foundation Model for Enhancing Pathology Image Quality

Ziyi Liu, Zhe Xu, Jiabo Ma, Wenqaing Li, Junlin Hou, Fuxiang Huang, Xi Wang, Ronald Cheong Kin Chan, Terence Tsz Wai Wong, Hao Chen

Main category: cs.CV

TL;DR: Proposes LPFM, the first unified low-level pathology foundation model that handles image restoration (super-resolution, deblurring, denoising) and virtual staining tasks through a single architecture using contrastive pre-training and conditional diffusion.

DetailsMotivation: Real-world pathology images suffer from degradations like noise, blur, low resolution, and staining inconsistencies, while existing methods are task-specific and lack versatility for diverse low-level vision challenges.

Method: Uses contrastive pre-trained encoder on 190M unlabeled images to learn stain-invariant features, combined with unified conditional diffusion process that adapts to specific tasks via textual prompts.

Result: Achieves statistically significant improvements (p<0.01) over SOTA methods in 56/66 tasks, with 10-15% PSNR gains for restoration and 12-18% SSIM improvements for virtual staining.

Conclusion: LPFM successfully bridges the gap in low-level pathology image enhancement by providing a unified foundation model capable of handling multiple restoration and translation tasks with superior performance.

Abstract: Foundation models have revolutionized computational pathology by achieving remarkable success in high-level diagnostic tasks, yet the critical challenge of low-level image enhancement remains largely unaddressed. Real-world pathology images frequently suffer from degradations such as noise, blur, and low resolution due to slide preparation artifacts, staining variability, and imaging constraints, while the reliance on physical staining introduces significant costs, delays, and inconsistency. Although existing methods target individual problems like denoising or super-resolution, their task-specific designs lack the versatility to handle the diverse low-level vision challenges encountered in practice. To bridge this gap, we propose the first unified Low-level Pathology Foundation Model (LPFM), capable of enhancing image quality in restoration tasks, including super-resolution, deblurring, and denoising, as well as facilitating image translation tasks like virtual staining (H&E and special stains), all through a single adaptable architecture. Our approach introduces a contrastive pre-trained encoder that learns transferable, stain-invariant feature representations from 190 million unlabeled pathology images, enabling robust identification of degradation patterns. A unified conditional diffusion process dynamically adapts to specific tasks via textual prompts, ensuring precise control over output quality. Trained on a curated dataset of 87,810 whole slied images (WSIs) across 34 tissue types and 5 staining protocols, LPFM demonstrates statistically significant improvements (p<0.01) over state-of-the-art methods in most tasks (56/66), achieving Peak Signal-to-Noise Ratio (PSNR) gains of 10-15% for image restoration and Structural Similarity Index Measure (SSIM) improvements of 12-18% for virtual staining.

[335] SpectMamba: Integrating Frequency and State Space Models for Enhanced Medical Image Detection

Yao Wang, Dong Yang, Zhi Qiao, Wenjian Huang, Liuzhi Yang, Zhen Qian

Main category: cs.CV

TL;DR: SpectMamba is a novel Mamba-based architecture for medical image detection that uses Hybrid Spatial-Frequency Attention and Visual State-Space Module to capture global context efficiently with linear complexity.

DetailsMotivation: Existing CNN models have limited receptive fields while Transformers face prohibitive computational costs for high-resolution medical images. Mamba's linear complexity for long sequences offers a promising alternative for medical image analysis.

Method: Proposes SpectMamba with Hybrid Spatial-Frequency Attention block to separately learn high/low-frequency features, Visual State-Space Module for long-range dependencies, and Hilbert Curve Scanning to strengthen spatial correlations.

Result: Achieves state-of-the-art performance across various medical image detection tasks while maintaining both effectiveness and efficiency.

Conclusion: SpectMamba successfully addresses the limitations of CNNs and Transformers in medical imaging, providing an efficient and accurate solution for abnormality detection with global context capture capabilities.

Abstract: Abnormality detection in medical imaging is a critical task requiring both high efficiency and accuracy to support effective diagnosis. While convolutional neural networks (CNNs) and Transformer-based models are widely used, both face intrinsic challenges: CNNs have limited receptive fields, restricting their ability to capture broad contextual information, and Transformers encounter prohibitive computational costs when processing high-resolution medical images. Mamba, a recent innovation in natural language processing, has gained attention for its ability to process long sequences with linear complexity, offering a promising alternative. Building on this foundation, we present SpectMamba, the first Mamba-based architecture designed for medical image detection. A key component of SpectMamba is the Hybrid Spatial-Frequency Attention (HSFA) block, which separately learns high- and low-frequency features. This approach effectively mitigates the loss of high-frequency information caused by frequency bias and correlates frequency-domain features with spatial features, thereby enhancing the model’s ability to capture global context. To further improve long-range dependencies, we propose the Visual State-Space Module (VSSM) and introduce a novel Hilbert Curve Scanning technique to strengthen spatial correlations and local dependencies, further optimizing the Mamba framework. Comprehensive experiments show that SpectMamba achieves state-of-the-art performance while being both effective and efficient across various medical image detection tasks.

[336] Bidirectional Sparse Attention for Faster Video Diffusion Training

Chenlu Zhan, Wen Li, Chuyu Shen, Jun Zhang, Suhui Wu, Hao Zhang

Main category: cs.CV

TL;DR: BSA framework uses bidirectional sparse attention to dramatically accelerate video DiT training by 17.79x while maintaining generative quality, addressing quadratic complexity bottlenecks in full attention.

DetailsMotivation: Video diffusion Transformer models face prohibitive computational costs due to quadratic complexity of full attention when generating high-resolution, long-duration videos, with inefficiencies from sparse queries and redundant computation patterns.

Method: Bidirectional Sparse Attention (BSA) framework that dynamically sparsifies both queries and key-value pairs in 3D full attention. Query sparsity via semantic similarity selection and dynamic spatial-time training, KV sparsity through statistical dynamic thresholding to retain only salient blocks.

Result: BSA reduces FLOPs by up to 20x, achieves 17.79x faster attention training, and preserves or even surpasses the generative quality of full attention across long sequences.

Conclusion: BSA effectively overcomes computational bottlenecks in video DiTs by dynamically optimizing attention sparsity, enabling efficient high-quality video generation without sacrificing performance.

Abstract: Video diffusion Transformer (DiT) models excel in generative quality but hit major computational bottlenecks when producing high-resolution, long-duration videos. The quadratic complexity of full attention leads to prohibitively high training and inference costs. Full attention inefficiency stems from two key challenges: excessive computation due to the inherent sparsity of Queries and Key-Value pairs, and redundant computation as fixed sparse patterns fail to leverage DiT’s dynamic attention. To overcome this limitation, we propose a Bidirectional Sparse Attention (BSA) framework for faster video DiT training, the first to dynamically sparsify both Queries and Key-Value pairs within 3D full attention, thereby substantially improving training and inference efficiency. BSA addresses these issues through two key components. Query sparsity is optimized by selecting the most informative query tokens via semantic similarity and with a dynamic spatial-time training strategy, while KV sparsity is achieved by computing a statistical dynamic threshold to retain only the most salient KV blocks for computation. Extensive experiments demonstrate that BSA significantly accelerates DiT training across long sequences, reducing FLOPs by up to 20x and achieving 17.79x faster attention training, while preserving or even surpassing the generative quality of full attention.

[337] An End-to-End Framework for Video Multi-Person Pose Estimation

Zhihong Wei

Main category: cs.CV

TL;DR: VEPE is an end-to-end video pose estimation framework that uses spatio-temporal Transformers to address limitations of two-stage approaches, improving both accuracy and inference efficiency by 300%.

DetailsMotivation: Existing video pose estimation methods separate spatial and temporal processing, rely on complex post-processing, and cannot capture global spatio-temporal context for end-to-end optimization.

Method: Proposes VEPE framework with three spatio-temporal Transformer components: STPE, STDME, and STPD, plus an instance consistency mechanism for cross-frame pose query matching and instance tracking.

Result: Outperforms most two-stage models on Posetrack dataset and achieves 300% improvement in inference efficiency.

Conclusion: VEPE provides a simple and flexible end-to-end solution for video pose estimation that effectively utilizes temporal context while significantly improving efficiency.

Abstract: Video-based human pose estimation models aim to address scenarios that cannot be effectively solved by static image models such as motion blur, out-of-focus and occlusion. Most existing approaches consist of two stages: detecting human instances in each image frame and then using a temporal model for single-person pose estimation. This approach separates the spatial and temporal dimensions and cannot capture the global spatio-temporal context between spatial instances for end-to-end optimization. In addition, it relies on separate detectors and complex post-processing such as RoI cropping and NMS, which reduces the inference efficiency of the video scene. To address the above problems, we propose VEPE (Video End-to-End Pose Estimation), a simple and flexible framework for end-to-end pose estimation in video. The framework utilizes three crucial spatio-temporal Transformer components: the Spatio-Temporal Pose Encoder (STPE), the Spatio-Temporal Deformable Memory Encoder (STDME), and the Spatio-Temporal Pose Decoder (STPD). These components are designed to effectively utilize temporal context for optimizing human body pose estimation. Furthermore, to reduce the mismatch problem during the cross-frame pose query matching process, we propose an instance consistency mechanism, which aims to enhance the consistency and discrepancy of the cross-frame instance query and realize the instance tracking function, which in turn accurately guides the pose query to perform cross-frame matching. Extensive experiments on the Posetrack dataset show that our approach outperforms most two-stage models and improves inference efficiency by 300%.

[338] PVINet: Point-Voxel Interlaced Network for Point Cloud Compression

Xuan Deng, Xingtao Wang, Xiandong Meng, Xiaopeng Fan, Debin Zhao

Main category: cs.CV

TL;DR: PVINet is a point cloud compression method that processes global and local features in parallel with interactive communication, using conditional sparse convolution to enhance feature extraction and reconstruction.

DetailsMotivation: Existing point cloud compression methods process global and local information sequentially without communication between them, limiting reconstruction quality that depends on both structural and contextual features.

Method: Proposes PVINet with voxel-based encoder for global features and point-based encoder for local contexts, using novel conditional sparse convolution to dynamically customize kernels and enable feature interactions between encoders.

Result: Experiments on benchmark datasets show competitive performance compared to state-of-the-art methods.

Conclusion: PVINet effectively captures both global structural and local contextual features through parallel processing and interactive communication, achieving improved point cloud compression performance.

Abstract: In point cloud compression, the quality of a reconstructed point cloud relies on both the global structure and the local context, with existing methods usually processing global and local information sequentially and lacking communication between these two types of information. In this paper, we propose a point-voxel interlaced network (PVINet), which captures global structural features and local contextual features in parallel and performs interactions at each scale to enhance feature perception efficiency. Specifically, PVINet contains a voxel-based encoder (Ev) for extracting global structural features and a point-based encoder (Ep) that models local contexts centered at each voxel. Particularly, a novel conditional sparse convolution is introduced, which applies point embeddings to dynamically customize kernels for voxel feature extraction, facilitating feature interactions from Ep to Ev. During decoding, a voxel-based decoder employs conditional sparse convolutions to incorporate point embeddings as guidance to reconstruct the point cloud. Experiments on benchmark datasets show that PVINet delivers competitive performance compared to state-of-the-art methods.

[339] FICGen: Frequency-Inspired Contextual Disentanglement for Layout-driven Degraded Image Generation

Wenzhuang Wang, Yifan Zhao, Mingcan Ma, Ming Liu, Zhonglin Jiang, Yong Chen, Jia Li

Main category: cs.CV

TL;DR: FICGen is a new frequency-inspired generative paradigm that addresses layout-to-image generation challenges in degraded scenes by disentangling contextual frequency distributions and improving instance rendering through frequency-aware guidance.

DetailsMotivation: Layout-to-image generation struggles with limited fidelity and weak alignment in degraded conditions (low-light, underwater) due to the "contextual illusion dilemma" where foreground instances are overwhelmed by context-dominant frequency distributions.

Method: Proposes FICGen with: 1) learnable dual-query mechanism with frequency resamplers to extract contextual frequency prototypes, 2) visual-frequency enhanced attention to inject frequency knowledge, 3) instance coherence map for latent-space disentanglement, and 4) adaptive spatial-frequency aggregation module for mixed degraded representations.

Result: Extensive experiments on 5 benchmarks across various degraded scenarios (severe low-light to mild blur) show FICGen consistently outperforms existing L2I methods in generative fidelity, alignment, and downstream auxiliary trainability.

Conclusion: FICGen effectively addresses the contextual illusion problem in degraded image generation by transferring frequency knowledge to latent diffusion space, enabling better rendering of degraded instances and their surroundings through frequency-aware guidance.

Abstract: Layout-to-image (L2I) generation has exhibited promising results in natural domains, but suffers from limited generative fidelity and weak alignment with user-provided layouts when applied to degraded scenes (i.e., low-light, underwater). We primarily attribute these limitations to the “contextual illusion dilemma” in degraded conditions, where foreground instances are overwhelmed by context-dominant frequency distributions. Motivated by this, our paper proposes a new Frequency-Inspired Contextual Disentanglement Generative (FICGen) paradigm, which seeks to transfer frequency knowledge of degraded images into the latent diffusion space, thereby facilitating the rendering of degraded instances and their surroundings via contextual frequency-aware guidance. To be specific, FICGen consists of two major steps. Firstly, we introduce a learnable dual-query mechanism, each paired with a dedicated frequency resampler, to extract contextual frequency prototypes from pre-collected degraded exemplars in the training set. Secondly, a visual-frequency enhanced attention is employed to inject frequency prototypes into the degraded generation process. To alleviate the contextual illusion and attribute leakage, an instance coherence map is developed to regulate latent-space disentanglement between individual instances and their surroundings, coupled with an adaptive spatial-frequency aggregation module to reconstruct spatial-frequency mixed degraded representations. Extensive experiments on 5 benchmarks involving a variety of degraded scenarios-from severe low-light to mild blur-demonstrate that FICGen consistently surpasses existing L2I methods in terms of generative fidelity, alignment and downstream auxiliary trainability.

[340] GPSToken: Gaussian Parameterized Spatially-adaptive Tokenization for Image Representation and Generation

Zhengqiang Zhang, Rongyuan Wu, Lingchen Sun, Lei Zhang

Main category: cs.CV

TL;DR: GPSToken is a novel Gaussian parameterized spatially-adaptive tokenization framework that uses 2D Gaussians to dynamically model image regions with varying shapes, positions, and textures, achieving state-of-the-art performance in image reconstruction and generation.

DetailsMotivation: Conventional uniform grid tokenization methods are inflexible for representing regions with varying shapes and textures at different locations, limiting feature representation efficacy.

Method: Uses entropy-driven algorithm to partition images into texture-homogeneous regions, parameterizes each region as 2D Gaussian (mean for position, covariance for shape) with texture features, and employs specialized transformer to optimize Gaussian parameters with differentiable splatting-based renderer.

Result: Achieves rFID and FID scores of 0.65 and 1.50 on image reconstruction and generation tasks using only 128 tokens, demonstrating state-of-the-art performance.

Conclusion: GPSToken provides effective non-uniform image tokenization that disentangles spatial layout from texture features, enabling efficient two-stage generation and superior performance compared to conventional methods.

Abstract: Effective and efficient tokenization plays an important role in image representation and generation. Conventional methods, constrained by uniform 2D/1D grid tokenization, are inflexible to represent regions with varying shapes and textures and at different locations, limiting their efficacy of feature representation. In this work, we propose $\textbf{GPSToken}$, a novel $\textbf{G}$aussian $\textbf{P}$arameterized $\textbf{S}$patially-adaptive $\textbf{Token}$ization framework, to achieve non-uniform image tokenization by leveraging parametric 2D Gaussians to dynamically model the shape, position, and textures of different image regions. We first employ an entropy-driven algorithm to partition the image into texture-homogeneous regions of variable sizes. Then, we parameterize each region as a 2D Gaussian (mean for position, covariance for shape) coupled with texture features. A specialized transformer is trained to optimize the Gaussian parameters, enabling continuous adaptation of position/shape and content-aware feature extraction. During decoding, Gaussian parameterized tokens are reconstructed into 2D feature maps through a differentiable splatting-based renderer, bridging our adaptive tokenization with standard decoders for end-to-end training. GPSToken disentangles spatial layout (Gaussian parameters) from texture features to enable efficient two-stage generation: structural layout synthesis using lightweight networks, followed by structure-conditioned texture generation. Experiments demonstrate the state-of-the-art performance of GPSToken, which achieves rFID and FID scores of 0.65 and 1.50 on image reconstruction and generation tasks using 128 tokens, respectively. Codes and models of GPSToken can be found at $\href{https://github.com/xtudbxk/GPSToken}{https://github.com/xtudbxk/GPSToken}$.

[341] MetaSSL: A General Heterogeneous Loss for Semi-Supervised Medical Image Segmentation

Weiren Zhao, Lanfeng Zhong, Xin Liao, Wenjun Liao, Sichuan Zhang, Shaoting Zhang, Guotai Wang

Main category: cs.CV

TL;DR: MetaSSL is a universal semi-supervised learning framework that improves medical image segmentation by assigning spatially heterogeneous weights to pixels based on uncertainty and consistency between predictions, addressing both unlabeled data heterogeneity and labeled data noise.

DetailsMotivation: Existing SSL methods overlook noise in labeled data and ignore the heterogeneous values of different unlabeled pixels. The authors argue that effectively mining information from both reference and supervised predictions is more essential than specific reference generation strategies.

Method: Proposes a spatially heterogeneous loss that splits predictions into four regions (UC, US, DC, DS) with decreasing weights based on uncertainty and consistency. Uses adaptive thresholding to distinguish confident from suspicious predictions and applies the approach to both labeled and unlabeled data.

Result: Experimental results showed significant improvement in segmentation performance when integrated with existing SSL frameworks across different datasets.

Conclusion: MetaSSL provides a plug-and-play universal framework that enhances SSL methods by better leveraging prediction information through spatially adaptive weighting, making it robust to annotation noise and effective for medical image segmentation.

Abstract: Semi-Supervised Learning (SSL) is important for reducing the annotation cost for medical image segmentation models. State-of-the-art SSL methods such as Mean Teacher, FixMatch and Cross Pseudo Supervision (CPS) are mainly based on consistency regularization or pseudo-label supervision between a reference prediction and a supervised prediction. Despite the effectiveness, they have overlooked the potential noise in the labeled data, and mainly focus on strategies to generate the reference prediction, while ignoring the heterogeneous values of different unlabeled pixels. We argue that effectively mining the rich information contained by the two predictions in the loss function, instead of the specific strategy to obtain a reference prediction, is more essential for SSL, and propose a universal framework MetaSSL based on a spatially heterogeneous loss that assigns different weights to pixels by simultaneously leveraging the uncertainty and consistency information between the reference and supervised predictions. Specifically, we split the predictions on unlabeled data into four regions with decreasing weights in the loss: Unanimous and Confident (UC), Unanimous and Suspicious (US), Discrepant and Confident (DC), and Discrepant and Suspicious (DS), where an adaptive threshold is proposed to distinguish confident predictions from suspicious ones. The heterogeneous loss is also applied to labeled images for robust learning considering the potential annotation noise. Our method is plug-and-play and general to most existing SSL methods. The experimental results showed that it improved the segmentation performance significantly when integrated with existing SSL frameworks on different datasets. Code is available at https://github.com/HiLab-git/MetaSSL.

[342] MVTrajecter: Multi-View Pedestrian Tracking with Trajectory Motion Cost and Trajectory Appearance Cost

Taiga Yamane, Ryo Masumura, Satoshi Suzuki, Shota Orihashi

Main category: cs.CV

TL;DR: MVTrajecter is a novel end-to-end multi-view pedestrian tracking method that utilizes multiple past timestamps from trajectories for robust association, outperforming previous state-of-the-art methods.

DetailsMotivation: Previous end-to-end MVPT methods only use current and single adjacent past timestamp, discarding valuable information from complete past trajectories, which limits association robustness.

Method: Proposes MVTrajecter with trajectory motion cost and trajectory appearance cost to incorporate motion and appearance information from multiple past timestamps, using attention mechanism to capture relationships between timestamps.

Result: Extensive experiments demonstrate the effectiveness of each component and show that MVTrajecter outperforms previous state-of-the-art methods in multi-view pedestrian tracking.

Conclusion: Utilizing information from multiple past timestamps through trajectory-based costs and attention mechanisms significantly improves pedestrian association robustness in multi-view tracking systems.

Abstract: Multi-View Pedestrian Tracking (MVPT) aims to track pedestrians in the form of a bird’s eye view occupancy map from multi-view videos. End-to-end methods that detect and associate pedestrians within one model have shown great progress in MVPT. The motion and appearance information of pedestrians is important for the association, but previous end-to-end MVPT methods rely only on the current and its single adjacent past timestamp, discarding the past trajectories before that. This paper proposes a novel end-to-end MVPT method called Multi-View Trajectory Tracker (MVTrajecter) that utilizes information from multiple timestamps in past trajectories for robust association. MVTrajecter introduces trajectory motion cost and trajectory appearance cost to effectively incorporate motion and appearance information, respectively. These costs calculate which pedestrians at the current and each past timestamp are likely identical based on the information between those timestamps. Even if a current pedestrian could be associated with a false pedestrian at some past timestamp, these costs enable the model to associate that current pedestrian with the correct past trajectory based on other past timestamps. In addition, MVTrajecter effectively captures the relationships between multiple timestamps leveraging the attention mechanism. Extensive experiments demonstrate the effectiveness of each component in MVTrajecter and show that it outperforms the previous state-of-the-art methods.

[343] Do Video Language Models Really Know Where to Look? Diagnosing Attention Failures in Video Language Models

Hyunjong Ok, Jaeho Lee

Main category: cs.CV

TL;DR: Vision encoders used in multimodal LLMs for video understanding fail to identify the most informative frames, limiting efficient video processing.

DetailsMotivation: To investigate whether current vision-language encoders can truly identify the most informative frames for video understanding tasks in multimodal LLMs, given the computational cost of processing all frames.

Method: Provided empirical evidence through analysis of popular vision encoders’ capability to identify keyframes that MLLMs should focus on for given textual queries.

Result: Found that popular vision encoders critically suffer from limited capability to identify where MLLMs should look inside videos to appropriately handle textual queries.

Conclusion: Development of better keyframe identification techniques is necessary for efficient video multimodal large language models.

Abstract: Recent advances in multimodal large language models (MLLMs) have led to much progress in video understanding tasks. To avoid the heavy computational cost of processing all frames, these models typically rely on keyframe sampling methods guided by vision-language encoders (\textit{e.g.,} SigLIP). However, it remains unclear whether such encoders can truly identify the most informative frames. In this work, we provide several empirical pieces of evidence revealing that popular vision encoders critically suffer from their limited capability to identify where the MLLM should look inside the video to handle the given textual query appropriately. Our findings suggest that the development of better keyframe identification techniques may be necessary for efficient video MLLMs.

[344] DynaMind: Reconstructing Dynamic Visual Scenes from EEG by Aligning Temporal Dynamics and Multimodal Semantics to Guided Diffusion

Junxiang Liu, Junming Lin, Jiangtong Li, Jie Li

Main category: cs.CV

TL;DR: DynaMind is a novel EEG-to-video reconstruction framework that jointly models neural dynamics and semantic features using three modules: Regional-aware Semantic Mapper, Temporal-aware Dynamic Aligner, and Dual-Guidance Video Reconstructor, achieving state-of-the-art performance with significant improvements in accuracy and visual fidelity.

DetailsMotivation: Current EEG-based video reconstruction methods struggle with low spatial resolution of EEG, temporal mismatches between neural recordings and video dynamics, and insufficient use of semantic information, leading to inadequate dynamic coherence and semantic context resolution.

Method: Three core modules: 1) Regional-aware Semantic Mapper extracts multimodal semantic features from EEG across brain regions into a unified diffusion prior; 2) Temporal-aware Dynamic Aligner generates dynamic latent sequences for temporal consistency; 3) Dual-Guidance Video Reconstructor translates temporal blueprint into high-fidelity video using semantic guidance.

Result: State-of-the-art performance on SEED-DV dataset: 12.5% and 10.3% improvement in video- and frame-based accuracies respectively, 9.4% SSIM improvement, and 19.7% FVMD reduction, demonstrating exceptional visual fidelity and temporal coherence.

Conclusion: DynaMind represents a critical advancement in bridging neural dynamics with high-fidelity visual semantics, effectively overcoming limitations of previous EEG-based video reconstruction methods through joint modeling of neural dynamics and semantic features.

Abstract: Reconstruction dynamic visual scenes from electroencephalography (EEG) signals remains a primary challenge in brain decoding, limited by the low spatial resolution of EEG, a temporal mismatch between neural recordings and video dynamics, and the insufficient use of semantic information within brain activity. Therefore, existing methods often inadequately resolve both the dynamic coherence and the complex semantic context of the perceived visual stimuli. To overcome these limitations, we introduce DynaMind, a novel framework that reconstructs video by jointly modeling neural dynamics and semantic features via three core modules: a Regional-aware Semantic Mapper (RSM), a Temporal-aware Dynamic Aligner (TDA), and a Dual-Guidance Video Reconstructor (DGVR). The RSM first utilizes a regional-aware encoder to extract multimodal semantic features from EEG signals across distinct brain regions, aggregating them into a unified diffusion prior. In the mean time, the TDA generates a dynamic latent sequence, or blueprint, to enforce temporal consistency between the feature representations and the original neural recordings. Together, guided by the semantic diffusion prior, the DGVR translates the temporal-aware blueprint into a high-fidelity video reconstruction. On the SEED-DV dataset, DynaMind sets a new state-of-the-art (SOTA), boosting reconstructed video accuracies (video- and frame-based) by 12.5 and 10.3 percentage points, respectively. It also achieves a leap in pixel-level quality, showing exceptional visual fidelity and temporal coherence with a 9.4% SSIM improvement and a 19.7% FVMD reduction. This marks a critical advancement, bridging the gap between neural dynamics and high-fidelity visual semantics.

[345] FocusDPO: Dynamic Preference Optimization for Multi-Subject Personalized Image Generation via Adaptive Focus

Qiaoqiao Jin, Siming Fu, Dong She, Weinan Jia, Hualiang Wang, Mu Liu, Jidong Jiang

Main category: cs.CV

TL;DR: FocusDPO is a framework for multi-subject personalized image generation that uses adaptive focus regions based on semantic correspondence and image complexity to prevent attribute leakage while maintaining subject fidelity.

DetailsMotivation: Multi-subject personalized image generation faces challenges in preserving subject fidelity and preventing cross-subject attribute leakage without test-time optimization.

Method: Adaptively identifies focus regions based on dynamic semantic correspondence and supervision image complexity, progressively adjusting focal areas across noise timesteps with a weighted strategy that rewards information-rich patches and penalizes low-confidence regions.

Result: Substantially enhances performance of existing pre-trained personalized generation models, achieving state-of-the-art results on both single-subject and multi-subject benchmarks while effectively mitigating attribute leakage.

Conclusion: The framework advances controllable multi-subject image synthesis by dynamically adjusting focus allocation and establishing robust correspondence mappings between generated and reference subjects.

Abstract: Multi-subject personalized image generation aims to synthesize customized images containing multiple specified subjects without requiring test-time optimization. However, achieving fine-grained independent control over multiple subjects remains challenging due to difficulties in preserving subject fidelity and preventing cross-subject attribute leakage. We present FocusDPO, a framework that adaptively identifies focus regions based on dynamic semantic correspondence and supervision image complexity. During training, our method progressively adjusts these focal areas across noise timesteps, implementing a weighted strategy that rewards information-rich patches while penalizing regions with low prediction confidence. The framework dynamically adjusts focus allocation during the DPO process according to the semantic complexity of reference images and establishes robust correspondence mappings between generated and reference subjects. Extensive experiments demonstrate that our method substantially enhances the performance of existing pre-trained personalized generation models, achieving state-of-the-art results on both single-subject and multi-subject personalized image synthesis benchmarks. Our method effectively mitigates attribute leakage while preserving superior subject fidelity across diverse generation scenarios, advancing the frontier of controllable multi-subject image synthesis.

[346] SegAssess: Panoramic quality mapping for robust and transferable unsupervised segmentation assessment

Bingnan Yang, Mi Zhang, Zhili Zhang, Zhan Zhang, Yuanxin Zhao, Xiangyun Hu, Jianya Gong

Main category: cs.CV

TL;DR: SegAssess is a novel DL framework that introduces Panoramic Quality Mapping for comprehensive pixel-level segmentation quality assessment, achieving SOTA performance with excellent zero-shot transferability across 32 datasets.

DetailsMotivation: Existing unsupervised SQA methods suffer from coarse granularity, incomplete assessments, and poor transferability, creating a need for more robust pixel-level quality evaluation in remote sensing without ground truth.

Method: Formulates SQA as a four-class panoramic segmentation task (TP/FP/TN/FN) using enhanced SAM architecture with mask prompting, Edge Guided Compaction branch with ASF module, and Augmented Mixup Sampling training strategy.

Result: Achieves state-of-the-art performance across 32 datasets from 6 sources and demonstrates remarkable zero-shot transferability to unseen masks.

Conclusion: SegAssess establishes PQM as a robust and transferable solution for unsupervised segmentation quality assessment, providing comprehensive pixel-level evaluation.

Abstract: High-quality image segmentation is fundamental to pixel-level geospatial analysis in remote sensing, necessitating robust segmentation quality assessment (SQA), particularly in unsupervised settings lacking ground truth. Although recent deep learning (DL) based unsupervised SQA methods show potential, they often suffer from coarse evaluation granularity, incomplete assessments, and poor transferability. To overcome these limitations, this paper introduces Panoramic Quality Mapping (PQM) as a new paradigm for comprehensive, pixel-wise SQA, and presents SegAssess, a novel deep learning framework realizing this approach. SegAssess distinctively formulates SQA as a fine-grained, four-class panoramic segmentation task, classifying pixels within a segmentation mask under evaluation into true positive (TP), false positive (FP), true negative (TN), and false negative (FN) categories, thereby generating a complete quality map. Leveraging an enhanced Segment Anything Model (SAM) architecture, SegAssess uniquely employs the input mask as a prompt for effective feature integration via cross-attention. Key innovations include an Edge Guided Compaction (EGC) branch with an Aggregated Semantic Filter (ASF) module to refine predictions near challenging object edges, and an Augmented Mixup Sampling (AMS) training strategy integrating multi-source masks to significantly boost cross-domain robustness and zero-shot transferability. Comprehensive experiments across 32 datasets derived from 6 sources demonstrate that SegAssess achieves state-of-the-art (SOTA) performance and exhibits remarkable zero-shot transferability to unseen masks, establishing PQM via SegAssess as a robust and transferable solution for unsupervised SQA. The code is available at https://github.com/Yangbn97/SegAssess.

[347] PrediTree: A Multi-Temporal Sub-meter Dataset of Multi-Spectral Imagery Aligned With Canopy Height Maps

Hiyam Debary, Mustansar Fiaz, Levente Klein

Main category: cs.CV

TL;DR: PrediTree is the first open-source dataset for high-resolution tree height prediction, combining LiDAR canopy height maps with multi-temporal satellite imagery across French forests, enabling deep learning models to predict tree growth.

DetailsMotivation: Address the critical gap in forest monitoring capabilities by providing comprehensive training data for predicting tree height and growth at sub-meter resolution using deep learning methods.

Method: Created a dataset with 3,141,568 images combining 0.5m LiDAR-derived canopy height maps and multi-temporal multi-spectral imagery. Proposed an encoder-decoder framework using U-Net architecture that processes multi-temporal imagery and relative time differences to predict canopy height.

Result: U-Net architecture achieved the highest performance with 11.78% masked mean squared error, outperforming ResNet-50 by ~12% and RGB-only experiments by ~30%.

Conclusion: PrediTree enables effective training of deep learning models for tree height prediction and provides a valuable open-source resource for forest monitoring research, with the dataset and code publicly available.

Abstract: We present PrediTree, the first comprehensive open-source dataset designed for training and evaluating tree height prediction models at sub-meter resolution. This dataset combines very high-resolution (0.5m) LiDAR-derived canopy height maps, spatially aligned with multi-temporal and multi-spectral imagery, across diverse forest ecosystems in France, totaling 3,141,568 images. PrediTree addresses a critical gap in forest monitoring capabilities by enabling the training of deep learning methods that can predict tree growth based on multiple past observations. %\sout{Initially focused on French forests, PrediTree is designed as an expanding resource with ongoing efforts to incorporate data from other countries. } To make use of this PrediTree dataset, we propose an encoder-decoder framework that requires the multi-temporal multi-spectral imagery and the relative time differences in years between the canopy height map timestamp (target) and each image acquisition date for which this framework predicts the canopy height. The conducted experiments demonstrate that a U-Net architecture trained on the PrediTree dataset provides the highest masked mean squared error of $11.78%$, outperforming the next-best architecture, ResNet-50, by around $12%$, and cutting the error of the same experiments but on fewer bands (red, green, blue only), by around $30%$. This dataset is publicly available on \href{URL}{HuggingFace}, and both processing and training codebases are available on \href{URL}{GitHub}.

[348] DcMatch: Unsupervised Multi-Shape Matching with Dual-Level Consistency

Tianwei Ye, Yong Ma, Xiaoguang Mei

Main category: cs.CV

TL;DR: DcMatch is an unsupervised learning framework for non-rigid multi-shape matching that uses shape graph attention networks and dual-domain consistency to achieve state-of-the-art performance.

DetailsMotivation: Existing methods learn canonical embeddings from single shapes, but there's a need to capture the underlying manifold structure of entire shape collections for more robust correspondences.

Method: Uses shape graph attention network to capture manifold structure, constructs shared latent space, employs universe predictor for consistent correspondences, and enforces dual-level consistency through spatial/spectral domain alignment with cycle consistency loss.

Result: Extensive experiments show consistent outperformance of previous state-of-the-art approaches across diverse multi-shape matching scenarios.

Conclusion: DcMatch provides a novel unsupervised framework that effectively handles non-rigid multi-shape matching through manifold-aware learning and dual-domain consistency, achieving superior performance.

Abstract: Establishing point-to-point correspondences across multiple 3D shapes is a fundamental problem in computer vision and graphics. In this paper, we introduce DcMatch, a novel unsupervised learning framework for non-rigid multi-shape matching. Unlike existing methods that learn a canonical embedding from a single shape, our approach leverages a shape graph attention network to capture the underlying manifold structure of the entire shape collection. This enables the construction of a more expressive and robust shared latent space, leading to more consistent shape-to-universe correspondences via a universe predictor. Simultaneously, we represent these correspondences in both the spatial and spectral domains and enforce their alignment in the shared universe space through a novel cycle consistency loss. This dual-level consistency fosters more accurate and coherent mappings. Extensive experiments on several challenging benchmarks demonstrate that our method consistently outperforms previous state-of-the-art approaches across diverse multi-shape matching scenarios. Code is available at https://github.com/YeTianwei/DcMatch.

[349] Generalizable Self-supervised Monocular Depth Estimation with Mixture of Low-Rank Experts for Diverse Endoscopic Scenes

Liangjing Shao, Benshuang Chen, Chenkang Du, Xueli Liu, Xinrong Chen

Main category: cs.CV

TL;DR: Self-supervised monocular depth estimation framework for endoscopic scenes using block-wise mixture of dynamic low-rank experts to handle illumination variations and diverse tissue features.

DetailsMotivation: Address challenges in generalizable depth estimation for endoscopic scenes due to varying illumination conditions and diverse tissue features, enabling low-cost 3D scene perception for minimally invasive procedures.

Method: Proposes a novel block-wise mixture of dynamic low-rank experts that adaptively selects different experts based on input features, using small trainable parameters. Also introduces a self-supervised training framework to handle brightness and reflectance inconsistencies.

Result: Outperforms state-of-the-art methods on both realistic and simulated endoscopic datasets, achieving best generalization through zero-shot depth estimation on diverse endoscopic scenes.

Conclusion: The proposed framework enables accurate endoscopic perception for minimally invasive measurement and surgery, with code to be released upon acceptance.

Abstract: Self-supervised monocular depth estimation is a significant task for low-cost and efficient three-dimensional scene perception in endoscopy. The variety of illumination conditions and scene features is still the primary challenge for generalizable depth estimation in endoscopic scenes. In this work, a self-supervised framework is proposed for monocular depth estimation in various endoscopy. Firstly, due to various features in endoscopic scenes with different tissues, a novel block-wise mixture of dynamic low-rank experts is proposed to efficiently finetuning the foundation model for endoscopic depth estimation. In the proposed module, based on the input feature, different experts with a small amount of trainable parameters are adaptively selected for weighted inference, from various mixture of low-rank experts which are allocated based on the training quality of each block. Moreover, a novel self-supervised training framework is proposed to jointly cope with the inconsistency of brightness and reflectance. The proposed method outperform state-of-the-art works on both realistic and simulated endoscopic datasets. Furthermore, the proposed network also achieves the best generalization based on zero-shot depth estimation on diverse endoscopic scenes. The proposed method could contribute to accurate endoscopic perception for minimally invasive measurement and surgery. The code will be released upon acceptance, while the demo video can be found on here: https://endo-gede.netlify.app/.

[350] Measuring Image-Relation Alignment: Reference-Free Evaluation of VLMs and Synthetic Pre-training for Open-Vocabulary Scene Graph Generation

Maëlic Neau, Zoe Falomir, Cédric Buche, Akihiro Sugimoto

Main category: cs.CV

TL;DR: Proposes a new reference-free metric for evaluating open-vocabulary scene graph generation and a method for generating high-quality synthetic training data through region-specific prompt tuning of VLMs.

DetailsMotivation: Current SGG benchmarks have limited vocabulary, making open-vocabulary evaluation inefficient, and existing weakly supervised pre-training data is of poor quality.

Method: Develops a reference-free evaluation metric for open-vocabulary relation prediction and creates a region-specific prompt tuning approach for VLMs to generate high-quality synthetic training data.

Result: Experimental results demonstrate that pre-training with the new synthetic data improves the generalization capabilities of open-vocabulary SGG models.

Conclusion: The proposed metric enables fair evaluation of open-vocabulary capabilities, and the synthetic data generation method effectively enhances model performance for scene graph generation tasks.

Abstract: Scene Graph Generation (SGG) encodes visual relationships between objects in images as graph structures. Thanks to the advances of Vision-Language Models (VLMs), the task of Open-Vocabulary SGG has been recently proposed where models are evaluated on their functionality to learn a wide and diverse range of relations. Current benchmarks in SGG, however, possess a very limited vocabulary, making the evaluation of open-source models inefficient. In this paper, we propose a new reference-free metric to fairly evaluate the open-vocabulary capabilities of VLMs for relation prediction. Another limitation of Open-Vocabulary SGG is the reliance on weakly supervised data of poor quality for pre-training. We also propose a new solution for quickly generating high-quality synthetic data through region-specific prompt tuning of VLMs. Experimental results show that pre-training with this new data split can benefit the generalization capabilities of Open-Voc SGG models.

[351] POINTS-Reader: Distillation-Free Adaptation of Vision-Language Models for Document Conversion

Yuan Liu, Zhongyin Zhao, Le Tian, Haicheng Wang, Xubing Ye, Yangxiu You, Zilin Yu, Chuhan Wu, Xiao Zhou, Yang Yu, Jie Zhou

Main category: cs.CV

TL;DR: Proposes a fully automated, distillation-free framework with synthetic data generation and self-improvement stages to create high-quality document extraction datasets and models that handle diverse formats without manual annotation.

DetailsMotivation: Manual annotation is costly and time-consuming, while automatic labeling with existing models lacks accuracy in complex document formats. Teacher-student distillation limits real-world performance.

Method: Two-stage framework: 1) Generate large-scale synthetic data for initial model training, 2) Self-improvement approach using fine-tuned model to annotate real documents with quality filtering, then iterative retraining.

Result: Trained POINTS-1.5 model achieves POINTS-Reader that surpasses many existing public and proprietary models of comparable or larger size.

Conclusion: The framework enables automated creation of high-quality document extraction datasets and models capable of handling diverse document formats without manual annotation, achieving state-of-the-art performance.

Abstract: High-quality labeled data is essential for training accurate document conversion models, particularly in domains with complex formats such as tables, formulas, and multi-column text. However, manual annotation is both costly and time-consuming, while automatic labeling using existing models often lacks accuracy in handling such challenging scenarios. Consequently, training student models by distilling outputs from teacher models can significantly limit their performance in real-world applications. In this paper, we propose a fully automated, distillation-free framework comprising two stages for constructing high-quality document extraction datasets and models capable of handling diverse document formats and layouts. In the first stage, we introduce a method for generating large-scale, diverse synthetic data, which enables a model to extract key elements in a unified format with strong initial performance. In the second stage, we present a self-improvement approach that further adapts the model, initially trained on synthetic data, to real-world documents. Specifically, we first use the fine-tuned model to annotate real documents, then apply a suite of filtering strategies to verify annotation quality, and finally retrain the model on the verified dataset. By iteratively repeating this process, we progressively enhance both the model’s conversion capabilities and the quality of the generated data. We train a public POINTS-1.5 model to obtain POINTS-Reader, which surpasses many existing public and proprietary models of comparable or larger size. Our model is available at https://github.com/Tencent/POINTS-Reader.

[352] FantasyHSI: Video-Generation-Centric 4D Human Synthesis In Any Scene through A Graph-based Multi-Agent Framework

Lingzhou Mu, Qiang Wang, Fan Jiang, Mengchao Wang, Yaqi Fan, Mu Xu, Kai Zhang

Main category: cs.CV

TL;DR: FantasyHSI is a novel Human-Scene Interaction framework that uses video generation and multi-agent systems to generate realistic human behaviors in complex environments without paired data, addressing challenges in long-horizon tasks and scene generalization.

DetailsMotivation: Human-Scene Interaction faces challenges in handling long-horizon, high-level tasks and generalizing to unseen scenes. Existing methods struggle with trajectory drift and physical realism issues.

Method: Models interaction as a dynamic directed graph with a collaborative multi-agent system: scene navigator for environmental perception and path planning, planning agent for goal decomposition, and critic agent for closed-loop feedback correction. Uses Direct Preference Optimization for action generation to enhance physical realism.

Result: Extensive experiments on SceneBench benchmark show FantasyHSI significantly outperforms existing methods in generalization, long-horizon task completion, and physical realism, reducing artifacts like limb distortion and foot-sliding.

Conclusion: FantasyHSI provides an effective framework for realistic human-scene interaction through multi-agent collaboration and feedback mechanisms, demonstrating superior performance in complex, unseen environments.

Abstract: Human-Scene Interaction (HSI) seeks to generate realistic human behaviors within complex environments, yet it faces significant challenges in handling long-horizon, high-level tasks and generalizing to unseen scenes. To address these limitations, we introduce FantasyHSI, a novel HSI framework centered on video generation and multi-agent systems that operates without paired data. We model the complex interaction process as a dynamic directed graph, upon which we build a collaborative multi-agent system. This system comprises a scene navigator agent for environmental perception and high-level path planning, and a planning agent that decomposes long-horizon goals into atomic actions. Critically, we introduce a critic agent that establishes a closed-loop feedback mechanism by evaluating the deviation between generated actions and the planned path. This allows for the dynamic correction of trajectory drifts caused by the stochasticity of the generative model, thereby ensuring long-term logical consistency. To enhance the physical realism of the generated motions, we leverage Direct Preference Optimization (DPO) to train the action generator, significantly reducing artifacts such as limb distortion and foot-sliding. Extensive experiments on our custom SceneBench benchmark demonstrate that FantasyHSI significantly outperforms existing methods in terms of generalization, long-horizon task completion, and physical realism. Ours project page: https://fantasy-amap.github.io/fantasy-hsi/

[353] RT-DETRv2 Explained in 8 Illustrations

Ethan Qi Yang Chua, Jen Hong Tan

Main category: cs.CV

TL;DR: This paper provides visual explanations of RT-DETRv2 architecture through eight detailed illustrations to make this complex object detection model more understandable.

DetailsMotivation: Object detection architectures like RT-DETRv2 are notoriously difficult to understand, and existing diagrams fail to clarify how components actually work and fit together.

Method: The authors use a series of eight carefully designed illustrations to explain the architecture, moving from the overall pipeline down to critical components including encoder, decoder, and multi-scale deformable attention.

Result: The paper provides visualizations of tensor flow and unpacks the logic behind each module to create a clearer mental model of RT-DETRv2’s internal workings.

Conclusion: The goal is to make RT-DETRv2 genuinely understandable for researchers and practitioners by providing comprehensive visual explanations of its architecture and component interactions.

Abstract: Object detection architectures are notoriously difficult to understand, often more so than large language models. While RT-DETRv2 represents an important advance in real-time detection, most existing diagrams do little to clarify how its components actually work and fit together. In this article, we explain the architecture of RT-DETRv2 through a series of eight carefully designed illustrations, moving from the overall pipeline down to critical components such as the encoder, decoder, and multi-scale deformable attention. Our goal is to make the existing one genuinely understandable. By visualizing the flow of tensors and unpacking the logic behind each module, we hope to provide researchers and practitioners with a clearer mental model of how RT-DETRv2 works under the hood.

[354] Learning Correlation-aware Aleatoric Uncertainty for 3D Hand Pose Estimation

Lee Chae-Yeon, Nam Hyeon-Woo, Tae-Hyun Oh

Main category: cs.CV

TL;DR: Proposes a novel aleatoric uncertainty modeling method for 3D hand pose estimation that captures joint correlations using a single linear layer, achieving better accuracy and uncertainty modeling capabilities.

DetailsMotivation: Existing 3D hand pose estimation methods cannot estimate aleatoric uncertainty and lack uncertainty modeling that incorporates joint correlation knowledge, which is crucial for handling complex hand movements, self-similarity, and occlusions.

Method: Introduces aleatoric uncertainty modeling by formulating hand joint output as a probabilistic distribution and using a single linear layer to capture intrinsic correlations among hand joints. This parameterization serves as an add-on module that can be applied to existing models.

Result: The proposed parameterization for uncertainty modeling outperforms existing approaches, and the 3D hand pose estimation model equipped with this uncertainty head achieves favorable accuracy while adding new uncertainty modeling capability.

Conclusion: The method successfully addresses limitations in existing 3D hand pose estimation by incorporating aleatoric uncertainty modeling with joint correlation knowledge, achieving a good balance between modeling accuracy and computational efficiency.

Abstract: 3D hand pose estimation is a fundamental task in understanding human hands. However, accurately estimating 3D hand poses remains challenging due to the complex movement of hands, self-similarity, and frequent occlusions. In this work, we address two limitations: the inability of existing 3D hand pose estimation methods to estimate aleatoric (data) uncertainty, and the lack of uncertainty modeling that incorporates joint correlation knowledge, which has not been thoroughly investigated. To this end, we introduce aleatoric uncertainty modeling into the 3D hand pose estimation framework, aiming to achieve a better trade-off between modeling joint correlations and computational efficiency. We propose a novel parameterization that leverages a single linear layer to capture intrinsic correlations among hand joints. This is enabled by formulating the hand joint output space as a probabilistic distribution, allowing the linear layer to capture joint correlations. Our proposed parameterization is used as a task head layer, and can be applied as an add-on module on top of the existing models. Our experiments demonstrate that our parameterization for uncertainty modeling outperforms existing approaches. Furthermore, the 3D hand pose estimation model equipped with our uncertainty head achieves favorable accuracy in 3D hand pose estimation while introducing new uncertainty modeling capability to the model. The project page is available at https://hand-uncertainty.github.io/.

[355] Towards More Diverse and Challenging Pre-training for Point Cloud Learning: Self-Supervised Cross Reconstruction with Decoupled Views

Xiangdong Zhang, Shaofeng Zhang, Junchi Yan

Main category: cs.CV

TL;DR: Point-PQAE is a novel two-view cross-reconstruction method for point cloud self-supervised learning that outperforms single-view self-reconstruction approaches by generating decoupled views and reconstructing one from the other.

DetailsMotivation: Most existing generative approaches for point cloud self-supervised learning focus on single-view self-reconstruction. The authors recognize that two-view pre-training introduces greater diversity and variance, enabling more challenging and informative pre-training.

Method: Proposes Point-PQAE with a cross-reconstruction generative paradigm that generates two decoupled point cloud views and reconstructs one from the other. Introduces a crop mechanism for view generation and a novel positional encoding to represent 3D relative position between views.

Result: Outperforms self-reconstruction baseline (Point-MAE) by 6.5%, 7.0%, and 6.7% in three variants of ScanObjectNN with Mlp-Linear evaluation protocol.

Conclusion: Cross-reconstruction significantly increases pre-training difficulty compared to self-reconstruction, enabling superior performance in 3D self-supervised learning over single-modal self-reconstruction methods.

Abstract: Point cloud learning, especially in a self-supervised way without manual labels, has gained growing attention in both vision and learning communities due to its potential utility in a wide range of applications. Most existing generative approaches for point cloud self-supervised learning focus on recovering masked points from visible ones within a single view. Recognizing that a two-view pre-training paradigm inherently introduces greater diversity and variance, it may thus enable more challenging and informative pre-training. Inspired by this, we explore the potential of two-view learning in this domain. In this paper, we propose Point-PQAE, a cross-reconstruction generative paradigm that first generates two decoupled point clouds/views and then reconstructs one from the other. To achieve this goal, we develop a crop mechanism for point cloud view generation for the first time and further propose a novel positional encoding to represent the 3D relative position between the two decoupled views. The cross-reconstruction significantly increases the difficulty of pre-training compared to self-reconstruction, which enables our method to surpass previous single-modal self-reconstruction methods in 3D self-supervised learning. Specifically, it outperforms the self-reconstruction baseline (Point-MAE) by 6.5%, 7.0%, and 6.7% in three variants of ScanObjectNN with the Mlp-Linear evaluation protocol. The code is available at https://github.com/aHapBean/Point-PQAE.

[356] ReCap: Event-Aware Image Captioning with Article Retrieval and Semantic Gaussian Normalization

Thinh-Phuc Nguyen, Thanh-Hai Nguyen, Gia-Huy Dinh, Lam-Huy Nguyen, Minh-Triet Tran, Trung-Nghia Le

Main category: cs.CV

TL;DR: ReCap is a novel pipeline that enhances image captioning by incorporating contextual information from relevant articles to generate narrative-rich, factually grounded captions, achieving 2nd place in the EVENTA 2025 challenge.

DetailsMotivation: Standard image captioning systems produce generic descriptions that fail to capture event-level semantics crucial for applications like news reporting and digital archiving, missing temporal, social, and historical contexts.

Method: Three-component pipeline: (1) two-stage article retrieval using DINOv2 embeddings with global feature similarity and patch-level mutual nearest neighbor re-ranking, (2) context extraction framework synthesizing article summaries, generic captions, and source metadata, (3) LLM-based caption generation with Semantic Gaussian Normalization.

Result: Achieved overall score of 0.54666 on OpenEvents V1 dataset, ranking 2nd on the private test set in EVENTA 2025 Grand Challenge Track 1.

Conclusion: ReCap effectively bridges visual perception with real-world knowledge, offering a practical solution for context-aware image understanding in high-stakes domains.

Abstract: Image captioning systems often produce generic descriptions that fail to capture event-level semantics which are crucial for applications like news reporting and digital archiving. We present ReCap, a novel pipeline for event-enriched image retrieval and captioning that incorporates broader contextual information from relevant articles to generate narrative-rich, factually grounded captions. Our approach addresses the limitations of standard vision-language models that typically focus on visible content while missing temporal, social, and historical contexts. ReCap comprises three integrated components: (1) a robust two-stage article retrieval system using DINOv2 embeddings with global feature similarity for initial candidate selection followed by patch-level mutual nearest neighbor similarity re-ranking; (2) a context extraction framework that synthesizes information from article summaries, generic captions, and original source metadata; and (3) a large language model-based caption generation system with Semantic Gaussian Normalization to enhance fluency and relevance. Evaluated on the OpenEvents V1 dataset as part of Track 1 in the EVENTA 2025 Grand Challenge, ReCap achieved a strong overall score of 0.54666, ranking 2nd on the private test set. These results highlight ReCap’s effectiveness in bridging visual perception with real-world knowledge, offering a practical solution for context-aware image understanding in high-stakes domains. The code is available at https://github.com/Noridom1/EVENTA2025-Event-Enriched-Image-Captioning.

[357] Novel Category Discovery with X-Agent Attention for Open-Vocabulary Semantic Segmentation

Jiahao Li Yang Lu, Yachao Zhang, Fangyong Wang, Yuan Xie, Yanyun Qu

Main category: cs.CV

TL;DR: X-Agent is a novel open-vocabulary semantic segmentation framework that uses latent semantic-aware agents to optimize cross-modal attention and enhance latent semantic perception, achieving state-of-the-art performance.

DetailsMotivation: Address the domain discrepancy challenge in open-vocabulary semantic segmentation where existing VLM-based approaches lack exploration of fundamental latent semantic comprehension mechanisms.

Method: Proposes X-Agent framework with latent semantic-aware agents that orchestrate cross-modal attention mechanisms to optimize latent semantic dynamics and amplify perceptibility.

Result: Achieves state-of-the-art performance in benchmark evaluations while effectively enhancing latent semantic saliency.

Conclusion: X-Agent successfully addresses the bottleneck in OVSS by improving latent semantic comprehension through innovative agent-based cross-modal attention optimization.

Abstract: Open-vocabulary semantic segmentation (OVSS) conducts pixel-level classification via text-driven alignment, where the domain discrepancy between base category training and open-vocabulary inference poses challenges in discriminative modeling of latent unseen category. To address this challenge, existing vision-language model (VLM)-based approaches demonstrate commendable performance through pre-trained multi-modal representations. However, the fundamental mechanisms of latent semantic comprehension remain underexplored, making the bottleneck for OVSS. In this work, we initiate a probing experiment to explore distribution patterns and dynamics of latent semantics in VLMs under inductive learning paradigms. Building on these insights, we propose X-Agent, an innovative OVSS framework employing latent semantic-aware ``agent’’ to orchestrate cross-modal attention mechanisms, simultaneously optimizing latent semantic dynamic and amplifying its perceptibility. Extensive benchmark evaluations demonstrate that X-Agent achieves state-of-the-art performance while effectively enhancing the latent semantic saliency.

[358] SAR-NAS: Lightweight SAR Object Detection with Neural Architecture Search

Xinyi Yu, Zhiwei Lin, Yongtao Wang

Main category: cs.CV

TL;DR: This paper applies Neural Architecture Search (NAS) to optimize YOLOv10 for SAR object detection, achieving superior accuracy with lower computational overhead compared to existing methods.

DetailsMotivation: SAR object detection faces challenges from speckle noise, small target ambiguities, and computational constraints. Existing approaches focus on SAR-specific architectural modifications, but this paper explores leveraging lightweight object detectors enhanced through NAS.

Method: Employed Neural Architecture Search (NAS) to systematically optimize YOLOv10’s network structure, with focus on backbone architecture search. Used evolutionary search within an extensive search space to identify optimal architecture balancing accuracy, parameter efficiency, and computational cost.

Result: Experimental results on the large-scale SARDet-100K dataset show the optimized model outperforms existing SAR detection methods, achieving superior detection accuracy while maintaining lower computational overhead.

Conclusion: This work introduces NAS to SAR object detection for the first time and demonstrates its effectiveness in optimizing lightweight detectors for real-world applications, offering a novel perspective on leveraging NAS in practical scenarios.

Abstract: Synthetic Aperture Radar (SAR) object detection faces significant challenges from speckle noise, small target ambiguities, and on-board computational constraints. While existing approaches predominantly focus on SAR-specific architectural modifications, this paper explores the application of the existing lightweight object detector, i.e., YOLOv10, for SAR object detection and enhances its performance through Neural Architecture Search (NAS). Specifically, we employ NAS to systematically optimize the network structure, especially focusing on the backbone architecture search. By constructing an extensive search space and leveraging evolutionary search, our method identifies a favorable architecture that balances accuracy, parameter efficiency, and computational cost. Notably, this work introduces NAS to SAR object detection for the first time. The experimental results on the large-scale SARDet-100K dataset demonstrate that our optimized model outperforms existing SAR detection methods, achieving superior detection accuracy while maintaining lower computational overhead. We hope this work offers a novel perspective on leveraging NAS for real-world applications.

[359] Multi-Representation Adapter with Neural Architecture Search for Efficient Range-Doppler Radar Object Detection

Zhiwei Lin, Weicheng Zheng, Yongtao Wang

Main category: cs.CV

TL;DR: Efficient radar object detection model using multi-representation inputs and neural architecture search to achieve state-of-the-art performance on RADDet and CARRADA datasets.

DetailsMotivation: Radar sensors offer robustness against adverse lighting and weather conditions compared to cameras, making them valuable for reliable object detection in challenging environments.

Method: Uses multi-representation RD radar maps (heatmaps + grayscale images), designs Adapter branch, Exchanger Module, and Primary-Auxiliary Fusion Module for feature extraction and fusion, and employs One-Shot Neural Architecture Search for efficiency optimization.

Result: Achieves mAP@50 of 71.9 on RADDet and 57.1 on CARRADA datasets, demonstrating favorable accuracy-efficiency trade-off and state-of-the-art performance.

Conclusion: The proposed model effectively combines multi-representation radar data with optimized architecture search to deliver efficient and high-performance object detection for radar applications.

Abstract: Detecting objects efficiently from radar sensors has recently become a popular trend due to their robustness against adverse lighting and weather conditions compared with cameras. This paper presents an efficient object detection model for Range-Doppler (RD) radar maps. Specifically, we first represent RD radar maps with multi-representation, i.e., heatmaps and grayscale images, to gather high-level object and fine-grained texture features. Then, we design an additional Adapter branch, an Exchanger Module with two modes, and a Primary-Auxiliary Fusion Module to effectively extract, exchange, and fuse features from the multi-representation inputs, respectively. Furthermore, we construct a supernet with various width and fusion operations in the Adapter branch for the proposed model and employ a One-Shot Neural Architecture Search method to further improve the model’s efficiency while maintaining high performance. Experimental results demonstrate that our model obtains favorable accuracy and efficiency trade-off. Moreover, we achieve new state-of-the-art performance on RADDet and CARRADA datasets with mAP@50 of 71.9 and 57.1, respectively.

[360] Cross-Domain Few-Shot Segmentation via Ordinary Differential Equations over Time Intervals

Huan Ni, Qingshan Liu, Xiaonan Niu, Danfeng Hong, Lingli Zhao, Haiyan Guan

Main category: cs.CV

TL;DR: Proposes FSS-TIs, an all-in-one module using ODEs and Fourier transform for cross-domain few-shot segmentation, achieving superior performance over existing methods.

DetailsMotivation: Existing CD-FSS methods use multiple independent modules that hinder knowledge flow and limit collective potential. Need for a more integrated approach to enhance cross-domain generalization.

Method: Uses ordinary differential equations and Fourier transform to model relationship between domain-specific and domain-agnostic features. Applies iterative transformation with randomized perturbations to spectra, reformulating as ODE optimization problem.

Result: Experimental results on five diverse datasets show FSS-TIs outperforms existing CD-FSS methods. Ablation studies validate cross-domain adaptability.

Conclusion: FSS-TIs provides a structurally concise and effective solution for cross-domain few-shot segmentation, demonstrating superior generalization across substantially different domains.

Abstract: Cross-domain few-shot segmentation (CD-FSS) not only enables the segmentation of unseen categories with very limited samples, but also improves cross-domain generalization ability within the few-shot segmentation framework. Currently, existing CD-FSS studies typically design multiple independent modules to enhance the cross-domain generalization ability of feature representations. However, the independence among these modules hinders the effective flow of knowledge, making it difficult to fully leverage their collective potential. In contrast, this paper proposes an all-in-one module based on ordinary differential equations and Fourier transform, resulting in a structurally concise method–Few-Shot Segmentation over Time Intervals (FSS-TIs). FSS-TIs assumes the existence of an ODE relationship between the spectra (including amplitude and phase spectra) of domain-specific features and domain-agnostic features. This ODE formulation yields an iterative transformation process along a sequence of time intervals, while simultaneously applying affine transformations with randomized perturbations to the spectra. In doing so, the exploration of domain-agnostic feature representation spaces and the simulation of diverse potential target-domain distributions are reformulated as an optimization process over the intrinsic parameters of the ODE. Moreover, we strictly constrain the support-sample selection during target-domain fine-tuning so that it is consistent with the requirements of real-world few-shot segmentation tasks. For evaluation, we introduce five datasets from substantially different domains and define two sets of cross-domain few-shot segmentation tasks to comprehensively analyze the performance of FSS-TIs. Experimental results demonstrate the superiority of FSS-TIs over existing CD-FSS methods, and in-depth ablation studies further validate the cross-domain adaptability of FSS-TIs.

[361] Guided Model-based LiDAR Super-Resolution for Resource-Efficient Automotive scene Segmentation

Alexandros Gkillas, Nikos Piperigkos, Aris S. Lalos

Main category: cs.CV

TL;DR: End-to-end framework that jointly optimizes LiDAR super-resolution and semantic segmentation, achieving comparable performance to high-cost 64-channel LiDAR using sparse 16-channel data with fewer parameters.

DetailsMotivation: High-resolution LiDAR is expensive and limits large-scale deployment for autonomous driving, while low-cost 16-channel LiDAR produces sparse point clouds that degrade segmentation accuracy.

Method: Joint optimization framework with SR module that incorporates semantic cues, new SR loss function focusing on regions of interest, and lightweight model-based architecture with fewer parameters.

Result: Achieves segmentation performance comparable to models using high-resolution 64-channel LiDAR data.

Conclusion: Proposed framework enables cost-effective autonomous driving deployment by effectively enhancing sparse LiDAR data through joint super-resolution and semantic segmentation optimization.

Abstract: High-resolution LiDAR data plays a critical role in 3D semantic segmentation for autonomous driving, but the high cost of advanced sensors limits large-scale deployment. In contrast, low-cost sensors such as 16-channel LiDAR produce sparse point clouds that degrade segmentation accuracy. To overcome this, we introduce the first end-to-end framework that jointly addresses LiDAR super-resolution (SR) and semantic segmentation. The framework employs joint optimization during training, allowing the SR module to incorporate semantic cues and preserve fine details, particularly for smaller object classes. A new SR loss function further directs the network to focus on regions of interest. The proposed lightweight, model-based SR architecture uses significantly fewer parameters than existing LiDAR SR approaches, while remaining easily compatible with segmentation networks. Experiments show that our method achieves segmentation performance comparable to models operating on high-resolution and costly 64-channel LiDAR data.

[362] Prior-Guided Residual Diffusion: Calibrated and Efficient Medical Image Segmentation

Fuyou Mao, Beining Wu, Yanfeng Jiang, Han Xue, Yan Tang, Hao Zhang

Main category: cs.CV

TL;DR: PGRD is a diffusion-based framework for medical image segmentation that learns voxel-wise distributions using prior guidance and residual learning to improve calibration and sampling efficiency.

DetailsMotivation: Medical image segmentation often involves ambiguity that requires capturing full conditional distributions rather than single point estimates, necessitating probabilistic models that maintain calibration while being computationally practical.

Method: PGRD embeds discrete labels in continuous space, uses a coarse prior predictor for step-wise guidance, learns residual to the prior for faster convergence, and employs deep diffusion supervision to stabilize training by supervising intermediate time steps.

Result: PGRD achieves higher Dice scores and lower NLL/ECE values compared to Bayesian, ensemble, Probabilistic U-Net, and vanilla diffusion baselines on MRI and CT datasets, while requiring fewer sampling steps for strong performance.

Conclusion: The proposed Prior-Guided Residual Diffusion framework effectively captures conditional distributions in medical image segmentation with improved calibration, accuracy, and computational efficiency compared to existing probabilistic methods.

Abstract: Ambiguity in medical image segmentation calls for models that capture full conditional distributions rather than a single point estimate. We present Prior-Guided Residual Diffusion (PGRD), a diffusion-based framework that learns voxel-wise distributions while maintaining strong calibration and practical sampling efficiency. PGRD embeds discrete labels as one-hot targets in a continuous space to align segmentation with diffusion modeling. A coarse prior predictor provides step-wise guidance; the diffusion network then learns the residual to the prior, accelerating convergence and improving calibration. A deep diffusion supervision scheme further stabilizes training by supervising intermediate time steps. Evaluated on representative MRI and CT datasets, PGRD achieves higher Dice scores and lower NLL/ECE values than Bayesian, ensemble, Probabilistic U-Net, and vanilla diffusion baselines, while requiring fewer sampling steps to reach strong performance.

[363] Street-Level Geolocalization Using Multimodal Large Language Models and Retrieval-Augmented Generation

Yunus Serhat Bicakci, Joseph Shingleton, Anahid Basiri

Main category: cs.CV

TL;DR: Novel geolocalization method using multimodal LLMs with retrieval-augmented generation, achieving state-of-the-art accuracy without fine-tuning.

DetailsMotivation: Street-level geolocalization from images is essential for navigation, location-based services, and urban planning, but traditional computer vision methods struggle with social media and smartphone camera data.

Method: Integrates open-weight multimodal LLMs with retrieval-augmented generation. Uses SigLIP encoder to build vector database from EMP-16 and OSV-5M datasets. Augments query images with retrieved similar/dissimilar geolocation prompts before processing.

Result: Achieved state-of-the-art performance with higher accuracy on benchmark datasets (IM2GPS, IM2GPS3k, YFCC4k). Eliminates need for expensive fine-tuning and scales seamlessly with new data.

Conclusion: Retrieval-augmented generation with multimodal LLMs provides an effective alternative to traditional training-from-scratch methods, enabling more accessible and scalable geolocation solutions in GeoAI.

Abstract: Street-level geolocalization from images is crucial for a wide range of essential applications and services, such as navigation, location-based recommendations, and urban planning. With the growing popularity of social media data and cameras embedded in smartphones, applying traditional computer vision techniques to localize images has become increasingly challenging, yet highly valuable. This paper introduces a novel approach that integrates open-weight and publicly accessible multimodal large language models with retrieval-augmented generation. The method constructs a vector database using the SigLIP encoder on two large-scale datasets (EMP-16 and OSV-5M). Query images are augmented with prompts containing both similar and dissimilar geolocation information retrieved from this database before being processed by the multimodal large language models. Our approach has demonstrated state-of-the-art performance, achieving higher accuracy compared against three widely used benchmark datasets (IM2GPS, IM2GPS3k, and YFCC4k). Importantly, our solution eliminates the need for expensive fine-tuning or retraining and scales seamlessly to incorporate new data sources. The effectiveness of retrieval-augmented generation-based multimodal large language models in geolocation estimation demonstrated by this paper suggests an alternative path to the traditional methods which rely on the training models from scratch, opening new possibilities for more accessible and scalable solutions in GeoAI.

[364] AgroSense: An Integrated Deep Learning System for Crop Recommendation via Soil Image Analysis and Nutrient Profiling

Vishal Pandey, Ranjita Das, Debasmita Biswas

Main category: cs.CV

TL;DR: AgroSense is a deep learning framework that combines soil image classification and nutrient analysis to provide real-time crop recommendations with 98% accuracy, addressing limitations of traditional soil analysis methods.

DetailsMotivation: Traditional soil analysis techniques are slow, labor-intensive and unsuitable for on-field decision making, creating a need for intelligent crop recommendation systems to meet global food security and sustainable farming demands.

Method: Uses multimodal deep learning with two modules: Soil Classification Module (ResNet-18, EfficientNet-B0, Vision Transformer for soil type categorization from images) and Crop Recommendation Module (MLP, XGBoost, LightGBM, TabNet for analyzing structured soil data including nutrients, pH, and rainfall).

Result: Achieved 98.0% accuracy, 97.8% precision, 97.7% recall, 96.75% F1-score, with RMSE of 0.32 and MAE of 0.27. Ablation studies confirmed importance of multimodal coupling, and statistical validation showed significant improvements.

Conclusion: AgroSense provides a practical, scalable solution for real-time decision support in precision agriculture and enables future lightweight multimodal AI systems for resource-constrained environments.

Abstract: Meeting the increasing global demand for food security and sustainable farming requires intelligent crop recommendation systems that operate in real time. Traditional soil analysis techniques are often slow, labor-intensive, and not suitable for on-field decision-making. To address these limitations, we introduce AgroSense, a deep-learning framework that integrates soil image classification and nutrient profiling to produce accurate and contextually relevant crop recommendations. AgroSense comprises two main components: a Soil Classification Module, which leverages ResNet-18, EfficientNet-B0, and Vision Transformer architectures to categorize soil types from images; and a Crop Recommendation Module, which employs a Multi-Layer Perceptron, XGBoost, LightGBM, and TabNet to analyze structured soil data, including nutrient levels, pH, and rainfall. We curated a multimodal dataset of 10,000 paired samples drawn from publicly available Kaggle repositories, approximately 50,000 soil images across seven classes, and 25,000 nutrient profiles for experimental evaluation. The fused model achieves 98.0% accuracy, with a precision of 97.8%, a recall of 97.7%, and an F1-score of 96.75%, while RMSE and MAE drop to 0.32 and 0.27, respectively. Ablation studies underscore the critical role of multimodal coupling, and statistical validation via t-tests and ANOVA confirms the significance of our improvements. AgroSense offers a practical, scalable solution for real-time decision support in precision agriculture and paves the way for future lightweight multimodal AI systems in resource-constrained environments.

[365] M3Ret: Unleashing Zero-shot Multimodal Medical Image Retrieval via Self-Supervision

Che Liu, Zheng Jiang, Chengyu Fang, Heng Guo, Yan-Jie Zhou, Jiaqi Qu, Le Lu, Minfeng Xu

Main category: cs.CV

TL;DR: M3Ret is a unified visual encoder that achieves state-of-the-art medical image retrieval across 2D, 3D, and video modalities without modality-specific customization, demonstrating strong cross-modal alignment and generalization to unseen MRI tasks.

DetailsMotivation: Current medical image retrieval methods are fragmented with separate architectures for different modalities (2D, 3D, video), which hampers scalability and prevents unified representation learning.

Method: Curated a large-scale hybrid-modality dataset (867,653 samples including 2D X-rays, ultrasounds, endoscopy videos, and 3D CT scans) and trained M3Ret using both generative (MAE) and contrastive (SimDINO) self-supervised learning paradigms without modality-specific customization.

Result: Achieved new state-of-the-art in zero-shot image-to-image retrieval across all modalities, surpassing DINOv3 and BMC-CLIP. Demonstrated strong cross-modal alignment without paired data and generalization to unseen MRI tasks despite no MRI in pretraining.

Conclusion: M3Ret represents a step toward foundation models for visual self-supervised learning in multimodal medical imaging, showing promising scalability and generalizability across modalities and data sizes.

Abstract: Medical image retrieval is essential for clinical decision-making and translational research, relying on discriminative visual representations. Yet, current methods remain fragmented, relying on separate architectures and training strategies for 2D, 3D, and video-based medical data. This modality-specific design hampers scalability and inhibits the development of unified representations. To enable unified learning, we curate a large-scale hybrid-modality dataset comprising 867,653 medical imaging samples, including 2D X-rays and ultrasounds, RGB endoscopy videos, and 3D CT scans. Leveraging this dataset, we train M3Ret, a unified visual encoder without any modality-specific customization. It successfully learns transferable representations using both generative (MAE) and contrastive (SimDINO) self-supervised learning (SSL) paradigms. Our approach sets a new state-of-the-art in zero-shot image-to-image retrieval across all individual modalities, surpassing strong baselines such as DINOv3 and the text-supervised BMC-CLIP. More remarkably, strong cross-modal alignment emerges without paired data, and the model generalizes to unseen MRI tasks, despite never observing MRI during pretraining, demonstrating the generalizability of purely visual self-supervision to unseen modalities. Comprehensive analyses further validate the scalability of our framework across model and data sizes. These findings deliver a promising signal to the medical imaging community, positioning M3Ret as a step toward foundation models for visual SSL in multimodal medical image understanding.

[366] Uirapuru: Timely Video Analytics for High-Resolution Steerable Cameras on Edge Devices

Guilherme H. Apostolo, Pablo Bauszat, Vinod Nigade, Henri E. Bal, Lin Wang

Main category: cs.CV

TL;DR: Uirapuru is a framework for real-time video analytics on steerable cameras that addresses challenges from camera movement, achieving significant accuracy and speed improvements over static camera approaches.

DetailsMotivation: Existing real-time video analytics systems focus on static cameras, but steerable cameras with pan-tilt-zoom capabilities introduce scene dynamism that breaks traditional approaches like frame tiling.

Method: Uirapuru incorporates comprehensive camera actuation understanding into system design and uses fast adaptive tiling at per-frame level to handle the dynamism from steerable camera movements.

Result: Uirapuru provides up to 1.45x accuracy improvement while meeting latency budgets, or achieves up to 4.53x inference speedup with comparable accuracy compared to state-of-the-art static camera approaches.

Conclusion: The framework successfully addresses the challenges of real-time video analytics on steerable cameras by integrating camera actuation awareness with adaptive tiling techniques.

Abstract: Real-time video analytics on high-resolution cameras has become a popular technology for various intelligent services like traffic control and crowd monitoring. While extensive work has been done on improving analytics accuracy with timing guarantees, virtually all of them target static viewpoint cameras. In this paper, we present Uirapuru, a novel framework for real-time, edge-based video analytics on high-resolution steerable cameras. The actuation performed by those cameras brings significant dynamism to the scene, presenting a critical challenge to existing popular approaches such as frame tiling. To address this problem, Uirapuru incorporates a comprehensive understanding of camera actuation into the system design paired with fast adaptive tiling at a per-frame level. We evaluate Uirapuru on a high-resolution video dataset, augmented by pan-tilt-zoom (PTZ) movements typical for steerable cameras and on real-world videos collected from an actual PTZ camera. Our experimental results show that Uirapuru provides up to 1.45x improvement in accuracy while respecting specified latency budgets or reaches up to 4.53x inference speedup with on-par accuracy compared to state-of-the-art static camera approaches.

[367] Unsupervised Ultra-High-Resolution UAV Low-Light Image Enhancement: A Benchmark, Metric and Framework

Wei Lu, Lingyu Zhu, Si-Bao Chen

Main category: cs.CV

TL;DR: A comprehensive solution for UAV low-light image enhancement with a new dataset, evaluation metric, and efficient framework that achieves real-time 4K processing.

DetailsMotivation: Low light conditions degrade UAV performance, and existing methods struggle with aerial imagery challenges like ultra-high resolution, lack of paired data, non-uniform illumination, and deployment constraints.

Method: Proposes U3D dataset (first unsupervised UHR UAV dataset), Edge Efficiency Index metric, and U3LIE framework with Adaptive Pre-enhancement Augmentation and Luminance Interval Loss for efficient training.

Result: U3LIE achieves state-of-the-art results, processing 4K images at 23.8 FPS on a single GPU, enabling real-time on-board deployment.

Conclusion: Provides a holistic solution (dataset, metric, method) for advancing robust 24/7 UAV vision with available code and datasets.

Abstract: Low light conditions significantly degrade Unmanned Aerial Vehicles (UAVs) performance in critical applications. Existing Low-light Image Enhancement (LIE) methods struggle with the unique challenges of aerial imagery, including Ultra-High Resolution (UHR), lack of paired data, severe non-uniform illumination, and deployment constraints. To address these issues, we propose three key contributions. First, we present U3D, the first unsupervised UHR UAV dataset for LIE, with a unified evaluation toolkit. Second, we introduce the Edge Efficiency Index (EEI), a novel metric balancing perceptual quality with key deployment factors: speed, resolution, model complexity, and memory footprint. Third, we develop U3LIE, an efficient framework with two training-only designs-Adaptive Pre-enhancement Augmentation (APA) for input normalization and a Luminance Interval Loss (L_int) for exposure control. U3LIE achieves SOTA results, processing 4K images at 23.8 FPS on a single GPU, making it ideal for real-time on-board deployment. In summary, these contributions provide a holistic solution (dataset, metric, and method) for advancing robust 24/7 UAV vision. The code and datasets are available at https://github.com/lwCVer/U3D_Toolkit.

[368] RibPull: Implicit Occupancy Fields and Medial Axis Extraction for CT Ribcage Scans

Emmanouil Nikolakakis, Amine Ouasfi, Julie Digne, Razvan Marinescu

Main category: cs.CV

TL;DR: RibPull uses neural occupancy fields to represent ribcages from CT scans as continuous implicit functions, enabling better handling of sparse data and allowing geometric operations like medial axis extraction through Laplacian-based contraction.

DetailsMotivation: Voxel grids in medical imaging suffer from resolution limitations, topological information loss, and inefficient handling of sparse data. Implicit 3D representations using continuous functions can better handle sparse and noisy data while preserving complex geometrical information.

Method: Utilizes neural occupancy fields to predict whether 3D points lie inside or outside ribcage objects from CT scans. Applies Laplacian-based contraction to extract the medial axis of the ribcage, demonstrating geometric operations that benefit from continuous coordinate-based representations.

Result: Methodology evaluated on 20 medical scans from the RibSeg dataset (extension of RibFrac dataset). The approach shows advantages over traditional voxel-based representations for handling sparse data and performing morphological operations.

Conclusion: Implicit occupancy fields provide a superior solution for representing medical imaging data, particularly for sparse CT-scanned ribcages, enabling effective geometric operations that are challenging with discrete voxel-based methods.

Abstract: We present RibPull, a methodology that utilizes implicit occupancy fields to bridge computational geometry and medical imaging. Implicit 3D representations use continuous functions that handle sparse and noisy data more effectively than discrete methods. While voxel grids are standard for medical imaging, they suffer from resolution limitations, topological information loss, and inefficient handling of sparsity. Coordinate functions preserve complex geometrical information and represent a better solution for sparse data representation, while allowing for further morphological operations. Implicit scene representations enable neural networks to encode entire 3D scenes within their weights. The result is a continuous function that can implicitly compesate for sparse signals and infer further information about the 3D scene by passing any combination of 3D coordinates as input to the model. In this work, we use neural occupancy fields that predict whether a 3D point lies inside or outside an object to represent CT-scanned ribcages. We also apply a Laplacian-based contraction to extract the medial axis of the ribcage, thus demonstrating a geometrical operation that benefits greatly from continuous coordinate-based 3D scene representations versus voxel-based representations. We evaluate our methodology on 20 medical scans from the RibSeg dataset, which is itself an extension of the RibFrac dataset. We will release our code upon publication.

[369] Neural Scene Designer: Self-Styled Semantic Image Manipulation

Jianman Lin, Tianshui Chen, Chunmei Qing, Zhijing Yang, Shuangping Huang, Yuheng Ren, Liang Lin

Main category: cs.CV

TL;DR: NSD is a novel framework for photo-realistic image editing that maintains both semantic alignment with user intent and stylistic consistency with the surrounding environment using advanced diffusion models and contrastive style learning.

DetailsMotivation: Existing image editing methods focus primarily on semantic control but neglect stylistic consistency, which is crucial for cohesive and aesthetically appealing images.

Method: Uses Neural Scene Designer (NSD) with parallel cross-attention mechanisms for text and style processing, plus Progressive Self-style Representational Learning (PSRL) module with style contrastive loss to capture fine-grained style representations.

Result: Extensive experiments on a comprehensive benchmark demonstrate the framework’s effectiveness in achieving both semantic control and style consistency.

Conclusion: NSD successfully addresses the gap in stylistic consistency preservation while maintaining semantic alignment, establishing a new standard for style-consistent image editing.

Abstract: Maintaining stylistic consistency is crucial for the cohesion and aesthetic appeal of images, a fundamental requirement in effective image editing and inpainting. However, existing methods primarily focus on the semantic control of generated content, often neglecting the critical task of preserving this consistency. In this work, we introduce the Neural Scene Designer (NSD), a novel framework that enables photo-realistic manipulation of user-specified scene regions while ensuring both semantic alignment with user intent and stylistic consistency with the surrounding environment. NSD leverages an advanced diffusion model, incorporating two parallel cross-attention mechanisms that separately process text and style information to achieve the dual objectives of semantic control and style consistency. To capture fine-grained style representations, we propose the Progressive Self-style Representational Learning (PSRL) module. This module is predicated on the intuitive premise that different regions within a single image share a consistent style, whereas regions from different images exhibit distinct styles. The PSRL module employs a style contrastive loss that encourages high similarity between representations from the same image while enforcing dissimilarity between those from different images. Furthermore, to address the lack of standardized evaluation protocols for this task, we establish a comprehensive benchmark. This benchmark includes competing algorithms, dedicated style-related metrics, and diverse datasets and settings to facilitate fair comparisons. Extensive experiments conducted on our benchmark demonstrate the effectiveness of the proposed framework.

[370] MILO: A Lightweight Perceptual Quality Metric for Image and Latent-Space Optimization

Uğur Çoğalan, Mojtaba Bemana, Karol Myszkowski, Hans-Peter Seidel, Colin Groth

Main category: cs.CV

TL;DR: MILO is a lightweight perceptual metric for image quality assessment that outperforms existing metrics and serves as an effective perceptual loss for image optimization tasks.

DetailsMotivation: To develop an efficient full-reference image quality assessment metric that doesn't require large human-labeled datasets and can be used for perceptual optimization in generative models.

Method: Trained using pseudo-MOS supervision with reproducible distortions scored by an ensemble of quality metrics, featuring spatial masking and curriculum learning for optimization tasks.

Result: Outperforms existing metrics across standard benchmarks, enables efficient perceptual optimization in image and latent domains, and improves performance in denoising, super-resolution, and face restoration tasks.

Conclusion: MILO serves as both a state-of-the-art image quality metric and a practical tool for perceptual optimization in generative pipelines with fast inference suitable for real-time applications.

Abstract: We present MILO (Metric for Image- and Latent-space Optimization), a lightweight, multiscale, perceptual metric for full-reference image quality assessment (FR-IQA). MILO is trained using pseudo-MOS (Mean Opinion Score) supervision, in which reproducible distortions are applied to diverse images and scored via an ensemble of recent quality metrics that account for visual masking effects. This approach enables accurate learning without requiring large-scale human-labeled datasets. Despite its compact architecture, MILO outperforms existing metrics across standard FR-IQA benchmarks and offers fast inference suitable for real-time applications. Beyond quality prediction, we demonstrate the utility of MILO as a perceptual loss in both image and latent domains. In particular, we show that spatial masking modeled by MILO, when applied to latent representations from a VAE encoder within Stable Diffusion, enables efficient and perceptually aligned optimization. By combining spatial masking with a curriculum learning strategy, we first process perceptually less relevant regions before progressively shifting the optimization to more visually distorted areas. This strategy leads to significantly improved performance in tasks like denoising, super-resolution, and face restoration, while also reducing computational overhead. MILO thus functions as both a state-of-the-art image quality metric and as a practical tool for perceptual optimization in generative pipelines.

[371] Bangladeshi Street Food Calorie Estimation Using Improved YOLOv8 and Regression Model

Aparup Dhar, MD Tamim Hossain, Pritom Barua

Main category: cs.CV

TL;DR: A modified YOLOv8-based system for accurate calorie estimation of Bangladeshi street foods, achieving 96.0% R^2 score with low error rates.

DetailsMotivation: Addressing limitations in existing automated calorie tracking systems that struggle with multiple food recognition, image scaling issues, and Western cuisine bias, particularly for diverse Bangladeshi street foods.

Method: Constructed a diverse Bangladeshi street food dataset, modified YOLOv8 for improved classification and segmentation, and integrated a machine learning regression model for calorie estimation.

Result: Achieved 6.94 MAE, 11.03 RMSE, and 96.0% R^2 score in calorie estimation with only slight computational complexity increase compared to base YOLOv8.

Conclusion: The proposed system provides highly effective and accurate real-world calorie calculation specifically tailored for Bangladeshi street foods, overcoming previous limitations in automated food calorie tracking.

Abstract: As obesity rates continue to increase, automated calorie tracking has become a vital tool for people seeking to maintain a healthy lifestyle or adhere to a diet plan. Although numerous research efforts have addressed this issue, existing approaches often face key limitations, such as providing only constant caloric output, struggling with multiple food recognition challenges, challenges in image scaling and normalization, and a predominant focus on Western cuisines. In this paper, we propose a tailored solution that specifically targets Bangladeshi street food. We first construct a diverse dataset of popular street foods found across Bangladesh. Then, we develop a refined calorie estimation system by modifying the state-of-the-art vision model YOLOv8. Our modified model achieves superior classification and segmentation results, with only a slight increase in computational complexity compared to the base variant. Coupled with a machine learning regression model, our system achieves an impressive 6.94 mean absolute error (MAE), 11.03 root mean squared error (RMSE), and a 96.0% R^2 score in calorie estimation, making it both highly effective and accurate for real-world food calorie calculations.

[372] InfoScale: Unleashing Training-free Variable-scaled Image Generation via Effective Utilization of Information

Guohui Zhang, Jiangtong Tan, Linjiang Huang, Zhonghang Yuan, Naishan Zheng, Jie Huang, Feng Zhao

Main category: cs.CV

TL;DR: InfoScale is a plug-and-play framework that addresses resolution-dependent performance issues in diffusion models by solving information loss, aggregation inflexibility, and noise distribution misalignment for variable-scale image generation.

DetailsMotivation: Diffusion models suffer performance degradation when generating images at resolutions different from their training scale due to information handling challenges across varying resolutions.

Method: Proposes InfoScale framework with three modules: Progressive Frequency Compensation for high-frequency information loss, Adaptive Information Aggregation for flexible attention mechanisms, and Noise Adaptation for proper initial noise distribution alignment.

Result: Extensive experiments demonstrate effective variable-scaled image generation, showing the framework’s plug-and-play capability with existing diffusion models.

Conclusion: InfoScale successfully addresses the three critical challenges in variable-scale generation and provides a unified solution that can be integrated with existing diffusion models.

Abstract: Diffusion models (DMs) have become dominant in visual generation but suffer performance drop when tested on resolutions that differ from the training scale, whether lower or higher. In fact, the key challenge in generating variable-scale images lies in the differing amounts of information across resolutions, which requires information conversion procedures to be varied for generating variable-scaled images. In this paper, we investigate the issues of three critical aspects in DMs for a unified analysis in variable-scaled generation: dilated convolution, attention mechanisms, and initial noise. Specifically, 1) dilated convolution in DMs for the higher-resolution generation loses high-frequency information. 2) Attention for variable-scaled image generation struggles to adjust the information aggregation adaptively. 3) The spatial distribution of information in the initial noise is misaligned with variable-scaled image. To solve the above problems, we propose \textbf{InfoScale}, an information-centric framework for variable-scaled image generation by effectively utilizing information from three aspects correspondingly. For information loss in 1), we introduce Progressive Frequency Compensation module to compensate for high-frequency information lost by dilated convolution in higher-resolution generation. For information aggregation inflexibility in 2), we introduce Adaptive Information Aggregation module to adaptively aggregate information in lower-resolution generation and achieve an effective balance between local and global information in higher-resolution generation. For information distribution misalignment in 3), we design Noise Adaptation module to re-distribute information in initial noise for variable-scaled generation. Our method is plug-and-play for DMs and extensive experiments demonstrate the effectiveness in variable-scaled image generation.

[373] Mamba-CNN: A Hybrid Architecture for Efficient and Accurate Facial Beauty Prediction

Djamel Eddine Boukhari

Main category: cs.CV

TL;DR: Mamba-CNN is a novel hybrid architecture combining CNNs with Mamba-inspired SSM gating to efficiently model global facial context for attractiveness assessment, achieving state-of-the-art results on SCUT-FBP5500 benchmark.

DetailsMotivation: Address the trade-off between CNNs (efficient but limited receptive fields) and ViTs (global context but quadratic computational cost) for facial attractiveness assessment, a challenging subjective regression task.

Method: Integrates lightweight Mamba-inspired State Space Model gating mechanism into hierarchical convolutional backbone to dynamically modulate feature maps and emphasize salient facial features with long-range spatial relationships.

Result: Achieves new state-of-the-art on SCUT-FBP5500: Pearson Correlation of 0.9187, MAE of 0.2022, and RMSE of 0.2610.

Conclusion: Validates synergistic potential of combining CNNs with selective SSMs, presenting a powerful new architectural paradigm for nuanced visual understanding tasks.

Abstract: The computational assessment of facial attractiveness, a challenging subjective regression task, is dominated by architectures with a critical trade-off: Convolutional Neural Networks (CNNs) offer efficiency but have limited receptive fields, while Vision Transformers (ViTs) model global context at a quadratic computational cost. To address this, we propose Mamba-CNN, a novel and efficient hybrid architecture. Mamba-CNN integrates a lightweight, Mamba-inspired State Space Model (SSM) gating mechanism into a hierarchical convolutional backbone. This core innovation allows the network to dynamically modulate feature maps and selectively emphasize salient facial features and their long-range spatial relationships, mirroring human holistic perception while maintaining computational efficiency. We conducted extensive experiments on the widely-used SCUT-FBP5500 benchmark, where our model sets a new state-of-the-art. Mamba-CNN achieves a Pearson Correlation (PC) of 0.9187, a Mean Absolute Error (MAE) of 0.2022, and a Root Mean Square Error (RMSE) of 0.2610. Our findings validate the synergistic potential of combining CNNs with selective SSMs and present a powerful new architectural paradigm for nuanced visual understanding tasks.

[374] Traces of Image Memorability in Vision Encoders: Activations, Attention Distributions and Autoencoder Losses

Ece Takmaz, Albert Gatt, Jakub Dotlacil

Main category: cs.CV

TL;DR: This paper explores how pretrained vision encoders’ internal features (latent activations, attention distributions, patch uniformity) correlate with image memorability, finding that sparse autoencoder loss on ViT representations outperforms previous CNN-based methods.

DetailsMotivation: To understand what makes images memorable to humans by examining correlates of memorability in pretrained vision encoders, building on findings from cognitive science and computer vision.

Method: Analyzed latent activations, attention distributions, and patch uniformity in pretrained vision encoders. Explored sparse autoencoder loss over vision transformer representations as a proxy for memorability.

Result: Found that these features correlate with memorability to some extent. Sparse autoencoder loss on ViT representations outperformed past methods using CNN representations.

Conclusion: Model-internal features are informative predictors of image memorability, shedding light on the relationship between these features and what makes images memorable to humans.

Abstract: Images vary in how memorable they are to humans. Inspired by findings from cognitive science and computer vision, this paper explores the correlates of image memorability in pretrained vision encoders, focusing on latent activations, attention distributions, and the uniformity of image patches. We find that these features correlate with memorability to some extent. Additionally, we explore sparse autoencoder loss over the representations of vision transformers as a proxy for memorability, which yields results outperforming past methods using convolutional neural network representations. Our results shed light on the relationship between model-internal features and memorability. They show that some features are informative predictors of what makes images memorable to humans.

[375] Im2Haircut: Single-view Strand-based Hair Reconstruction for Human Avatars

Vanessa Sklyarova, Egor Zakharov, Malte Prinzler, Giorgio Becherini, Michael J. Black, Justus Thies

Main category: cs.CV

TL;DR: Novel 3D hair reconstruction method combining synthetic data for internal geometry and real data for outer structure, using transformer-based prior and Gaussian splatting for single/multi-image reconstruction.

DetailsMotivation: Existing methods struggle with reconstructing complete 3D hair from single images due to hair complexity and lack of ground truth data. Synthetic data alone is limited in quantity and quality, requiring manual artist work.

Method: Transformer-based prior model trained on synthetic data for internal geometry knowledge, combined with real data learning for outer structure. Gaussian-splatting-based reconstruction from one or more images.

Result: Qualitative and quantitative comparisons show superior performance in capturing detailed hair orientation, overall silhouette, and backside consistency compared to existing methods.

Conclusion: The hybrid approach using both synthetic and real data effectively models complete 3D hairstyles from single photographs, overcoming limitations of previous methods.

Abstract: We present a novel approach for 3D hair reconstruction from single photographs based on a global hair prior combined with local optimization. Capturing strand-based hair geometry from single photographs is challenging due to the variety and geometric complexity of hairstyles and the lack of ground truth training data. Classical reconstruction methods like multi-view stereo only reconstruct the visible hair strands, missing the inner structure of hairstyles and hampering realistic hair simulation. To address this, existing methods leverage hairstyle priors trained on synthetic data. Such data, however, is limited in both quantity and quality since it requires manual work from skilled artists to model the 3D hairstyles and create near-photorealistic renderings. To address this, we propose a novel approach that uses both, real and synthetic data to learn an effective hairstyle prior. Specifically, we train a transformer-based prior model on synthetic data to obtain knowledge of the internal hairstyle geometry and introduce real data in the learning process to model the outer structure. This training scheme is able to model the visible hair strands depicted in an input image, while preserving the general 3D structure of hairstyles. We exploit this prior to create a Gaussian-splatting-based reconstruction method that creates hairstyles from one or more images. Qualitative and quantitative comparisons with existing reconstruction pipelines demonstrate the effectiveness and superior performance of our method for capturing detailed hair orientation, overall silhouette, and backside consistency. For additional results and code, please refer to https://im2haircut.is.tue.mpg.de.

[376] PointSlice: Accurate and Efficient Slice-Based Representation for 3D Object Detection from Point Clouds

Liu Qifeng, Zhao Dawei, Dong Yabo, Xiao Liang, Wang Juan, Min Chen, Li Fuyang, Jiang Weizhong, Lu Dongming, Nie Yiming

Main category: cs.CV

TL;DR: PointSlice is a novel 3D object detection method that slices point clouds into 2D horizontal planes and uses a Slice Interaction Network to maintain vertical relationships, achieving both high speed and accuracy.

DetailsMotivation: Current voxel-based methods offer high accuracy but slow inference, while pillar-based methods are faster but less accurate. There's a need for a method that balances both speed and accuracy in 3D point cloud processing for autonomous driving.

Method: Proposes PointSlice which converts 3D point clouds into multiple 2D data slices along horizontal planes, and introduces a Slice Interaction Network (SIN) to maintain vertical relationships across slices while learning only 2D data distributions.

Result: Achieves 1.13x faster speed and 0.79x fewer parameters than state-of-the-art voxel methods on Waymo with only 1.2 mAPH accuracy reduction. Achieves 66.74 mAP on nuScenes (SOTA) and 1.10x faster speed with 0.66x fewer parameters on Argoverse 2.

Conclusion: PointSlice effectively balances speed and accuracy by treating 3D point clouds as batches of 2D data with vertical interaction, making it suitable for real-time autonomous driving applications.

Abstract: 3D object detection from point clouds plays a critical role in autonomous driving. Currently, the primary methods for point cloud processing are voxel-based and pillarbased approaches. Voxel-based methods offer high accuracy through fine-grained spatial segmentation but suffer from slower inference speeds. Pillar-based methods enhance inference speed but still fall short of voxel-based methods in accuracy. To address these issues, we propose a novel point cloud processing method, PointSlice, which slices point clouds along the horizontal plane and includes a dedicated detection network. The main contributions of PointSlice are: (1) A new point cloud processing technique that converts 3D point clouds into multiple sets of 2D (x-y) data slices. The model only learns 2D data distributions, treating the 3D point cloud as separate batches of 2D data, which reduces the number of model parameters and enhances inference speed; (2) The introduction of a Slice Interaction Network (SIN). To maintain vertical relationships across slices, we incorporate SIN into the 2D backbone network, which improves the model’s 3D object perception capability. Extensive experiments demonstrate that PointSlice achieves high detection accuracy and inference speed. On the Waymo dataset, PointSlice is 1.13x faster and has 0.79x fewer parameters than the state-of-the-art voxel-based method (SAFDNet), with only a 1.2 mAPH accuracy reduction. On the nuScenes dataset, we achieve a state-of-the-art detection result of 66.74 mAP. On the Argoverse 2 dataset, PointSlice is 1.10x faster, with 0.66x fewer parameters and a 1.0 mAP accuracy reduction. The code will be available at https://github.com/qifeng22/PointSlice2.

[377] A Continuous-Time Consistency Model for 3D Point Cloud Generation

Sebastian Eilermann, René Heesch, Oliver Niggemann

Main category: cs.CV

TL;DR: ConTiCoM-3D is a continuous-time consistency model for fast 3D shape generation from point clouds without diffusion steps or latent encodings, achieving state-of-the-art results with efficient 1-2 step inference.

DetailsMotivation: Fast and accurate 3D shape generation from point clouds is essential for robotics, AR/VR, and digital content creation, but existing methods rely on iterative denoising or latent decoders which are computationally expensive.

Method: Integrates TrigFlow-inspired continuous noise schedule with Chamfer Distance-based geometric loss, uses time-conditioned neural network operating in continuous time without discretized diffusion steps or pre-trained teacher models.

Result: Matches or outperforms state-of-the-art diffusion and latent consistency models on ShapeNet benchmark in both quality and efficiency, with high geometric fidelity.

Conclusion: ConTiCoM-3D establishes a practical framework for scalable 3D shape generation with efficient one- to two-step inference while maintaining high quality.

Abstract: Fast and accurate 3D shape generation from point clouds is essential for applications in robotics, AR/VR, and digital content creation. We introduce ConTiCoM-3D, a continuous-time consistency model that synthesizes 3D shapes directly in point space, without discretized diffusion steps, pre-trained teacher models, or latent-space encodings. The method integrates a TrigFlow-inspired continuous noise schedule with a Chamfer Distance-based geometric loss, enabling stable training on high-dimensional point sets while avoiding expensive Jacobian-vector products. This design supports efficient one- to two-step inference with high geometric fidelity. In contrast to previous approaches that rely on iterative denoising or latent decoders, ConTiCoM-3D employs a time-conditioned neural network operating entirely in continuous time, thereby achieving fast generation. Experiments on the ShapeNet benchmark show that ConTiCoM-3D matches or outperforms state-of-the-art diffusion and latent consistency models in both quality and efficiency, establishing it as a practical framework for scalable 3D shape generation.

[378] Reinforced Visual Perception with Tools

Zetong Zhou, Dongping Chen, Zixian Ma, Zhihan Hu, Mingyang Fu, Sinan Wang, Yao Wan, Zhou Zhao, Ranjay Krishna

Main category: cs.CV

TL;DR: ReVPT enhances multi-modal LLMs’ visual reasoning through reinforcement learning with visual tools, achieving SOTA performance on perception-heavy benchmarks.

DetailsMotivation: Address limitations of supervised finetuning for visual reasoning (expensive data generation, poor generalization) by using RL to train models to effectively use visual tools.

Method: Propose ReVPT with novel RL algorithm based on GRPO to train models to reason with a suite of four visual tools through reinforcement learning.

Result: Achieves state-of-the-art performance on SAT, CV-Bench, BLINK and MMStar benchmarks. ReVPT-3B/7B outperform instruct models by 9.03%/9.44% on CV-Bench.

Conclusion: RL-based approach effectively enhances visual tool usage and reasoning capabilities, providing new insights for visual reasoning systems.

Abstract: Visual reasoning, a cornerstone of human intelligence, encompasses complex perceptual and logical processes essential for solving diverse visual problems. While advances in computer vision have produced powerful models for various perceptual tasks, leveraging these for general visual reasoning remains challenging. Prior work demonstrates that augmenting LLMs with vision models via supervised finetuning improves performance, but faces key limitations such as expensive data generation, reliance on careful data filtering, and poor generalization. To address these issues, we propose ReVPT to enhance multi-modal LLMs’ abilities to reason about and use visual tools through reinforcement learning. We introduce a novel RL algorithm based on GRPO, designed to train models to reason with a suite of four visual tools. Through extensive experiments, we show that our method achieves state-of-the-art performance on several perception-heavy benchmarks, including SAT, CV-Bench, BLINK and MMStar, significantly outperforming the supervised and text-based RL finetuning baselines. Notably, Our ReVPT-3B and ReVPT-7B outperform the instruct models by 9.03% and 9.44% on CV-Bench. Finally, we bring to the community new insights on RL-based visual tool-usage through extensive ablations. Our code is available at https://github.com/ls-kelvin/REVPT.

[379] MSA2-Net: Utilizing Self-Adaptive Convolution Module to Extract Multi-Scale Information in Medical Image Segmentation

Chao Deng, Xiaosen Li, Xiao Qin

Main category: cs.CV

TL;DR: MSA2-Net introduces a Self-Adaptive Convolution Module that dynamically adjusts kernel sizes based on dataset characteristics, improving medical image segmentation performance across multiple datasets.

DetailsMotivation: nnUNet automatically tunes most hyperparameters but overlooks internal network hyperparameters, limiting model generalization. This study addresses this limitation by enabling dynamic adaptation to different datasets.

Method: A Self-Adaptive Convolution Module that adjusts convolution kernel sizes based on dataset fingerprints. Integrated into Multi-Scale Convolution Bridge and Multi-Scale Amalgamation Decoder of MSA2-Net to capture both global and local features while eliminating redundant data.

Result: Achieved Dice coefficients of 86.49% (Synapse), 92.56% (ACDC), 93.37% (Kvasir), and 92.98% (ISIC2017), demonstrating robust performance across diverse medical image segmentation datasets.

Conclusion: MSA2-Net with Self-Adaptive Convolution Module shows exceptional precision and robustness in medical image segmentation, effectively addressing nnUNet’s limitations in internal hyperparameter tuning.

Abstract: The nnUNet segmentation framework adeptly adjusts most hyperparameters in training scripts automatically, but it overlooks the tuning of internal hyperparameters within the segmentation network itself, which constrains the model’s ability to generalize. Addressing this limitation, this study presents a novel Self-Adaptive Convolution Module that dynamically adjusts the size of the convolution kernels depending on the unique fingerprints of different datasets. This adjustment enables the MSA2-Net, when equipped with this module, to proficiently capture both global and local features within the feature maps. Self-Adaptive Convolution Module is strategically integrated into two key components of the MSA2-Net: the Multi-Scale Convolution Bridge and the Multi-Scale Amalgamation Decoder. In the MSConvBridge, the module enhances the ability to refine outputs from various stages of the CSWin Transformer during the skip connections, effectively eliminating redundant data that could potentially impair the decoder’s performance. Simultaneously, the MSADecoder, utilizing the module, excels in capturing detailed information of organs varying in size during the decoding phase. This capability ensures that the decoder’s output closely reproduces the intricate details within the feature maps, thus yielding highly accurate segmentation images. MSA2-Net, bolstered by this advanced architecture, has demonstrated exceptional performance, achieving Dice coefficient scores of 86.49%, 92.56%, 93.37%, and 92.98% on the Synapse, ACDC, Kvasir, and Skin Lesion Segmentation (ISIC2017) datasets, respectively. This underscores MSA2-Net’s robustness and precision in medical image segmentation tasks across various datasets.

[380] RSCC: A Large-Scale Remote Sensing Change Caption Dataset for Disaster Events

Zhenyuan Chen, Chenxi Wang, Ningyu Zhang, Feng Zhang

Main category: cs.CV

TL;DR: The RSCC dataset provides 62,315 pre-/post-disaster image pairs with detailed captions to enable vision-language models for disaster monitoring and temporal change analysis in remote sensing.

DetailsMotivation: Existing remote sensing datasets lack temporal image pairs and detailed textual annotations, making it difficult to capture dynamic disaster impacts over time and train robust vision-language models for disaster monitoring.

Method: Created the Remote Sensing Change Caption (RSCC) dataset - a large-scale benchmark with 62,315 pre-/post-disaster image pairs covering earthquakes, floods, wildfires and other disasters, paired with rich human-like change captions.

Result: RSCC enables robust training and evaluation of vision-language models for disaster-aware bi-temporal understanding, facilitating detailed disaster-related analysis.

Conclusion: The dataset bridges the temporal and semantic divide in remote sensing data, paving the way for more accurate, interpretable, and scalable vision-language applications in remote sensing for disaster monitoring.

Abstract: Remote sensing is critical for disaster monitoring, yet existing datasets lack temporal image pairs and detailed textual annotations. While single-snapshot imagery dominates current resources, it fails to capture dynamic disaster impacts over time. To address this gap, we introduce the Remote Sensing Change Caption (RSCC) dataset, a large-scale benchmark comprising 62,315 pre-/post-disaster image pairs (spanning earthquakes, floods, wildfires, and more) paired with rich, human-like change captions. By bridging the temporal and semantic divide in remote sensing data, RSCC enables robust training and evaluation of vision-language models for disaster-aware bi-temporal understanding. Our results highlight RSCC’s ability to facilitate detailed disaster-related analysis, paving the way for more accurate, interpretable, and scalable vision-language applications in remote sensing. Code and dataset are available at https://github.com/Bili-Sakura/RSCC.

[381] Variation-aware Vision Token Dropping for Faster Large Vision-Language Models

Junjie Chen, Xuyang Liu, Zichen Wen, Yiyu Wang, Siteng Huang, Honggang Chen

Main category: cs.CV

TL;DR: V²Drop is a token compression method that removes visual tokens with minimal variation during LVLM inference, maintaining high performance while significantly reducing latency and memory usage.

DetailsMotivation: Large vision-language models face efficiency issues with high-resolution images and long videos due to substantial token counts, and existing compression methods suffer from positional bias and incompatibility with efficient operators.

Method: Proposes Variation-aware Vision Token Dropping (V²Drop) which progressively removes visual tokens with minimal variation during inference, based on the observation that visual token variations exhibit task-agnostic properties.

Result: Maintains 94.0% of original performance for image understanding and 98.6% for video understanding while reducing LLM generation latency by 31.5% and 74.2% respectively, with additional GPU memory reduction when combined with efficient operators.

Conclusion: V²Drop provides an effective token compression solution that addresses limitations of existing methods and significantly improves computational efficiency for LVLMs while preserving performance.

Abstract: Large vision-language models (LVLMs) have demonstrated remarkable capabilities in multimodal understanding tasks. However, the increasing demand for high-resolution image and long-video understanding results in substantial token counts, leading to reduced inference efficiency. Token compression offers a direct solution by reducing the number of tokens to be processed, thereby improving computational efficiency. Through extensive analysis, we identify two critical limitations in existing inner-LLM token compression methods: positional bias and incompatibility with efficient operators, which hinder their practical deployment for LVLM acceleration. This paper presents the first approach from a token variation perspective, revealing that visual token variations within LLMs exhibit task-agnostic properties. We propose Variation-aware Vision Token Dropping (\textit{i.e.}, \textbf{V$^2$Drop}), which progressively removes visual tokens with minimal variation during LVLM inference, thereby enhancing computational efficiency. Extensive experiments across multiple models and benchmarks demonstrate that our V$^2$Drop is able to maintain \textbf{94.0%} and \textbf{98.6%} of the original model performance for image and video understanding tasks respectively, while reducing LLM generation latency by \textbf{31.5%} and \textbf{74.2%}. When combined with efficient operators, V$^2$Drop further reduces GPU peak memory usage.

[382] Unified Supervision For Vision-Language Modeling in 3D Computed Tomography

Hao-Chih Lee, Zelong Liu, Hamza Ahmed, Spencer Kim, Sean Huver, Vishwesh Nath, Zahi A. Fayad, Timothy Deyer, Xueyan Mei

Main category: cs.CV

TL;DR: Uniferum is a volumetric vision-language model that unifies diverse supervision signals from classification labels and segmentation masks, achieving state-of-the-art performance on CT-RATE benchmark with 7% AUROC improvement and demonstrating robust out-of-distribution generalization.

DetailsMotivation: General-purpose VLMs lack discriminative precision for reliable clinical use in radiology, compounded by scarcity and heterogeneity of publicly available volumetric CT datasets with varying annotation formats and granularity.

Method: Uniferum harmonizes three public 3D CT datasets with distinct annotations by unifying classification labels and segmentation masks into a single training framework, integrating heterogeneous annotations and body segmentation.

Result: Achieves 7% AUROC improvement on CT-RATE benchmark compared to CLIP-based and conventional multi-label convolutional models, with robust out-of-distribution generalization and unexpected zero-shot performance on RAD-CHEST and INSPECT datasets.

Conclusion: The approach effectively integrates heterogeneous annotations and body segmentation to enhance model performance, setting a new direction for clinically reliable, data-efficient VLMs in 3D medical imaging.

Abstract: General-purpose vision-language models (VLMs) have emerged as promising tools in radiology, offering zero-shot capabilities that mitigate the need for large labeled datasets. However, in high-stakes domains like diagnostic radiology, these models often lack the discriminative precision required for reliable clinical use. This challenge is compounded by the scarcity and heterogeneity of publicly available volumetric CT datasets, which vary widely in annotation formats and granularity. To address these limitations, we introduce Uniferum, a volumetric VLM that unifies diverse supervision signals, encoded in classification labels and segmentation masks, into a single training framework. By harmonizing three public 3D CT datasets with distinct annotations, Uniferum achieves state-of-the-art performance, improving AUROC on the CT-RATE benchmark by 7% compared to CLIP-based and conventional multi-label convolutional models. The model demonstrates robust out-of-distribution generalization, with observed evidence of unexpected zero-shot performance on the RAD-CHEST and INSPECT datasets. Our results highlight the effectiveness of integrating heterogeneous annotations and body segmentation to enhance model performance, setting a new direction for clinically reliable, data-efficient VLMs in 3D medical imaging.

[383] Acoustic Interference Suppression in Ultrasound images for Real-Time HIFU Monitoring Using an Image-Based Latent Diffusion Model

Dejia Cai, Yao Ran, Kun Yang, Xinwang Shi, Yingying Zhou, Kexian Wu, Yang Xu, Yi Hu, Xiaowei Zhou

Main category: cs.CV

TL;DR: HIFU-ILDiff is a deep learning approach using latent diffusion models to remove HIFU interference from ultrasound images in real-time, significantly outperforming traditional Notch Filter methods.

DetailsMotivation: HIFU treatments require real-time monitoring, but ultrasound guidance is hindered by HIFU-induced interference, affecting treatment success and safety.

Method: Uses Vector Quantized Variational Autoencoder (VQ-VAE) to encode noisy images into latent space, then applies latent diffusion model to iteratively remove interference, followed by decoding to reconstruct clean images.

Result: Achieves SSIM of 0.796 and PSNR of 23.780 (vs 0.443 and 14.420 for Notch Filter) with real-time processing at 15 FPS (vs 5 seconds per frame for Notch Filter).

Conclusion: HIFU-ILDiff enables effective real-time denoising of HIFU interference, improving treatment precision for clinical HIFU applications.

Abstract: High-Intensity Focused Ultrasound (HIFU) is a non-invasive therapeutic technique widely used for treating various diseases. However, the success and safety of HIFU treatments depend on real-time monitoring, which is often hindered by interference when using ultrasound to guide HIFU treatment. To address these challenges, we developed HIFU-ILDiff, a novel deep learning-based approach leveraging latent diffusion models to suppress HIFU-induced interference in ultrasound images. The HIFU-ILDiff model employs a Vector Quantized Variational Autoencoder (VQ-VAE) to encode noisy ultrasound images into a lower-dimensional latent space, followed by a latent diffusion model that iteratively removes interference. The denoised latent vectors are then decoded to reconstruct high-resolution, interference-free ultrasound images. We constructed a comprehensive dataset comprising 18,872 image pairs from in vitro phantoms, ex vivo tissues, and in vivo animal data across multiple imaging modalities and HIFU power levels to train and evaluate the model. Experimental results demonstrate that HIFU-ILDiff significantly outperforms the commonly used Notch Filter method, achieving a Structural Similarity Index (SSIM) of 0.796 and Peak Signal-to-Noise Ratio (PSNR) of 23.780 compared to SSIM of 0.443 and PSNR of 14.420 for the Notch Filter under in vitro scenarios. Additionally, HIFU-ILDiff achieves real-time processing at 15 frames per second, markedly faster than the Notch Filter’s 5 seconds per frame. These findings indicate that HIFU-ILDiff is able to denoise HIFU interference in ultrasound guiding images for real-time monitoring during HIFU therapy, which will greatly improve the treatment precision in current clinical applications.

[384] Kwai Keye-VL 1.5 Technical Report

Biao Yang, Bin Wen, Boyang Ding, Changyi Liu, Chenglong Chu, Chengru Song, Chongling Rao, Chuan Yi, Da Li, Dunju Zang, Fan Yang, Guorui Zhou, Guowang Zhang, Han Shen, Hao Peng, Haojie Ding, Hao Wang, Hengrui Ju, Jiaming Huang, Jiangxia Cao, Jiankang Chen, Jingyun Hua, Kaibing Chen, Kaiyu Jiang, Kaiyu Tang, Kun Gai, Muhao Wei, Qiang Wang, Ruitao Wang, Sen Na, Shengnan Zhang, Siyang Mao, Sui Huang, Tianke Zhang, Tingting Gao, Wei Chen, Wei Yuan, Xiangyu Wu, Xiao Hu, Xingyu Lu, Yi-Fan Zhang, Yiping Yang, Yulong Chen, Zeyi Lu, Zhenhua Wu, Zhixin Ling, Zhuoran Yang, Ziming Li, Di Xu, Haixuan Gao, Hang Li, Jing Wang, Lejian Ren, Qigen Hu, Qianqian Wang, Shiyao Wang, Xinchen Luo, Yan Li, Yuhang Hu, Zixing Zhang

Main category: cs.CV

TL;DR: Keye-VL-1.5 introduces three innovations for video understanding: Slow-Fast video encoding, progressive pre-training to 128K context length, and comprehensive post-training pipeline with reasoning enhancement.

DetailsMotivation: Video understanding remains challenging due to dynamic and information-dense nature of videos, with existing models struggling with spatial resolution vs temporal coverage trade-offs.

Method: 1) Slow-Fast video encoding strategy with dynamic resource allocation based on inter-frame similarity; 2) Progressive 4-stage pre-training extending context from 8K to 128K tokens; 3) Post-training pipeline with 5-step chain-of-thought data, iterative GSPO-based RL, and alignment training.

Result: Significant improvements over existing models on public benchmarks and internal human assessment, particularly excelling in video understanding tasks while maintaining competitive performance on general multimodal benchmarks.

Conclusion: Keye-VL-1.5 successfully addresses fundamental challenges in video comprehension through its three key innovations, demonstrating state-of-the-art performance in video understanding tasks.

Abstract: In recent years, the development of Large Language Models (LLMs) has significantly advanced, extending their capabilities to multimodal tasks through Multimodal Large Language Models (MLLMs). However, video understanding remains a challenging area due to the dynamic and information-dense nature of videos. Existing models struggle with the trade-off between spatial resolution and temporal coverage when processing video content. We present Keye-VL-1.5, which addresses fundamental challenges in video comprehension through three key innovations. First, we introduce a novel Slow-Fast video encoding strategy that dynamically allocates computational resources based on inter-frame similarity, processing key frames with significant visual changes at higher resolution (Slow pathway) while handling relatively static frames with increased temporal coverage at lower resolution (Fast pathway). Second, we implement a progressive four-stage pre-training methodology that systematically extends the model’s context length from 8K to 128K tokens, enabling processing of longer videos and more complex visual content. Third, we develop a comprehensive post-training pipeline focusing on reasoning enhancement and human preference alignment, incorporating a 5-step chain-of-thought data construction process, iterative GSPO-based reinforcement learning with progressive prompt hinting for difficult cases, and alignment training. Through extensive evaluation on public benchmarks and rigorous internal human assessment, Keye-VL-1.5 demonstrates significant improvements over existing models, particularly excelling in video understanding tasks while maintaining competitive performance on general multimodal benchmarks.

[385] Understanding Space Is Rocket Science - Only Top Reasoning Models Can Solve Spatial Understanding Tasks

Nils Hoehing, Mayug Maniparambil, Ellen Rushe, Noel E. O’Connor, Anthony Ventresque

Main category: cs.CV

TL;DR: RocketScience is a new benchmark that tests VLMs’ spatial relation understanding using real-world image-text pairs, showing current VLMs struggle significantly with spatial reasoning while humans perform well.

DetailsMotivation: Current vision-language models lack proper spatial relation understanding capabilities, and there's a need for a benchmark that specifically tests this ability using real-world scenarios that are easy for humans but challenging for machines.

Method: Created a new open-source benchmark with real-world image-text pairs focusing on relative spatial relationships and object ordering. Used contrastive evaluation and performed disentanglement analysis to separate object localization from spatial reasoning capabilities.

Result: VLMs show striking lack of spatial relation understanding, with reasoning models performing surprisingly well. Performance is bottlenecked by spatial reasoning rather than object localization capabilities.

Conclusion: Spatial reasoning remains a significant challenge for current VLMs, and the RocketScience benchmark effectively exposes this weakness while providing a valuable tool for future model development and evaluation.

Abstract: We propose RocketScience, an open-source contrastive VLM benchmark that tests for spatial relation understanding. It is comprised of entirely new real-world image-text pairs covering mostly relative spatial understanding and the order of objects. The benchmark is designed to be very easy for humans and hard for the current generation of VLMs, and this is empirically verified. Our results show a striking lack of spatial relation understanding in open source and frontier commercial VLMs and a surprisingly high performance of reasoning models. Additionally, we perform a disentanglement analysis to separate the contributions of object localization and spatial reasoning in chain-of-thought-based models and find that the performance on the benchmark is bottlenecked by spatial reasoning and not object localization capabilities. We release the dataset with a CC-BY-4.0 license and make the evaluation code available at: https://github.com/nilshoehing/rocketscience

[386] ViSTA-SLAM: Visual SLAM with Symmetric Two-view Association

Ganlin Zhang, Shenhan Qian, Xi Wang, Daniel Cremers

Main category: cs.CV

TL;DR: ViSTA-SLAM is a real-time monocular visual SLAM system that works without camera intrinsics, using a lightweight two-view association model and Sim(3) pose graph for superior performance.

DetailsMotivation: To create a broadly applicable visual SLAM system that doesn't require camera calibration parameters (intrinsics) and can work across diverse camera setups with reduced complexity.

Method: Uses a lightweight symmetric two-view association (STA) model as frontend to estimate relative camera poses and regress local pointmaps from two RGB images. Backend employs a specially designed Sim(3) pose graph with loop closures to handle accumulated drift.

Result: Achieves superior performance in both camera tracking and dense 3D reconstruction quality compared to current methods, with frontend size reduced to 35% of comparable state-of-the-art methods.

Conclusion: ViSTA-SLAM demonstrates that a lightweight, calibration-free approach can outperform existing methods while being more broadly applicable across different camera configurations.

Abstract: We present ViSTA-SLAM as a real-time monocular visual SLAM system that operates without requiring camera intrinsics, making it broadly applicable across diverse camera setups. At its core, the system employs a lightweight symmetric two-view association (STA) model as the frontend, which simultaneously estimates relative camera poses and regresses local pointmaps from only two RGB images. This design reduces model complexity significantly, the size of our frontend is only 35% that of comparable state-of-the-art methods, while enhancing the quality of two-view constraints used in the pipeline. In the backend, we construct a specially designed Sim(3) pose graph that incorporates loop closures to address accumulated drift. Extensive experiments demonstrate that our approach achieves superior performance in both camera tracking and dense 3D reconstruction quality compared to current methods. Github repository: https://github.com/zhangganlin/vista-slam

[387] O-DisCo-Edit: Object Distortion Control for Unified Realistic Video Editing

Yuqing Chen, Junjie Wang, Lin Liu, Ruihang Chu, Xiaopeng Zhang, Qi Tian, Yujiu Yang

Main category: cs.CV

TL;DR: O-DisCo-Edit is a unified video editing framework that uses object distortion control signals based on random and adaptive noise to handle diverse editing tasks with a single representation, achieving state-of-the-art performance across various video editing applications.

DetailsMotivation: Current video editing methods require different control signals for different editing tasks, which complicates model design and demands significant training resources. There's a need for a unified framework that can handle diverse editing tasks efficiently.

Method: Proposes O-DisCo-Edit framework with novel object distortion control (O-DisCo) signal based on random and adaptive noise, paired with a “copy-form” preservation module to maintain non-edited regions. Uses an effective training paradigm for efficient, high-fidelity editing.

Result: Extensive experiments and comprehensive human evaluations show that O-DisCo-Edit surpasses both specialized and multitask state-of-the-art methods across various video editing tasks.

Conclusion: O-DisCo-Edit provides a unified solution for controllable video editing that outperforms existing methods while simplifying the model design and reducing training resource requirements through its flexible object distortion control representation.

Abstract: Diffusion models have recently advanced video editing, yet controllable editing remains challenging due to the need for precise manipulation of diverse object properties. Current methods require different control signal for diverse editing tasks, which complicates model design and demands significant training resources. To address this, we propose O-DisCo-Edit, a unified framework that incorporates a novel object distortion control (O-DisCo). This signal, based on random and adaptive noise, flexibly encapsulates a wide range of editing cues within a single representation. Paired with a “copy-form” preservation module for preserving non-edited regions, O-DisCo-Edit enables efficient, high-fidelity editing through an effective training paradigm. Extensive experiments and comprehensive human evaluations consistently demonstrate that O-DisCo-Edit surpasses both specialized and multitask state-of-the-art methods across various video editing tasks. https://cyqii.github.io/O-DisCo-Edit.github.io/

[388] TransForSeg: A Multitask Stereo ViT for Joint Stereo Segmentation and 3D Force Estimation in Catheterization

Pedram Fekri, Mehrdad Zadeh, Javad Dargahi

Main category: cs.CV

TL;DR: A novel stereo Vision Transformer model for simultaneous catheter segmentation from dual X-ray angles and 3D force estimation, outperforming existing methods.

DetailsMotivation: To improve catheterization procedures by enhancing tactile and visual perception through a more effective end-to-end architecture that captures long-range dependencies in X-ray images without gradually expanding receptive fields.

Method: Proposes an encoder-decoder Vision Transformer that processes two input X-ray images as separate sequences of patches. Uses shared segmentation heads for both encoder and decoder embeddings, and a regression head with fused decoder information for 3D force estimation.

Result: The model outperforms state-of-the-art pure segmentation models, vision-based catheter force estimation methods, and multitask approaches. Sets new state-of-the-art in both catheter segmentation and force estimation across various noise levels in synthetic X-ray images.

Conclusion: The stereo Vision Transformer approach effectively captures long-range dependencies and demonstrates superior performance for simultaneous catheter segmentation and 3D force estimation in medical imaging applications.

Abstract: Recently, the emergence of multitask deep learning models has enhanced catheterization procedures by providing tactile and visual perception data through an end-to-end architec- ture. This information is derived from a segmentation and force estimation head, which localizes the catheter in X-ray images and estimates the applied pressure based on its deflection within the image. These stereo vision architectures incorporate a CNN- based encoder-decoder that captures the dependencies between X-ray images from two viewpoints, enabling simultaneous 3D force estimation and stereo segmentation of the catheter. With these tasks in mind, this work approaches the problem from a new perspective. We propose a novel encoder-decoder Vision Transformer model that processes two input X-ray images as separate sequences. Given sequences of X-ray patches from two perspectives, the transformer captures long-range dependencies without the need to gradually expand the receptive field for either image. The embeddings generated by both the encoder and decoder are fed into two shared segmentation heads, while a regression head employs the fused information from the decoder for 3D force estimation. The proposed model is a stereo Vision Transformer capable of simultaneously segmenting the catheter from two angles while estimating the generated forces at its tip in 3D. This model has undergone extensive experiments on synthetic X-ray images with various noise levels and has been compared against state-of-the-art pure segmentation models, vision-based catheter force estimation methods, and a multitask catheter segmentation and force estimation approach. It outperforms existing models, setting a new state-of-the-art in both catheter segmentation and force estimation.

[389] Improving Large Vision and Language Models by Learning from a Panel of Peers

Jefferson Hernandez, Jing Shi, Simon Jenni, Vicente Ordonez, Kushal Kafle

Main category: cs.CV

TL;DR: Panel-of-Peers learning framework uses multiple LVLMs to evaluate and refine each other’s outputs through iterative peer review, achieving significant performance improvements without human-labeled data.

DetailsMotivation: Traditional LVLM alignment methods rely on costly human-curated preference data, limited-quality machine-generated data, or hallucination-prone self-supervised data, creating scalability and quality challenges.

Method: A collaborative learning framework where a panel of LVLMs generate, assess, and refine outputs through iterative peer review, simulating a classroom learning environment with curated prompts.

Result: Significant performance improvement across multiple benchmarks, with average score increasing from 48% to 57% on fifteen benchmarks.

Conclusion: Panel-of-Peers provides a scalable alternative to self-supervised alignment that enhances model performance without requiring extensive human-labeled datasets.

Abstract: Traditional alignment methods for Large Vision and Language Models (LVLMs) primarily rely on human-curated preference data. Human-generated preference data is costly; machine-generated preference data is limited in quality; and self-supervised preference data often introduces hallucinations. To overcome these limitations, we propose a novel Panel-of-Peers learning framework inspired by collaborative learning among humans. This approach leverages a panel of LVLMs, each evaluating and learning from their collective outputs through an iterative self-improvement process. By simulating a peer review system, our models generate, assess, and refine outputs in response to a curated set of prompts, mimicking a classroom learning environment. We demonstrate that this methodology enhances model performance without requiring extensive human-labeled datasets. Our experiments show significant improvement across multiple benchmarks, demonstrating the potential of peer evaluations as a scalable alternative to self-supervised alignment. Notably, we show that Panel-of-Peers increases the average score on fifteen benchmarks from 48% to 57%

[390] Q-Sched: Pushing the Boundaries of Few-Step Diffusion Models with Quantization-Aware Scheduling

Natalia Frumkin, Diana Marculescu

Main category: cs.CV

TL;DR: Q-Sched is a post-training quantization method that modifies the diffusion scheduler instead of model weights, achieving 4x model size reduction while maintaining full-precision accuracy through quantization-aware trajectory adjustment.

DetailsMotivation: Text-to-image diffusion models are computationally expensive, requiring dozens of forward passes through large transformer backbones. Few-step models reduce cost but still depend on large, uncompressed backbones that are too costly for full-precision inference without datacenter GPUs.

Method: Q-Sched modifies the diffusion model scheduler rather than model weights, adjusting the few-step sampling trajectory. It uses JAQ loss (combining text-image compatibility with image quality metric) to learn quantization-aware pre-conditioning coefficients with reference-free calibration using only a handful of prompts.

Result: Achieves 4x reduction in model size while maintaining full-precision accuracy. Shows 15.5% FID improvement over FP16 4-step Latent Consistency Model and 16.6% improvement over FP16 8-step Phased Consistency Model. Large-scale user study with 80,000+ annotations confirms effectiveness on FLUX.1 and SDXL-Turbo.

Conclusion: Quantization and few-step distillation are complementary for high-fidelity generation. Q-Sched provides an effective post-training quantization approach that avoids full-precision inference during calibration and delivers substantial performance gains.

Abstract: Text-to-image diffusion models are computationally intensive, often requiring dozens of forward passes through large transformer backbones. For instance, Stable Diffusion XL generates high-quality images with 50 evaluations of a 2.6B-parameter model, an expensive process even for a single batch. Few-step diffusion models reduce this cost to 2-8 denoising steps but still depend on large, uncompressed U-Net or diffusion transformer backbones, which are often too costly for full-precision inference without datacenter GPUs. These requirements also limit existing post-training quantization methods that rely on full-precision calibration. We introduce Q-Sched, a new paradigm for post-training quantization that modifies the diffusion model scheduler rather than model weights. By adjusting the few-step sampling trajectory, Q-Sched achieves full-precision accuracy with a 4x reduction in model size. To learn quantization-aware pre-conditioning coefficients, we propose the JAQ loss, which combines text-image compatibility with an image quality metric for fine-grained optimization. JAQ is reference-free and requires only a handful of calibration prompts, avoiding full-precision inference during calibration. Q-Sched delivers substantial gains: a 15.5% FID improvement over the FP16 4-step Latent Consistency Model and a 16.6% improvement over the FP16 8-step Phased Consistency Model, showing that quantization and few-step distillation are complementary for high-fidelity generation. A large-scale user study with more than 80,000 annotations further confirms Q-Sched’s effectiveness on both FLUX.1[schnell] and SDXL-Turbo.

[391] OpenVision 2: A Family of Generative Pretrained Visual Encoders for Multimodal Learning

Yanqing Liu, Xianhang Li, Letian Zhang, Zirui Wang, Zeyu Zheng, Yuyin Zhou, Cihang Xie

Main category: cs.CV

TL;DR: OpenVision 2 simplifies the original architecture by removing text encoder and contrastive loss, using only captioning loss, achieving competitive performance with significantly improved training efficiency.

DetailsMotivation: To enhance training efficiency of vision-language models by simplifying architecture while maintaining performance, following trends from CapPa, AIMv2 and LLaVA.

Method: Removed text encoder and contrastive loss, retained only captioning loss for purely generative training signal. Scaled to over 1 billion parameters.

Result: 1.5x faster training (83h to 57h), 1.8x lower memory usage (24.5GB to 13.8GB), 4x larger batch size (2k to 8k), while maintaining competitive multimodal benchmark performance.

Conclusion: The lightweight generative-only paradigm is compelling for future vision encoder development in multimodal foundation models due to superior training efficiency.

Abstract: This paper provides a simplification on OpenVision’s architecture and loss design for enhancing its training efficiency. Following the prior vision-language pretraining works CapPa and AIMv2, as well as modern multimodal designs like LLaVA, our changes are straightforward: we remove the text encoder (and therefore the contrastive loss), retaining only the captioning loss as a purely generative training signal. We name this new version OpenVision 2. The initial results are promising: despite this simplification, OpenVision 2 competitively matches the original model’s performance on a broad set of multimodal benchmarks while substantially cutting both training time and memory consumption. For example, with ViT-L/14, it reduces training time by about 1.5x (from 83h to 57h), and memory usage by about 1.8x (from 24.5GB to 13.8GB, equivalently allowing the maximum batch size to grow from 2k to 8k). This superior training efficiency also allows us to scale far beyond the largest vision encoder used in OpenVision, reaching more than 1 billion parameters. We hold a strong belief that this lightweight, generative-only paradigm is compelling for future vision encoder development in multimodal foundation models.

[392] GaussianGAN: Real-Time Photorealistic controllable Human Avatars

Mohamed Ilyes Lakhal, Richard Bowden

Main category: cs.CV

TL;DR: GaussianGAN is a real-time animatable avatar method that uses Gaussian splatting densification and novel view segmentation to create photorealistic human avatars with reduced blurring, achieving state-of-the-art results at 79 FPS.

DetailsMotivation: Current neural rendering solutions for human avatars suffer from noticeable blurring, limiting their photorealistic quality and real-time performance.

Method: Proposes Gaussian splatting densification strategy on cylindrical skeletal structures, novel view segmentation module for accurate semantic maps, and UNet generator combining Gaussian features with segmentation for photorealistic rendering.

Result: Achieves 79 FPS rendering speed with state-of-the-art pixel fidelity: 32.94dB on ZJU Mocap dataset and 33.39dB on Thuman4 dataset, outperforming previous methods in visual perception and quality.

Conclusion: GaussianGAN successfully addresses blurring issues in current avatar solutions, providing real-time photorealistic rendering with superior visual quality through innovative Gaussian splatting and segmentation techniques.

Abstract: Photorealistic and controllable human avatars have gained popularity in the research community thanks to rapid advances in neural rendering, providing fast and realistic synthesis tools. However, a limitation of current solutions is the presence of noticeable blurring. To solve this problem, we propose GaussianGAN, an animatable avatar approach developed for photorealistic rendering of people in real-time. We introduce a novel Gaussian splatting densification strategy to build Gaussian points from the surface of cylindrical structures around estimated skeletal limbs. Given the camera calibration, we render an accurate semantic segmentation with our novel view segmentation module. Finally, a UNet generator uses the rendered Gaussian splatting features and the segmentation maps to create photorealistic digital avatars. Our method runs in real-time with a rendering speed of 79 FPS. It outperforms previous methods regarding visual perception and quality, achieving a state-of-the-art results in terms of a pixel fidelity of 32.94db on the ZJU Mocap dataset and 33.39db on the Thuman4 dataset.

[393] Examination of PCA Utilisation for Multilabel Classifier of Multispectral Images

Filip Karpowicz, Wiktor Kępiński, Bartosz Staszyński, Grzegorz Sarwas

Main category: cs.CV

TL;DR: PCA’s effectiveness for multi-label multispectral image classification depends on deep learning architecture and training strategy, with ResNet50 and DINOv2 showing varied results.

DetailsMotivation: High dimensionality of multispectral images and complexity of multi-label classification create processing challenges that need dimensionality reduction solutions.

Method: Used PCA to reduce data to 3 dimensions before feeding into a three-layer classifier, testing with ResNet50 and DINOv2 architectures.

Result: PCA effectiveness varies significantly based on the chosen deep learning architecture and training approach.

Conclusion: Findings open avenues for future research into self-supervised pre-training and alternative dimensionality reduction methods for multispectral image classification.

Abstract: This paper investigates the utility of Principal Component Analysis (PCA) for multi-label classification of multispectral images using ResNet50 and DINOv2, acknowledging the high dimensionality of such data and the associated processing challenges. Multi-label classification, where each image may belong to multiple classes, adds further complexity to feature extraction. Our pipeline includes an optional PCA step that reduces the data to three dimensions before feeding it into a three-layer classifier. The findings demonstrate that the effectiveness of PCA for multi-label multispectral image classification depends strongly on the chosen deep learning architecture and training strategy, opening avenues for future research into self-supervised pre-training and alternative dimensionality reduction approaches.

[394] Deep Learning-Based Rock Particulate Classification Using Attention-Enhanced ConvNeXt

Anthony Amankwah, Chris Aldrich

Main category: cs.CV

TL;DR: Enhanced ConvNeXt model with self-attention and channel attention mechanisms for improved rock size classification accuracy

DetailsMotivation: Accurate rock size classification is crucial for geotechnical engineering, mining, and resource management as it impacts operational efficiency and safety

Method: Proposed CNSCA model based on ConvNeXt architecture augmented with self-attention for long-range spatial dependencies and channel attention for informative feature channels

Result: Model significantly outperforms three strong baselines on rock size classification dataset, demonstrating improved accuracy and robustness

Conclusion: Incorporation of attention mechanisms enhances deep learning models’ capability for fine-grained classification tasks involving natural textures like rocks

Abstract: Accurate classification of rock sizes is a vital component in geotechnical engineering, mining, and resource management, where precise estimation influences operational efficiency and safety. In this paper, we propose an enhanced deep learning model based on the ConvNeXt architecture, augmented with both self-attention and channel attention mechanisms. Building upon the foundation of ConvNext, our proposed model, termed CNSCA, introduces self-attention to capture long-range spatial dependencies and channel attention to emphasize informative feature channels. This hybrid design enables the model to effectively capture both fine-grained local patterns and broader contextual relationships within rock imagery, leading to improved classification accuracy and robustness. We evaluate our model on a rock size classification dataset and compare it against three strong baseline. The results demonstrate that the incorporation of attention mechanisms significantly enhances the models capability for fine-grained classification tasks involving natural textures like rocks.

[395] Clinical Metadata Guided Limited-Angle CT Image Reconstruction

Yu Shi, Shuyi Fan, Changsheng Fang, Shuo Han, Haodong Li, Li Zhou, Bahareh Morovati, Dayang Wang, Hengyong Yu

Main category: cs.CV

TL;DR: A two-stage diffusion framework using clinical metadata to improve limited-angle CT reconstruction, achieving superior image quality with physics-based consistency enforcement.

DetailsMotivation: Limited-angle CT offers better temporal resolution and lower radiation but suffers from severe artifacts due to truncated projections, requiring methods to address this ill-posed reconstruction problem.

Method: Two-stage diffusion framework: 1) Transformer-based diffusion model conditioned on metadata (acquisition parameters, demographics, diagnostic impressions) generates coarse anatomical priors, 2) Refinement stage integrates coarse prior and metadata with physics-based data consistency using ADMM at each sampling step.

Result: Significantly improved reconstruction fidelity under severe angular truncation, outperforming metadata-free baselines in SSIM, PSNR, nMI, and PCC metrics. Different metadata types provide complementary benefits, especially diagnostic and demographic priors.

Conclusion: Clinical metadata plays a dual role in improving both reconstruction quality and efficiency, supporting their integration into future metadata-guided medical imaging frameworks.

Abstract: Limited-angle computed tomography (LACT) offers improved temporal resolution and reduced radiation dose for cardiac imaging, but suffers from severe artifacts due to truncated projections. To address the ill-posedness of LACT reconstruction, we propose a two-stage diffusion framework guided by structured clinical metadata. In the first stage, a transformer-based diffusion model conditioned exclusively on metadata, including acquisition parameters, patient demographics, and diagnostic impressions, generates coarse anatomical priors from noise. The second stage further refines the images by integrating both the coarse prior and metadata to produce high-fidelity results. Physics-based data consistency is enforced at each sampling step in both stages using an Alternating Direction Method of Multipliers module, ensuring alignment with the measured projections. Extensive experiments on both synthetic and real cardiac CT datasets demonstrate that incorporating metadata significantly improves reconstruction fidelity, particularly under severe angular truncation. Compared to existing metadata-free baselines, our method achieves superior performance in SSIM, PSNR, nMI, and PCC. Ablation studies confirm that different types of metadata contribute complementary benefits, particularly diagnostic and demographic priors under limited-angle conditions. These findings highlight the dual role of clinical metadata in improving both reconstruction quality and efficiency, supporting their integration into future metadata-guided medical imaging frameworks.

[396] TransMatch: A Transfer-Learning Framework for Defect Detection in Laser Powder Bed Fusion Additive Manufacturing

Mohsen Asghari Ilani, Yaser Mike Banad

Main category: cs.CV

TL;DR: TransMatch is a new framework combining transfer learning and semi-supervised few-shot learning to detect surface defects in LPBF additive manufacturing with high accuracy using limited labeled data.

DetailsMotivation: Surface defects in Laser Powder Bed Fusion pose significant risks to structural integrity, but there's a scarcity of labeled defect data for training detection models.

Method: TransMatch merges transfer learning and semi-supervised few-shot learning to leverage both labeled and unlabeled novel-class images, overcoming limitations of previous meta-learning approaches.

Result: Achieved 98.91% accuracy with minimal loss, along with high precision, recall, and F1-scores for multiple defect classes including cracks, pinholes, holes, and spatter on a dataset of 8,284 images.

Conclusion: TransMatch represents a significant advancement in additive manufacturing defect detection, offering a practical and scalable solution for quality assurance across industrial applications.

Abstract: Surface defects in Laser Powder Bed Fusion (LPBF) pose significant risks to the structural integrity of additively manufactured components. This paper introduces TransMatch, a novel framework that merges transfer learning and semi-supervised few-shot learning to address the scarcity of labeled AM defect data. By effectively leveraging both labeled and unlabeled novel-class images, TransMatch circumvents the limitations of previous meta-learning approaches. Experimental evaluations on a Surface Defects dataset of 8,284 images demonstrate the efficacy of TransMatch, achieving 98.91% accuracy with minimal loss, alongside high precision, recall, and F1-scores for multiple defect classes. These findings underscore its robustness in accurately identifying diverse defects, such as cracks, pinholes, holes, and spatter. TransMatch thus represents a significant leap forward in additive manufacturing defect detection, offering a practical and scalable solution for quality assurance and reliability across a wide range of industrial applications.

[397] Mixture of Balanced Information Bottlenecks for Long-Tailed Visual Recognition

Yifan Lan, Xin Cai, Jun Cheng, Shan Tan

Main category: cs.CV

TL;DR: Proposed Balanced Information Bottleneck (BIB) and Mixture of BIB (MBIB) methods that integrate loss re-balancing and self-distillation into information bottleneck framework for long-tailed visual recognition, achieving state-of-the-art performance.

DetailsMotivation: Real-world visual recognition data is usually long-tailed, creating challenges for DNN training and deployment. Information bottleneck is an effective representation learning approach that needs adaptation for imbalanced data.

Method: BIB integrates loss function re-balancing and self-distillation techniques into original IB network. MBIB extends this with multiple BIBs combining knowledge from different network layers, enabling end-to-end simultaneous representation and classification learning.

Result: Experiments on CIFAR100-LT, ImageNet-LT, and iNaturalist 2018 show both BIB and MBIB achieve state-of-the-art performance for long-tailed visual recognition.

Conclusion: The proposed BIB and MBIB approaches effectively address long-tailed recognition by preserving essential label-related information through information bottleneck framework enhanced with re-balancing and multi-layer knowledge integration.

Abstract: Deep neural networks (DNNs) have achieved significant success in various applications with large-scale and balanced data. However, data in real-world visual recognition are usually long-tailed, bringing challenges to efficient training and deployment of DNNs. Information bottleneck (IB) is an elegant approach for representation learning. In this paper, we propose a balanced information bottleneck (BIB) approach, in which loss function re-balancing and self-distillation techniques are integrated into the original IB network. BIB is thus capable of learning a sufficient representation with essential label-related information fully preserved for long-tailed visual recognition. To further enhance the representation learning capability, we also propose a novel structure of mixture of multiple balanced information bottlenecks (MBIB), where different BIBs are responsible for combining knowledge from different network layers. MBIB facilitates an end-to-end learning strategy that trains representation and classification simultaneously from an information theory perspective. We conduct experiments on commonly used long-tailed datasets, including CIFAR100-LT, ImageNet-LT, and iNaturalist 2018. Both BIB and MBIB reach state-of-the-art performance for long-tailed visual recognition.

[398] PractiLight: Practical Light Control Using Foundational Diffusion Models

Yotam Erel, Rishabh Dabral, Vladislav Golyanik, Amit H. Bermano, Christian Theobalt

Main category: cs.CV

TL;DR: PractiLight is a practical approach for controlling lighting in generated images by leveraging foundational generative models’ understanding. It uses a lightweight LoRA regressor to produce irradiance maps and incorporates desired lighting through Classifier Guidance, achieving state-of-the-art performance with minimal training data.

DetailsMotivation: Existing approaches for light control in generated images require extensive domain-specific datasets, limiting generalization. The authors aim to develop a practical method that leverages foundational knowledge from recent generative models without extensive retraining.

Method: The approach trains a lightweight LoRA regressor to produce direct irradiance maps from a small set of training images. Key insight is that lighting relationships are similar to token interactions in self-attention layers. Uses Classifier Guidance to incorporate desired lighting into image generation.

Result: State-of-the-art performance in quality and control with proven parameter and data efficiency across diverse scenes and image domains. The method generalizes well to various conditions without extensive retraining.

Conclusion: Image lighting can be effectively controlled by tapping into foundational knowledge of generative models, enabling practical and general relighting applications with minimal training requirements.

Abstract: Light control in generated images is a difficult task, posing specific challenges, spanning over the entire image and frequency spectrum. Most approaches tackle this problem by training on extensive yet domain-specific datasets, limiting the inherent generalization and applicability of the foundational backbones used. Instead, PractiLight is a practical approach, effectively leveraging foundational understanding of recent generative models for the task. Our key insight is that lighting relationships in an image are similar in nature to token interaction in self-attention layers, and hence are best represented there. Based on this and other analyses regarding the importance of early diffusion iterations, PractiLight trains a lightweight LoRA regressor to produce the direct irradiance map for a given image, using a small set of training images. We then employ this regressor to incorporate the desired lighting into the generation process of another image using Classifier Guidance. This careful design generalizes well to diverse conditions and image domains. We demonstrate state-of-the-art performance in terms of quality and control with proven parameter and data efficiency compared to leading works over a wide variety of scenes types. We hope this work affirms that image lighting can feasibly be controlled by tapping into foundational knowledge, enabling practical and general relighting.

[399] Latent Gene Diffusion for Spatial Transcriptomics Completion

Paula Cárdenas, Leonardo Manrique, Daniela Vega, Daniela Ruiz, Pablo Arbeláez

Main category: cs.CV

TL;DR: LGDiST is a reference-free latent gene diffusion model that addresses data dropout in spatial transcriptomics by using context genes to build a biologically meaningful latent space, achieving 18% lower MSE than previous methods.

DetailsMotivation: Current models for predicting spatially resolved gene expression from histopathology images suffer from data dropout limitations and dependence on single-cell RNA sequencing references, which introduces alignment issues, batch effects, and inherited dropout problems.

Method: LGDiST uses a diffusion model approach with context genes previously considered uninformative to construct a rich genetic latent space. The model incorporates neighbor conditioning and operates without external references.

Result: LGDiST outperforms previous state-of-the-art methods with 18% lower average MSE across 26 datasets. It improves gene expression prediction performance by up to 10% MSE on six state-of-the-art methods. Ablation studies show that removing key components (context genes, ST latent space, neighbor conditioning) leads to significant performance drops.

Conclusion: LGDiST represents a significant advancement in spatial transcriptomics data completion, demonstrating that reference-free approaches using context genes and diffusion modeling can effectively address data dropout while providing biologically meaningful representations.

Abstract: Computer Vision has proven to be a powerful tool for analyzing Spatial Transcriptomics (ST) data. However, current models that predict spatially resolved gene expression from histopathology images suffer from significant limitations due to data dropout. Most existing approaches rely on single-cell RNA sequencing references, making them dependent on alignment quality and external datasets while also risking batch effects and inherited dropout. In this paper, we address these limitations by introducing LGDiST, the first reference-free latent gene diffusion model for ST data dropout. We show that LGDiST outperforms the previous state-of-the-art in gene expression completion, with an average Mean Squared Error that is 18% lower across 26 datasets. Furthermore, we demonstrate that completing ST data with LGDiST improves gene expression prediction performance on six state-of-the-art methods up to 10% in MSE. A key innovation of LGDiST is using context genes previously considered uninformative to build a rich and biologically meaningful genetic latent space. Our experiments show that removing key components of LGDiST, such as the context genes, the ST latent space, and the neighbor conditioning, leads to considerable drops in performance. These findings underscore that the full architecture of LGDiST achieves substantially better performance than any of its isolated components.

[400] Enabling Federated Object Detection for Connected Autonomous Vehicles: A Deployment-Oriented Evaluation

Komala Subramanyam Cherukuri, Kewei Sha, Zhenhua Huang

Main category: cs.CV

TL;DR: This paper presents the first comprehensive evaluation of federated learning for object detection in connected autonomous vehicles, analyzing performance, resource usage, and environmental robustness across multiple state-of-the-art detectors and datasets.

DetailsMotivation: Centralized training for object detection in CAVs lacks scalability, adaptability, and privacy, while federated learning offers collaborative and privacy-preserving training but faces deployment challenges due to computational demands and diverse operating conditions.

Method: The study evaluates YOLOv5, YOLOv8, YOLOv11, and Deformable DETR detectors on KITTI, BDD100K, and nuScenes datasets, analyzing trade-offs between detection accuracy, computational cost, and resource usage under various conditions including different resolutions, batch sizes, weather/lighting, and dynamic client participation.

Result: The paper provides a holistic deployment-oriented evaluation that integrates model performance, system-level resource profiling, and environmental robustness for FL-based object detection in CAVs.

Conclusion: This work paves the way for robust federated learning deployment in connected autonomous vehicles by addressing critical deployment factors including data heterogeneity, constrained hardware, and environmental variability through systematic evaluation.

Abstract: Object detection is crucial for Connected Autonomous Vehicles (CAVs) to perceive their surroundings and make safe driving decisions. Centralized training of object detection models often achieves promising accuracy, fast convergence, and simplified training process, but it falls short in scalability, adaptability, and privacy-preservation. Federated learning (FL), by contrast, enables collaborative, privacy-preserving, and continuous training across naturally distributed CAV fleets. However, deploying FL in real-world CAVs remains challenging due to the substantial computational demands of training and inference, coupled with highly diverse operating conditions. Practical deployment must address three critical factors: (i) heterogeneity from non-IID data distributions, (ii) constrained onboard computing hardware, and (iii) environmental variability such as lighting and weather, alongside systematic evaluation to ensure reliable performance. This work introduces the first holistic deployment-oriented evaluation of FL-based object detection in CAVs, integrating model performance, system-level resource profiling, and environmental robustness. Using state-of-the-art detectors, YOLOv5, YOLOv8, YOLOv11, and Deformable DETR, evaluated on the KITTI, BDD100K, and nuScenes datasets, we analyze trade-offs between detection accuracy, computational cost, and resource usage under diverse resolutions, batch sizes, weather and lighting conditions, and dynamic client participation, paving the way for robust FL deployment in CAVs.

[401] Doctoral Thesis: Geometric Deep Learning For Camera Pose Prediction, Registration, Depth Estimation, and 3D Reconstruction

Xueyang Kang

Main category: cs.CV

TL;DR: This dissertation develops geometric deep learning methods that integrate traditional geometric priors with deep learning to overcome challenges in 3D vision tasks like camera pose estimation, point cloud registration, and 3D reconstruction.

DetailsMotivation: Address limitations of current 3D deep learning approaches that struggle with high-dimensional data scarcity and traditional geometric methods that fail in unstructured environments with ambiguous features.

Method: Integration of geometric priors and constraints (depth information, surface normals, equivariance) into deep learning models to create geometry-aware architectures for 3D vision tasks.

Result: Enhanced accuracy and robustness in geometric representations for camera pose estimation, point cloud registration, depth prediction, and 3D reconstruction.

Conclusion: Combining geometric techniques with deep learning capabilities produces effective geometry-aware models suitable for real-world applications like cultural heritage preservation and VR/AR environments.

Abstract: Modern deep learning developments create new opportunities for 3D mapping technology, scene reconstruction pipelines, and virtual reality development. Despite advances in 3D deep learning technology, direct training of deep learning models on 3D data faces challenges due to the high dimensionality inherent in 3D data and the scarcity of labeled datasets. Structure-from-motion (SfM) and Simultaneous Localization and Mapping (SLAM) exhibit robust performance when applied to structured indoor environments but often struggle with ambiguous features in unstructured environments. These techniques often struggle to generate detailed geometric representations effective for downstream tasks such as rendering and semantic analysis. Current limitations require the development of 3D representation methods that combine traditional geometric techniques with deep learning capabilities to generate robust geometry-aware deep learning models. The dissertation provides solutions to the fundamental challenges in 3D vision by developing geometric deep learning methods tailored for essential tasks such as camera pose estimation, point cloud registration, depth prediction, and 3D reconstruction. The integration of geometric priors or constraints, such as including depth information, surface normals, and equivariance into deep learning models, enhances both the accuracy and robustness of geometric representations. This study systematically investigates key components of 3D vision, including camera pose estimation, point cloud registration, depth estimation, and high-fidelity 3D reconstruction, demonstrating their effectiveness across real-world applications such as digital cultural heritage preservation and immersive VR/AR environments.

[402] HydroVision: Predicting Optically Active Parameters in Surface Water Using Computer Vision

Shubham Laxmikant Deshmukh, Matthew Wilchek, Feras A. Batarseh

Main category: cs.CV

TL;DR: HydroVision is a deep learning framework that uses standard RGB images to estimate multiple water quality parameters, achieving high accuracy with DenseNet121 architecture as the best performer.

DetailsMotivation: To provide a scalable, cost-effective alternative to traditional multispectral/hyperspectral remote sensing for water quality monitoring, enabling early contamination detection and supporting regulatory agencies during environmental stressors.

Method: Developed a scene classification framework using transfer learning with five state-of-the-art CNN architectures (VGG-16, ResNet50, MobileNetV2, DenseNet121) and Vision Transformer, trained on 500,000+ seasonally varied RGB images from USGS.

Result: DenseNet121 achieved the highest validation performance with R2 score of 0.89 for predicting CDOM, demonstrating strong capability for real-world water quality monitoring across diverse conditions.

Conclusion: The framework shows promise for practical water quality monitoring but requires future improvements for low-light and obstructed scenarios to expand operational utility.

Abstract: Ongoing advancements in computer vision, particularly in pattern recognition and scene classification, have enabled new applications in environmental monitoring. Deep learning now offers non-contact methods for assessing water quality and detecting contamination, both critical for disaster response and public health protection. This work introduces HydroVision, a deep learning-based scene classification framework that estimates optically active water quality parameters including Chlorophyll-Alpha, Chlorophylls, Colored Dissolved Organic Matter (CDOM), Phycocyanins, Suspended Sediments, and Turbidity from standard Red-Green-Blue (RGB) images of surface water. HydroVision supports early detection of contamination trends and strengthens monitoring by regulatory agencies during external environmental stressors, industrial activities, and force majeure events. The model is trained on more than 500,000 seasonally varied images collected from the United States Geological Survey Hydrologic Imagery Visualization and Information System between 2022 and 2024. This approach leverages widely available RGB imagery as a scalable, cost-effective alternative to traditional multispectral and hyperspectral remote sensing. Four state-of-the-art convolutional neural networks (VGG-16, ResNet50, MobileNetV2, DenseNet121) and a Vision Transformer are evaluated through transfer learning to identify the best-performing architecture. DenseNet121 achieves the highest validation performance, with an R2 score of 0.89 in predicting CDOM, demonstrating the framework’s promise for real-world water quality monitoring across diverse conditions. While the current model is optimized for well-lit imagery, future work will focus on improving robustness under low-light and obstructed scenarios to expand its operational utility.

[403] Automated Wildfire Damage Assessment from Multi view Ground level Imagery Via Vision Language Models

Miguel Esparza, Archit Gupta, Ali Mostafavi, Kai Yin, Yiming Xiao

Main category: cs.CV

TL;DR: Zero-shot framework using vision language models for wildfire damage assessment from ground-level imagery, showing multi-view analysis dramatically improves classification accuracy over single-view approaches.

DetailsMotivation: Traditional damage assessment methods are time-consuming and modern computer vision approaches require extensive labeled datasets, hindering immediate post-disaster deployment. Need for rapid, accurate property damage assessment after wildfires.

Method: Proposed two pipelines: VLM only (Pipeline A) and VLM + large language model (Pipeline B) with structured prompts based on wildfire damage indicators. Applied to 2025 Eaton and Palisades fires in California using multi-view image analysis.

Result: Single-view assessments performed poorly (F1 scores 0.225-0.511) while multi-view analysis showed dramatic improvements (F1 scores 0.857-0.947). McNemar test confirmed multi-view yields statistically significant improvements. No significant difference between Pipeline A and B.

Conclusion: VLMs effectively synthesize multi-perspective information for nuanced damage identification. Framework provides immediately deployable, flexible workflow without supervised training, accelerating disaster response triage and prioritization.

Abstract: The escalating intensity and frequency of wildfires demand innovative computational methods for rapid and accurate property damage assessment. Traditional methods are often time consuming, while modern computer vision approaches typically require extensive labeled datasets, hindering immediate post-disaster deployment. This research introduces a novel, zero-shot framework leveraging pre-trained vision language models (VLMs) to classify damage from ground-level imagery. We propose and evaluate two pipelines applied to the 2025 Eaton and Palisades fires in California, a VLM (Pipeline A) and a VLM + large language model (LLM) approach (Pipeline B), that integrate structured prompts based on specific wildfire damage indicators. A primary scientific contribution of this study is demonstrating the VLMs efficacy in synthesizing information from multiple perspectives to identify nuanced damage, a critical limitation in existing literature. Our findings reveal that while single view assessments struggled to classify affected structures (F1 scores ranging from 0.225 to 0.511), the multi-view analysis yielded dramatic improvements (F1 scores ranging from 0.857 to 0.947). Moreover, the McNemar test confirmed that pipelines with a multi-view image assessment yields statistically significant classification improvements; however, the improvements this research observed between Pipeline A and B were not statistically significant. Thus, future research can explore the potential of LLM prompting in damage assessment. The practical contribution is an immediately deployable, flexible, and interpretable workflow that bypasses the need for supervised training, significantly accelerating triage and prioritization for disaster response practitioners.

[404] DroneSR: Rethinking Few-shot Thermal Image Super-Resolution from Drone-based Perspective

Zhipeng Weng, Xiaopeng Liu, Ce Liu, Xingyuan Guo, Yukai Shi, Liang Lin

Main category: cs.CV

TL;DR: Proposes Gaussian quantization representation learning for diffusion models to reduce overfitting in few-shot drone infrared image super-resolution tasks, with a monitoring mechanism and new benchmark dataset.

DetailsMotivation: Large-scale diffusion models for super-resolution suffer from severe overfitting when trained on few-shot drone-captured infrared data, undermining generalization ability.

Method: Gaussian quantization representation learning method for diffusion models with an overfitting monitoring mechanism during training, tested on a new multi-source drone infrared benchmark dataset.

Result: Outperforms existing super-resolution approaches and significantly mitigates overfitting of large-scale architectures under complex conditions with few training samples.

Conclusion: The proposed method effectively reduces overfitting while maintaining architecture complexity, providing a robust solution for few-shot drone-based infrared image reconstruction.

Abstract: Although large scale models achieve significant improvements in performance, the overfitting challenge still frequently undermines their generalization ability. In super resolution tasks on images, diffusion models as representatives of generative models typically adopt large scale architectures. However, few-shot drone-captured infrared training data frequently induces severe overfitting in large-scale architectures. To address this key challenge, our method proposes a new Gaussian quantization representation learning method oriented to diffusion models that alleviates overfitting and enhances robustness. At the same time, an effective monitoring mechanism tracks large scale architectures during training to detect signs of overfitting. By introducing Gaussian quantization representation learning, our method effectively reduces overfitting while maintaining architecture complexity. On this basis, we construct a multi source drone-based infrared image benchmark dataset for detection and use it to emphasize overfitting issues of large scale architectures in few sample, drone-based diverse drone-based image reconstruction scenarios. To verify the efficacy of the method in mitigating overfitting, experiments are conducted on the constructed benchmark. Experimental results demonstrate that our method outperforms existing super resolution approaches and significantly mitigates overfitting of large scale architectures under complex conditions. The code and DroneSR dataset will be available at: https://github.com/wengzp1/GARLSR.

[405] Towards Interpretable Geo-localization: a Concept-Aware Global Image-GPS Alignment Framework

Furong Jia, Lanxin Liu, Ce Hou, Fan Zhang, Xinyan Liu, Yu Liu

Main category: cs.CV

TL;DR: A novel framework that integrates global geo-localization with concept bottlenecks to enhance both accuracy and interpretability by projecting image and location embeddings onto shared geographic concepts.

DetailsMotivation: Current geo-localization models like GeoCLIP lack interpretability, and existing concept-based methods don't align well with geo-alignment objectives, resulting in poor interpretability and performance.

Method: Proposes a Concept-Aware Alignment Module that jointly projects image and location embeddings onto a shared bank of geographic concepts (e.g., tropical climate, mountain, cathedral) and minimizes concept-level loss for better alignment.

Result: Extensive experiments show the approach surpasses GeoCLIP in geo-localization accuracy and improves performance across diverse geospatial prediction tasks while providing richer semantic insights.

Conclusion: This is the first work to introduce interpretability into geo-localization, achieving both better performance and enhanced understanding of geographic decision-making processes.

Abstract: Worldwide geo-localization involves determining the exact geographic location of images captured globally, typically guided by geographic cues such as climate, landmarks, and architectural styles. Despite advancements in geo-localization models like GeoCLIP, which leverages images and location alignment via contrastive learning for accurate predictions, the interpretability of these models remains insufficiently explored. Current concept-based interpretability methods fail to align effectively with Geo-alignment image-location embedding objectives, resulting in suboptimal interpretability and performance. To address this gap, we propose a novel framework integrating global geo-localization with concept bottlenecks. Our method inserts a Concept-Aware Alignment Module that jointly projects image and location embeddings onto a shared bank of geographic concepts (e.g., tropical climate, mountain, cathedral) and minimizes a concept-level loss, enhancing alignment in a concept-specific subspace and enabling robust interpretability. To our knowledge, this is the first work to introduce interpretability into geo-localization. Extensive experiments demonstrate that our approach surpasses GeoCLIP in geo-localization accuracy and boosts performance across diverse geospatial prediction tasks, revealing richer semantic insights into geographic decision-making processes.

[406] A Diffusion-Based Framework for Configurable and Realistic Multi-Storage Trace Generation

Seohyun Kim, Junyoung Lee, Jongho Park, Jinhyung Koo, Sungjin Lee, Yeseong Kim

Main category: cs.CV

TL;DR: DiTTO is a diffusion-based framework for generating realistic, configurable multi-device storage traces with high fidelity and diversity.

DetailsMotivation: To create synthetic storage traces that accurately capture temporal dynamics and inter-device dependencies while allowing user-defined configurations, addressing the need for realistic and configurable trace data in storage system research.

Method: Leverages advanced diffusion techniques to synthesize high-fidelity continuous traces that maintain temporal patterns and device relationships while following user-specified configurations.

Result: Experimental results show DiTTO generates traces with high fidelity and diversity while closely aligning with guided configurations with only 8% errors.

Conclusion: DiTTO successfully demonstrates that diffusion models can effectively generate realistic and precisely configurable multi-device storage traces, providing a valuable tool for storage system research and development.

Abstract: We propose DiTTO, a novel diffusion-based framework for generating realistic, precisely configurable, and diverse multi-device storage traces. Leveraging advanced diffusion tech- niques, DiTTO enables the synthesis of high-fidelity continuous traces that capture temporal dynamics and inter-device dependencies with user-defined configurations. Our experimental results demonstrate that DiTTO can generate traces with high fidelity and diversity while aligning closely with guided configurations with only 8% errors.

[407] Structure-aware Contrastive Learning for Diagram Understanding of Multimodal Models

Hiroshi Sasaki

Main category: cs.CV

TL;DR: A novel training paradigm using hard samples and specialized loss functions to improve diagram understanding in vision-language models like CLIP, showing significant improvements on flowchart benchmarks.

DetailsMotivation: Standard multimodal models like CLIP struggle with specialized visual domains like diagrams that contain structured, symbolic information different from natural imagery.

Method: Proposed contrastive learning approach with two specialized loss functions that leverage the inherent structural properties of diagrams, using hard samples for training.

Result: Substantial improvements over standard CLIP and conventional hard negative CLIP learning on flowchart benchmarks for both image-text matching and visual question answering tasks.

Conclusion: Tailored training strategies are crucial for specialized tasks, and this approach advances diagrammatic understanding in vision-language integration.

Abstract: Multimodal models, such as the Contrastive Language-Image Pre-training (CLIP) model, have demonstrated remarkable success in aligning visual and linguistic representations. However, these models exhibit limitations when applied to specialised visual domains, such as diagrams, which encode structured, symbolic information distinct from that of natural imagery. In this paper, we introduce a novel training paradigm explicitly designed to enhance the comprehension of diagrammatic images within vision-language models. Our approach uses ``hard’’ samples for our proposed contrastive learning that incorporates two specialised loss functions that leverage the inherent structural properties of diagrams. By integrating these objectives into model training, our method enables models to develop a more structured and semantically coherent understanding of diagrammatic content. We empirically validate our approach on a benchmark dataset of flowcharts, as a representative class of diagrammatic imagery, demonstrating substantial improvements over standard CLIP and conventional hard negative CLIP learning paradigms for both image-text matching and visual question answering tasks. Our findings underscore the significance of tailored training strategies for specialised tasks and contribute to advancing diagrammatic understanding within the broader landscape of vision-language integration.

[408] 2D Gaussian Splatting with Semantic Alignment for Image Inpainting

Hongyu Li, Chaofeng Chen, Xiaoming Li, Guangming Lu

Main category: cs.CV

TL;DR: First image inpainting framework using 2D Gaussian Splatting, achieving competitive results through continuous rendering and DINO-guided semantic consistency.

DetailsMotivation: To explore Gaussian Splatting's untapped potential for image inpainting, which requires both local pixel coherence and global semantic restoration.

Method: Encodes incomplete images into continuous 2D Gaussian splat coefficients, uses differentiable rasterization with patch-wise strategy for efficiency, and incorporates DINO features for semantic guidance.

Result: Achieves competitive performance on standard benchmarks in both quantitative metrics and perceptual quality.

Conclusion: Establishes a new direction for applying Gaussian Splatting to 2D image processing with promising inpainting results.

Abstract: Gaussian Splatting (GS), a recent technique for converting discrete points into continuous spatial representations, has shown promising results in 3D scene modeling and 2D image super-resolution. In this paper, we explore its untapped potential for image inpainting, which demands both locally coherent pixel synthesis and globally consistent semantic restoration. We propose the first image inpainting framework based on 2D Gaussian Splatting, which encodes incomplete images into a continuous field of 2D Gaussian splat coefficients and reconstructs the final image via a differentiable rasterization process. The continuous rendering paradigm of GS inherently promotes pixel-level coherence in the inpainted results. To improve efficiency and scalability, we introduce a patch-wise rasterization strategy that reduces memory overhead and accelerates inference. For global semantic consistency, we incorporate features from a pretrained DINO model. We observe that DINO’s global features are naturally robust to small missing regions and can be effectively adapted to guide semantic alignment in large-mask scenarios, ensuring that the inpainted content remains contextually consistent with the surrounding scene. Extensive experiments on standard benchmarks demonstrate that our method achieves competitive performance in both quantitative metrics and perceptual quality, establishing a new direction for applying Gaussian Splatting to 2D image processing.

[409] Draw-In-Mind: Learning Precise Image Editing via Chain-of-Thought Imagination

Ziyun Zeng, Junhao Zhang, Wei Li, Mike Zheng Shou

Main category: cs.CV

TL;DR: DIM addresses imbalanced responsibilities in multimodal models by creating a dataset that enhances instruction comprehension and provides explicit design blueprints for image editing, resulting in state-of-the-art performance with a modest 4.6B parameter model.

DetailsMotivation: Current unified multimodal models struggle with precise image editing due to imbalanced division of responsibilities - understanding modules act as translators while generation modules must handle both design and painting tasks, despite understanding modules being trained with more data on complex reasoning.

Method: Introduces Draw-In-Mind (DIM) dataset with two subsets: DIM-T2I (14M image-text pairs for instruction comprehension) and DIM-Edit (233K chain-of-thought imaginations as design blueprints). Connects frozen Qwen2.5-VL-3B with trainable SANA1.5-1.6B via lightweight MLP, trained on DIM dataset.

Result: DIM-4.6B-Edit achieves SOTA or competitive performance on ImgEdit and GEdit-Bench benchmarks, outperforming much larger models like UniWorld-V1 and Step1X-Edit despite modest parameter scale.

Conclusion: Explicitly assigning design responsibility to the understanding module provides significant benefits for image editing tasks, demonstrating that better task division rather than simply scaling model size can lead to superior performance.

Abstract: In recent years, integrating multimodal understanding and generation into a single unified model has emerged as a promising paradigm. While this approach achieves strong results in text-to-image (T2I) generation, it still struggles with precise image editing. We attribute this limitation to an imbalanced division of responsibilities. The understanding module primarily functions as a translator that encodes user instructions into semantic conditions, while the generation module must simultaneously act as designer and painter, inferring the original layout, identifying the target editing region, and rendering the new content. This imbalance is counterintuitive because the understanding module is typically trained with several times more data on complex reasoning tasks than the generation module. To address this issue, we introduce Draw-In-Mind (DIM), a dataset comprising two complementary subsets: (i) DIM-T2I, containing 14M long-context image-text pairs to enhance complex instruction comprehension; and (ii) DIM-Edit, consisting of 233K chain-of-thought imaginations generated by GPT-4o, serving as explicit design blueprints for image edits. We connect a frozen Qwen2.5-VL-3B with a trainable SANA1.5-1.6B via a lightweight two-layer MLP, and train it on the proposed DIM dataset, resulting in DIM-4.6B-T2I/Edit. Despite its modest parameter scale, DIM-4.6B-Edit achieves SOTA or competitive performance on the ImgEdit and GEdit-Bench benchmarks, outperforming much larger models such as UniWorld-V1 and Step1X-Edit. These findings demonstrate that explicitly assigning the design responsibility to the understanding module provides significant benefits for image editing. Our dataset and models will be available at https://github.com/showlab/DIM.

[410] Ensemble-Based Event Camera Place Recognition Under Varying Illumination

Therese Joseph, Tobias Fischer, Michael Milford

Main category: cs.CV

TL;DR: Ensemble-based event camera place recognition method combining multiple reconstructions, feature extractors, and temporal resolutions achieves 57% improvement in day-night recall.

DetailsMotivation: Event cameras offer high dynamic range and low latency but robust visual place recognition under severe illumination changes remains challenging.

Method: Ensemble approach combining sequence-matched results from multiple event-to-frame reconstructions, VPR feature extractors, and temporal resolutions with broader fusion strategy.

Result: 57% relative improvement in Recall@1 across day-night transitions on long-term driving datasets (8km per traverse) without metric subsampling.

Conclusion: The ensemble fusion strategy significantly improves robustness to varied lighting conditions, and comprehensive analysis identifies critical components for robust performance.

Abstract: Compared to conventional cameras, event cameras provide a high dynamic range and low latency, offering greater robustness to rapid motion and challenging lighting conditions. Although the potential of event cameras for visual place recognition (VPR) has been established, developing robust VPR frameworks under severe illumination changes remains an open research problem. In this paper, we introduce an ensemble-based approach to event camera place recognition that combines sequence-matched results from multiple event-to-frame reconstructions, VPR feature extractors, and temporal resolutions. Unlike previous event-based ensemble methods, which only utilise temporal resolution, our broader fusion strategy delivers significantly improved robustness under varied lighting conditions (e.g., afternoon, sunset, night), achieving a 57% relative improvement in Recall@1 across day-night transitions. We evaluate our approach on two long-term driving datasets (with 8 km per traverse) without metric subsampling, thereby preserving natural variations in speed and stop duration that influence event density. We also conduct a comprehensive analysis of key design choices, including binning strategies, polarity handling, reconstruction methods, and feature extractors, to identify the most critical components for robust performance. Additionally, we propose a modification to the standard sequence matching framework that enhances performance at longer sequence lengths. To facilitate future research, we will release our codebase and benchmarking framework.

[411] MOSAIC: Multi-Subject Personalized Generation via Correspondence-Aware Alignment and Disentanglement

Dong She, Siming Fu, Mushui Liu, Qiaoqiao Jin, Hualiang Wang, Mu Liu, Jidong Jiang

Main category: cs.CV

TL;DR: MOSAIC is a representation-centric framework for multi-subject image generation that addresses identity blending and attribute leakage through semantic correspondence and orthogonal feature disentanglement.

DetailsMotivation: Existing multi-subject generation methods suffer from identity blending and attribute leakage due to inadequate modeling of how different subjects interact within shared representation spaces.

Method: Introduces SemAlign-MS dataset with fine-grained semantic correspondences, semantic correspondence attention loss for precise point-to-point alignment, and multi-reference disentanglement loss to push subjects into orthogonal attention subspaces.

Result: Achieves state-of-the-art performance on multiple benchmarks and maintains high fidelity with 4+ reference subjects, outperforming methods that typically degrade beyond 3 subjects.

Conclusion: MOSAIC enables complex multi-subject synthesis with precise identity preservation and semantic coherence, opening new possibilities for applications requiring multiple reference subjects.

Abstract: Multi-subject personalized generation presents unique challenges in maintaining identity fidelity and semantic coherence when synthesizing images conditioned on multiple reference subjects. Existing methods often suffer from identity blending and attribute leakage due to inadequate modeling of how different subjects should interact within shared representation spaces. We present MOSAIC, a representation-centric framework that rethinks multi-subject generation through explicit semantic correspondence and orthogonal feature disentanglement. Our key insight is that multi-subject generation requires precise semantic alignment at the representation level - knowing exactly which regions in the generated image should attend to which parts of each reference. To enable this, we introduce SemAlign-MS, a meticulously annotated dataset providing fine-grained semantic correspondences between multiple reference subjects and target images, previously unavailable in this domain. Building on this foundation, we propose the semantic correspondence attention loss to enforce precise point-to-point semantic alignment, ensuring high consistency from each reference to its designated regions. Furthermore, we develop the multi-reference disentanglement loss to push different subjects into orthogonal attention subspaces, preventing feature interference while preserving individual identity characteristics. Extensive experiments demonstrate that MOSAIC achieves state-of-the-art performance on multiple benchmarks. Notably, while existing methods typically degrade beyond 3 subjects, MOSAIC maintains high fidelity with 4+ reference subjects, opening new possibilities for complex multi-subject synthesis applications.

[412] Discrete Noise Inversion for Next-scale Autoregressive Text-based Image Editing

Quan Dao, Xiaoxiao He, Ligong Han, Ngan Hoai Nguyen, Amin Heyrani Nobar, Faez Ahmed, Han Zhang, Viet Anh Nguyen, Dimitris Metaxas

Main category: cs.CV

TL;DR: VARIN is the first noise inversion-based editing method for Visual AutoRegressive (VAR) models that enables precise text-guided image editing without additional training, using a novel Location-aware Argmax Inversion technique.

DetailsMotivation: While VAR models show strong text-to-image generation capabilities, their ability to perform prompt-guided image editing without additional training remains unexplored but critical for practical applications.

Method: Proposes VARIN with Location-aware Argmax Inversion (LAI) to generate inverse Gumbel noises, enabling precise image reconstruction and targeted edits aligned with textual prompts.

Result: Extensive experiments show VARIN effectively modifies source images according to text prompts while preserving original background and structural details.

Conclusion: VARIN validates VAR models’ capability for practical image editing applications through noise inversion techniques, offering precise control while maintaining image integrity.

Abstract: Visual autoregressive models (VAR) have recently emerged as a promising class of generative models, achieving performance comparable to diffusion models in text-to-image generation tasks. While conditional generation has been widely explored, the ability to perform prompt-guided image editing without additional training is equally critical, as it supports numerous practical real-world applications. This paper investigates the text-to-image editing capabilities of VAR by introducing Visual AutoRegressive Inverse Noise (VARIN), the first noise inversion-based editing technique designed explicitly for VAR models. VARIN leverages a novel pseudo-inverse function for argmax sampling, named Location-aware Argmax Inversion (LAI), to generate inverse Gumbel noises. These inverse noises enable precise reconstruction of the source image and facilitate targeted, controllable edits aligned with textual prompts. Extensive experiments demonstrate that VARIN effectively modifies source images according to specified prompts while significantly preserving the original background and structural details, thus validating its efficacy as a practical editing approach.

[413] Unsupervised Training of Vision Transformers with Synthetic Negatives

Nikolaos Giakoumoglou, Andreas Floros, Kleanthis Marios Papadopoulos, Tania Stathaki

Main category: cs.CV

TL;DR: Integrating synthetic hard negatives improves vision transformer representation learning, enhancing discriminative power without introducing novel methods.

DetailsMotivation: Addressing the neglected potential of hard negative samples in self-supervised learning for vision transformers, where previous works rarely explored synthetic hard negatives in this context.

Method: Building on existing observations by integrating synthetic hard negative samples into vision transformer representation learning frameworks.

Result: Notable improvement in discriminative power of learned representations, with performance gains demonstrated for both DeiT-S and Swin-T architectures.

Conclusion: A simple yet effective technique that enhances vision transformer performance through strategic use of synthetic hard negative samples in self-supervised learning.

Abstract: This paper does not introduce a novel method per se. Instead, we address the neglected potential of hard negative samples in self-supervised learning. Previous works explored synthetic hard negatives but rarely in the context of vision transformers. We build on this observation and integrate synthetic hard negatives to improve vision transformer representation learning. This simple yet effective technique notably improves the discriminative power of learned representations. Our experiments show performance improvements for both DeiT-S and Swin-T architectures.

[414] Explaining What Machines See: XAI Strategies in Deep Object Detection Models

FatemehSadat Seyedmomeni, Mohammad Ali Keyvanrad

Main category: cs.CV

TL;DR: A comprehensive review of explainable AI methods for object detection models, categorizing techniques and analyzing their application to various architectures like YOLO, SSD, and Faster R-CNN.

DetailsMotivation: Address the black-box nature and complexity of deep neural networks in object detection, particularly for critical applications like autonomous driving and medical imaging where interpretability is crucial.

Method: Categorizes XAI techniques into perturbation-based, gradient-based, backpropagation-based, and graph-based methods. Analyzes specific methods like D-RISE, BODEM, D-CLOSE, and FSOD, and investigates their applicability to various object detection architectures.

Result: Statistical analysis shows accelerating interest in explainable object detection from 2022 to mid-2025. The review provides a structured taxonomy and critical assessment of existing methods.

Conclusion: This review serves as a guide for researchers and practitioners to select suitable explainability techniques and promotes the development of more interpretable AI systems for object detection applications.

Abstract: In recent years, deep learning has achieved unprecedented success in various computer vision tasks, particularly in object detection. However, the black-box nature and high complexity of deep neural networks pose significant challenges for interpretability, especially in critical domains such as autonomous driving, medical imaging, and security systems. Explainable Artificial Intelligence (XAI) aims to address this challenge by providing tools and methods to make model decisions more transparent, interpretable, and trust-worthy for humans. This review provides a comprehensive analysis of state-of-the-art explain-ability methods specifically applied to object detection models. The paper be-gins by categorizing existing XAI techniques based on their underlying mechanisms-perturbation-based, gradient-based, backpropagation-based, and graph-based methods. Notable methods such as D-RISE, BODEM, D-CLOSE, and FSOD are discussed in detail. Furthermore, the paper investigates their applicability to various object detection architectures, including YOLO, SSD, Faster R-CNN, and EfficientDet. Statistical analysis of publication trends from 2022 to mid-2025 shows an accelerating interest in explainable object detection, indicating its increasing importance. The study also explores common datasets and evaluation metrics, and highlights the major challenges associated with model interpretability. By providing a structured taxonomy and a critical assessment of existing methods, this review aims to guide researchers and practitioners in selecting suitable explainability techniques for object detection applications and to foster the development of more interpretable AI systems.

[415] Fake & Square: Training Self-Supervised Vision Transformers with Synthetic Data and Synthetic Hard Negatives

Nikolaos Giakoumoglou, Andreas Floros, Kleanthis Marios Papadopoulos, Tania Stathaki

Main category: cs.CV

TL;DR: Syn2Co framework explores using synthetic data and synthetic hard negatives in self-supervised learning for vision transformers, showing promise but also limitations of synthetic approaches.

DetailsMotivation: To reduce reliance on vast real-world data and carefully curated hard negatives in contrastive self-supervised learning by exploring synthetic alternatives.

Method: Combines two approaches: 1) using generative models to create synthetic data for sample diversity, and 2) generating synthetic hard negatives in representation space. Evaluated on DeiT-S and Swin-T architectures.

Result: The framework demonstrates both promise and limitations of synthetic data in self-supervised learning, providing insights for future synthetic-enhanced training approaches.

Conclusion: Synthetic data augmentation and synthetic hard negative generation show potential for creating more robust and transferable visual representations, though with identified limitations that need addressing in future work.

Abstract: This paper does not introduce a new method per se. Instead, we build on existing self-supervised learning approaches for vision, drawing inspiration from the adage “fake it till you make it”. While contrastive self-supervised learning has achieved remarkable success, it typically relies on vast amounts of real-world data and carefully curated hard negatives. To explore alternatives to these requirements, we investigate two forms of “faking it” in vision transformers. First, we study the potential of generative models for unsupervised representation learning, leveraging synthetic data to augment sample diversity. Second, we examine the feasibility of generating synthetic hard negatives in the representation space, creating diverse and challenging contrasts. Our framework - dubbed Syn2Co - combines both approaches and evaluates whether synthetically enhanced training can lead to more robust and transferable visual representations on DeiT-S and Swin-T architectures. Our findings highlight the promise and limitations of synthetic data in self-supervised learning, offering insights for future work in this direction.

[416] Palette Aligned Image Diffusion

Elad Aharoni, Noy Porat, Dani Lischinski, Ariel Shamir

Main category: cs.CV

TL;DR: Palette-Adapter is a novel method that enables text-to-image diffusion models to generate images conditioned on user-specified color palettes with flexible control over palette adherence and color variation.

DetailsMotivation: Color palettes are widely used in creative workflows but introduce ambiguity and instability when conditioning image generation. Existing methods struggle with maintaining both palette adherence and image quality.

Method: Interprets palettes as sparse histograms with two control parameters (histogram entropy and palette-to-histogram distance), introduces negative histogram mechanism to suppress undesired hues, and trains on a curated dataset with balanced color coverage.

Result: Outperforms existing approaches in achieving both strong palette adherence and high image quality across a wide range of palettes and prompts, as validated through qualitative, quantitative, and user study evaluations.

Conclusion: The Palette-Adapter provides stable, semantically coherent generation with flexible palette control, addressing the challenge of palette conditioning in text-to-image diffusion models.

Abstract: We introduce the Palette-Adapter, a novel method for conditioning text-to-image diffusion models on a user-specified color palette. While palettes are a compact and intuitive tool widely used in creative workflows, they introduce significant ambiguity and instability when used for conditioning image generation. Our approach addresses this challenge by interpreting palettes as sparse histograms and introducing two scalar control parameters: histogram entropy and palette-to-histogram distance, which allow flexible control over the degree of palette adherence and color variation. We further introduce a negative histogram mechanism that allows users to suppress specific undesired hues, improving adherence to the intended palette under the standard classifier-free guidance mechanism. To ensure broad generalization across the color space, we train on a carefully curated dataset with balanced coverage of rare and common colors. Our method enables stable, semantically coherent generation across a wide range of palettes and prompts. We evaluate our method qualitatively, quantitatively, and through a user study, and show that it consistently outperforms existing approaches in achieving both strong palette adherence and high image quality.

[417] Vision-Based Embedded System for Noncontact Monitoring of Preterm Infant Behavior in Low-Resource Care Settings

Stanley Mugisha, Rashid Kisitu, Francis Komakech, Excellence Favor

Main category: cs.CV

TL;DR: A lightweight vision-based system using quantized MobileNet on Raspberry Pi achieves high accuracy (91.8% sleep detection, 97.7% crying detection) for non-invasive preterm infant monitoring in low-resource settings.

DetailsMotivation: Preterm birth is a leading cause of neonatal mortality, especially in low-resource settings where current monitoring methods are invasive, error-prone, and impractical. There's a need for automated, non-invasive monitoring solutions.

Method: Developed an embedded monitoring system using quantized MobileNet model deployed on Raspberry Pi for real-time behavioral state detection. The framework includes model quantization (68% size reduction), Raspberry Pi-optimized vision pipelines, and secure IoT communication for clinical alerts.

Result: Achieved state-of-the-art accuracy: 91.8% for sleep detection and 97.7% for crying/normal classification. The system maintains computational efficiency suitable for edge deployment while larger models like ResNet152 and VGG19 proved computationally prohibitive despite marginal accuracy gains.

Conclusion: Lightweight, optimized models like MobileNet provide the most viable foundation for scalable, low-cost NICU monitoring systems, enabling improved preterm care in resource-constrained environments through non-invasive, automated vision-based monitoring.

Abstract: Preterm birth remains a leading cause of neonatal mortality, disproportionately affecting low-resource settings with limited access to advanced neonatal intensive care units (NICUs).Continuous monitoring of infant behavior, such as sleep/awake states and crying episodes, is critical but relies on manual observation or invasive sensors, which are prone to error, impractical, and can cause skin damage. This paper presents a novel, noninvasive, and automated vision-based framework to address this gap. We introduce an embedded monitoring system that utilizes a quantized MobileNet model deployed on a Raspberry Pi for real-time behavioral state detection. When trained and evaluated on public neonatal image datasets, our system achieves state-of-the-art accuracy (91.8% for sleep detection and 97.7% for crying/normal classification) while maintaining computational efficiency suitable for edge deployment. Through comparative benchmarking, we provide a critical analysis of the trade-offs between model size, inference latency, and diagnostic accuracy. Our findings demonstrate that while larger architectures (e.g., ResNet152, VGG19) offer marginal gains in accuracy, their computational cost is prohibitive for real-time edge use. The proposed framework integrates three key innovations: model quantization for memory-efficient inference (68% reduction in size), Raspberry Pi-optimized vision pipelines, and secure IoT communication for clinical alerts. This work conclusively shows that lightweight, optimized models such as the MobileNet offer the most viable foundation for scalable, low-cost, and clinically actionable NICU monitoring systems, paving the way for improved preterm care in resource-constrained environments.

[418] See No Evil: Adversarial Attacks Against Linguistic-Visual Association in Referring Multi-Object Tracking Systems

Halima Bouzidi, Haoyu Liu, Mohammad Al Faruque

Main category: cs.CV

TL;DR: VEIL is an adversarial framework that exploits vulnerabilities in Referring Multi-Object Tracking (RMOT) systems, causing track ID switches and terminations through carefully crafted digital and physical perturbations.

DetailsMotivation: To examine security implications of RMOT systems from a design-logic perspective, identifying adversarial vulnerabilities that compromise both linguistic-visual referring and track-object matching components.

Method: VEIL framework targets unified referring-matching mechanisms of RMOT models, exploiting FIFO-based memory vulnerabilities through spatial-temporal reasoning attacks that persist in history buffers over multiple frames.

Result: Comprehensive evaluations on Refer-KITTI dataset show VEIL effectively corrupts tracking logic reliability, inducing track ID switches and terminations through both digital and physical perturbations.

Conclusion: The study demonstrates urgent need for security-aware RMOT designs for critical large-scale applications, revealing persistent vulnerabilities in advanced RMOT models with FIFO-based memory.

Abstract: Language-vision understanding has driven the development of advanced perception systems, most notably the emerging paradigm of Referring Multi-Object Tracking (RMOT). By leveraging natural-language queries, RMOT systems can selectively track objects that satisfy a given semantic description, guided through Transformer-based spatial-temporal reasoning modules. End-to-End (E2E) RMOT models further unify feature extraction, temporal memory, and spatial reasoning within a Transformer backbone, enabling long-range spatial-temporal modeling over fused textual-visual representations. Despite these advances, the reliability and robustness of RMOT remain underexplored. In this paper, we examine the security implications of RMOT systems from a design-logic perspective, identifying adversarial vulnerabilities that compromise both the linguistic-visual referring and track-object matching components. Additionally, we uncover a novel vulnerability in advanced RMOT models employing FIFO-based memory, whereby targeted and consistent attacks on their spatial-temporal reasoning introduce errors that persist within the history buffer over multiple subsequent frames. We present VEIL, a novel adversarial framework designed to disrupt the unified referring-matching mechanisms of RMOT models. We show that carefully crafted digital and physical perturbations can corrupt the tracking logic reliability, inducing track ID switches and terminations. We conduct comprehensive evaluations using the Refer-KITTI dataset to validate the effectiveness of VEIL and demonstrate the urgent need for security-aware RMOT designs for critical large-scale applications.

[419] ContextFusion and Bootstrap: An Effective Approach to Improve Slot Attention-Based Object-Centric Learning

Pinzhuo Tian, Shengjie Yang, Hang Yu, Alex C. Kot

Main category: cs.CV

TL;DR: Proposes ContextFusion and Bootstrap Branch to enhance slot attention models by incorporating semantic information and enabling flexible feature adaptation, addressing limitations in current object-centric learning approaches.

DetailsMotivation: Existing slot attention methods lack high-level semantic information and cannot fine-tune encoders, limiting their understanding of object semantics and flexibility in object-centric learning.

Method: Introduces ContextFusion stage to exploit semantic information from foreground/background with auxiliary indicators, and Bootstrap Branch to decouple feature adaptation from reconstruction using bootstrap strategy.

Result: Significantly improves performance of different state-of-the-art slot attention models on both simulated and real-world datasets.

Conclusion: The proposed enhancements effectively address key limitations in slot attention methods, enabling better semantic understanding and more flexible adaptation for improved object-centric learning.

Abstract: A key human ability is to decompose a scene into distinct objects and use their relationships to understand the environment. Object-centric learning aims to mimic this process in an unsupervised manner. Recently, the slot attention-based framework has emerged as a leading approach in this area and has been widely used in various downstream tasks. However, existing slot attention methods face two key limitations: (1) a lack of high-level semantic information. In current methods, image areas are assigned to slots based on low-level features such as color and texture. This makes the model overly sensitive to low-level features and limits its understanding of object contours, shapes, or other semantic characteristics. (2) The inability to fine-tune the encoder. Current methods require a stable feature space throughout training to enable reconstruction from slots, which restricts the flexibility needed for effective object-centric learning. To address these limitations, we propose a novel ContextFusion stage and a Bootstrap Branch, both of which can be seamlessly integrated into existing slot attention models. In the ContextFusion stage, we exploit semantic information from the foreground and background, incorporating an auxiliary indicator that provides additional contextual cues about them to enrich the semantic content beyond low-level features. In the Bootstrap Branch, we decouple feature adaptation from the original reconstruction phase and introduce a bootstrap strategy to train a feature-adaptive mechanism, allowing for more flexible adaptation. Experimental results show that our method significantly improves the performance of different SOTA slot attention models on both simulated and real-world datasets.

[420] SALAD – Semantics-Aware Logical Anomaly Detection

Matic Fučka, Vitjan Zavrtanik, Danijel Skočaj

Main category: cs.CV

TL;DR: SALAD is a semantics-aware discriminative method that significantly improves logical anomaly detection by modeling object composition maps, achieving 96.1% AUROC on MVTec LOCO benchmark.

DetailsMotivation: Current surface anomaly detection methods perform well on structural anomalies but struggle with logical anomalies like missing components. Existing approaches discard spatial and semantic information through aggregated pretrained features or handcrafted descriptors.

Method: Proposes SALAD with a novel composition branch to explicitly model object composition map distributions, learning semantic relationships. Introduces a new procedure for extracting composition maps without hand-made labels or category-specific information.

Result: Achieves state-of-the-art performance on MVTec LOCO benchmark with 96.1% image-level AUROC, significantly outperforming previous methods.

Conclusion: SALAD effectively addresses logical anomaly detection by preserving semantic and spatial information through composition map modeling, demonstrating substantial improvements over existing approaches.

Abstract: Recent surface anomaly detection methods excel at identifying structural anomalies, such as dents and scratches, but struggle with logical anomalies, such as irregular or missing object components. The best-performing logical anomaly detection approaches rely on aggregated pretrained features or handcrafted descriptors (most often derived from composition maps), which discard spatial and semantic information, leading to suboptimal performance. We propose SALAD, a semantics-aware discriminative logical anomaly detection method that incorporates a newly proposed composition branch to explicitly model the distribution of object composition maps, consequently learning important semantic relationships. Additionally, we introduce a novel procedure for extracting composition maps that requires no hand-made labels or category-specific information, in contrast to previous methods. By effectively modelling the composition map distribution, SALAD significantly improves upon state-of-the-art methods on the standard benchmark for logical anomaly detection, MVTec LOCO, achieving an impressive image-level AUROC of 96.1%. Code: https://github.com/MaticFuc/SALAD

[421] A Data-Centric Approach to Pedestrian Attribute Recognition: Synthetic Augmentation via Prompt-driven Diffusion Models

Alejandro Alonso, Sawaiz A. Chaudhry, Juan C. SanMiguel, Álvaro García-Martín, Pablo Ayuso-Albizu, Pablo Carballeira

Main category: cs.CV

TL;DR: A data-centric approach using synthetic data augmentation with diffusion models to improve pedestrian attribute recognition, particularly for underrepresented attributes, without changing model architecture.

DetailsMotivation: Pedestrian Attribute Recognition performance is constrained by training dataset limitations, especially under-representation of certain attributes, requiring better generalization across numerous real-world attributes.

Method: 1) Protocol to identify weakly recognized attributes 2) Prompt-driven pipeline using diffusion models to generate synthetic pedestrian images 3) Strategy to incorporate synthetic samples with prompt-based annotation rules and modified loss function

Result: Approach boosts recognition of underrepresented attributes and improves overall model performance beyond targeted attributes on popular PAR datasets, while strengthening zero-shot generalization.

Conclusion: Efficient and scalable solution that improves pedestrian attribute recognition in real-world scenarios without requiring architectural changes to existing models.

Abstract: Pedestrian Attribute Recognition (PAR) is a challenging task as models are required to generalize across numerous attributes in real-world data. Traditional approaches focus on complex methods, yet recognition performance is often constrained by training dataset limitations, particularly the under-representation of certain attributes. In this paper, we propose a data-centric approach to improve PAR by synthetic data augmentation guided by textual descriptions. First, we define a protocol to identify weakly recognized attributes across multiple datasets. Second, we propose a prompt-driven pipeline that leverages diffusion models to generate synthetic pedestrian images while preserving the consistency of PAR datasets. Finally, we derive a strategy to seamlessly incorporate synthetic samples into training data, which considers prompt-based annotation rules and modifies the loss function. Results on popular PAR datasets demonstrate that our approach not only boosts recognition of underrepresented attributes but also improves overall model performance beyond the targeted attributes. Notably, this approach strengthens zero-shot generalization without requiring architectural changes of the model, presenting an efficient and scalable solution to improve the recognition of attributes of pedestrians in the real world.

[422] NOOUGAT: Towards Unified Online and Offline Multi-Object Tracking

Benjamin Missaoui, Orcun Cetintas, Guillem Brasó, Tim Meinhardt, Laura Leal-Taixé

Main category: cs.CV

TL;DR: NOOUGAT is a unified multi-object tracker that bridges online/offline MOT with flexible temporal processing using GNNs and autoregressive layers.

DetailsMotivation: Current MOT methods are fragmented into online (frame-by-frame) and offline (batch processing) approaches, neither handling flexible temporal requirements or long-term occlusions effectively.

Method: Uses Graph Neural Network framework with non-overlapping subclips and novel Autoregressive Long-term Tracking layer, allowing adjustable subclip size for latency-context tradeoff.

Result: State-of-the-art performance: +2.3 AssA on DanceTrack, +9.2 on SportsMOT, +5.0 on MOT20 in online mode, with even better offline results.

Conclusion: NOOUGAT successfully unifies online and offline tracking with flexible temporal processing, overcoming limitations of both traditional approaches.

Abstract: The long-standing division between \textit{online} and \textit{offline} Multi-Object Tracking (MOT) has led to fragmented solutions that fail to address the flexible temporal requirements of real-world deployment scenarios. Current \textit{online} trackers rely on frame-by-frame hand-crafted association strategies and struggle with long-term occlusions, whereas \textit{offline} approaches can cover larger time gaps, but still rely on heuristic stitching for arbitrarily long sequences. In this paper, we introduce NOOUGAT, the first tracker designed to operate with arbitrary temporal horizons. NOOUGAT leverages a unified Graph Neural Network (GNN) framework that processes non-overlapping subclips, and fuses them through a novel Autoregressive Long-term Tracking (ALT) layer. The subclip size controls the trade-off between latency and temporal context, enabling a wide range of deployment scenarios, from frame-by-frame to batch processing. NOOUGAT achieves state-of-the-art performance across both tracking regimes, improving \textit{online} AssA by +2.3 on DanceTrack, +9.2 on SportsMOT, and +5.0 on MOT20, with even greater gains in \textit{offline} mode.

[423] SegFormer Fine-Tuning with Dropout: Advancing Hair Artifact Removal in Skin Lesion Analysis

Asif Mohammed Saad, Umme Niraj Mahi

Main category: cs.CV

TL;DR: Fine-tuned SegFormer with dropout regularization achieves precise hair mask segmentation in dermoscopic images with high accuracy (Dice ~0.96, IoU ~0.93).

DetailsMotivation: Hair artifacts in dermoscopic images obscure critical diagnostic features and challenge accurate skin lesion analysis, requiring effective preprocessing methods.

Method: SegFormerWithDropout architecture using MiT-B2 encoder pretrained on ImageNet, with dropout probability 0.3 in segmentation head. Trained on 500 dermoscopic images with hair mask annotations using 10-fold cross-validation, AdamW optimizer (lr=0.001), cross-entropy loss, and early stopping.

Result: Robust performance with average Dice coefficient ~0.96, IoU ~0.93, PSNR ~34 dB, SSIM 0.97, and low LPIPS 0.06, demonstrating accurate hair artifact segmentation.

Conclusion: The proposed method effectively segments hair artifacts in dermoscopic images and shows potential to enhance preprocessing for downstream skin cancer detection tasks.

Abstract: Hair artifacts in dermoscopic images present significant challenges for accurate skin lesion analysis, potentially obscuring critical diagnostic features in dermatological assessments. This work introduces a fine-tuned SegFormer model augmented with dropout regularization to achieve precise hair mask segmentation. The proposed SegformerWithDropout architecture leverages the MiT-B2 encoder, pretrained on ImageNet, with an in-channel count of 3 and 2 output classes, incorporating a dropout probability of 0.3 in the segmentation head to prevent overfitting. Training is conducted on a specialized dataset of 500 dermoscopic skin lesion images with fine-grained hair mask annotations, employing 10-fold cross-validation, AdamW optimization with a learning rate of 0.001, and cross-entropy loss. Early stopping is applied based on validation loss, with a patience of 3 epochs and a maximum of 20 epochs per fold. Performance is evaluated using a comprehensive suite of metrics, including Intersection over Union (IoU), Dice coefficient, Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM), and Learned Perceptual Image Patch Similarity (LPIPS). Experimental results from the cross-validation demonstrate robust performance, with average Dice coefficients reaching approximately 0.96 and IoU values of 0.93, alongside favorable PSNR (around 34 dB), SSIM (0.97), and low LPIPS (0.06), highlighting the model’s effectiveness in accurate hair artifact segmentation and its potential to enhance preprocessing for downstream skin cancer detection tasks.

[424] Enhancing Zero-Shot Pedestrian Attribute Recognition with Synthetic Data Generation: A Comparative Study with Image-To-Image Diffusion Models

Pablo Ayuso-Albizu, Juan C. SanMiguel, Pablo Carballeira

Main category: cs.CV

TL;DR: This paper explores using diffusion models to generate synthetic pedestrian images for Pedestrian Attribute Recognition (PAR) tasks, achieving 4.5% performance improvement through optimized text prompts and image properties.

DetailsMotivation: The scarcity of large-scale annotated datasets hinders PAR model generalization, especially in complex scenarios with occlusions, varying poses, and diverse environments. Diffusion models offer potential for generating diverse synthetic images to expand training data.

Method: The study investigates diffusion-based data expansion for PAR, identifying key parameters including text prompts, image properties, and latest diffusion-based augmentation enhancements. The best-performing approach is used to generate synthetic images to enrich zero-shot datasets.

Result: Experimental results show that prompt alignment and image properties are critical factors, with optimal selection leading to a 4.5% improvement in PAR recognition performance.

Conclusion: Diffusion models are effective for generating synthetic pedestrian images tailored to PAR tasks, and careful parameter selection can significantly enhance PAR model robustness and adaptability in real-world scenarios.

Abstract: Pedestrian Attribute Recognition (PAR) involves identifying various human attributes from images with applications in intelligent monitoring systems. The scarcity of large-scale annotated datasets hinders the generalization of PAR models, specially in complex scenarios involving occlusions, varying poses, and diverse environments. Recent advances in diffusion models have shown promise for generating diverse and realistic synthetic images, allowing to expand the size and variability of training data. However, the potential of diffusion-based data expansion for generating PAR-like images remains underexplored. Such expansion may enhance the robustness and adaptability of PAR models in real-world scenarios. This paper investigates the effectiveness of diffusion models in generating synthetic pedestrian images tailored to PAR tasks. We identify key parameters of img2img diffusion-based data expansion; including text prompts, image properties, and the latest enhancements in diffusion-based data augmentation, and examine their impact on the quality of generated images for PAR. Furthermore, we employ the best-performing expansion approach to generate synthetic images for training PAR models, by enriching the zero-shot datasets. Experimental results show that prompt alignment and image properties are critical factors in image generation, with optimal selection leading to a 4.5% improvement in PAR recognition performance.

[425] Omnidirectional Spatial Modeling from Correlated Panoramas

Xinshen Zhang, Tongxi Fu, Xu Zheng

Main category: cs.CV

TL;DR: CFpano is the first benchmark dataset for cross-frame correlated panoramas visual question answering, featuring 2700+ images and 8000+ QA pairs. The authors propose a multi-modal LLM with Group Relative Policy Optimization that achieves state-of-the-art performance (+5.37% overall improvement).

DetailsMotivation: Omnidirectional scene understanding is crucial for applications like embodied AI and autonomous driving, but existing methods only handle single frames, neglecting cross-frame correlated panoramas which are essential for holistic 360° scene comprehension.

Method: Introduced CFpano benchmark dataset with 2700+ images and 8000+ QA pairs. Developed a multi-modal LLM fine-tuned with Group Relative Policy Optimization (GRPO) and tailored reward functions for robust reasoning with cross-frame correlated panoramas.

Result: The proposed method achieves state-of-the-art performance, outperforming strong baselines by +5.37% in overall performance across both multiple-choice and open-ended VQA tasks. Experimental results demonstrate effectiveness across all major reasoning categories.

Conclusion: CFpano establishes a new benchmark for panoramic scene understanding. The GRPO-based multi-modal LLM effectively addresses cross-frame correlated panorama reasoning, validating the approach’s effectiveness and setting a new standard for 360° scene comprehension.

Abstract: Omnidirectional scene understanding is vital for various downstream applications, such as embodied AI, autonomous driving, and immersive environments, yet remains challenging due to geometric distortion and complex spatial relations in 360{\deg} imagery. Existing omnidirectional methods achieve scene understanding within a single frame while neglecting cross-frame correlated panoramas. To bridge this gap, we introduce \textbf{CFpano}, the \textbf{first} benchmark dataset dedicated to cross-frame correlated panoramas visual question answering in the holistic 360{\deg} scenes. CFpano consists of over 2700 images together with over 8000 question-answer pairs, and the question types include both multiple choice and open-ended VQA. Building upon our CFpano, we further present \methodname, a multi-modal large language model (MLLM) fine-tuned with Group Relative Policy Optimization (GRPO) and a set of tailored reward functions for robust and consistent reasoning with cross-frame correlated panoramas. Benchmark experiments with existing MLLMs are conducted with our CFpano. The experimental results demonstrate that \methodname achieves state-of-the-art performance across both multiple-choice and open-ended VQA tasks, outperforming strong baselines on all major reasoning categories (\textbf{+5.37%} in overall performance). Our analyses validate the effectiveness of GRPO and establish a new benchmark for panoramic scene understanding.

[426] ADVMEM: Adversarial Memory Initialization for Realistic Test-Time Adaptation via Tracklet-Based Benchmarking

Shyma Alhuwaider, Motasem Alfarra, Juan C. Perez, Merey Ramazanova, Bernard Ghanem

Main category: cs.CV

TL;DR: A novel tracklet-based dataset called ITD is introduced to benchmark test-time adaptation methods, addressing the lack of temporal dependencies in current benchmarks. The dataset reveals limitations of existing TTA methods and proposes an adversarial memory initialization strategy that significantly improves performance.

DetailsMotivation: Current TTA benchmarks fail to represent realistic scenarios with temporal dependencies, such as consecutive video frames showing the same object over time. Real-world applications like hand-held cameras and self-driving cars naturally exhibit these temporal patterns that are missing from existing datasets.

Method: Created the ITD dataset using tracklets (sequences of object-centric images) compiled from bounding boxes of an object-tracking dataset. Conducted experimental analysis of current TTA methods and proposed a novel adversarial memory initialization strategy to improve memory-based TTA approaches.

Result: The ITD dataset successfully captures natural temporal dependencies. Experimental analysis revealed limitations of current TTA methods when faced with temporal challenges. The proposed adversarial memory initialization strategy substantially boosted performance of various methods on the challenging benchmark.

Conclusion: The ITD dataset provides a more realistic benchmark for TTA methods by incorporating temporal dependencies. The proposed adversarial memory initialization effectively addresses the challenges posed by temporal patterns, significantly improving existing TTA method performance.

Abstract: We introduce a novel tracklet-based dataset for benchmarking test-time adaptation (TTA) methods. The aim of this dataset is to mimic the intricate challenges encountered in real-world environments such as images captured by hand-held cameras, self-driving cars, etc. The current benchmarks for TTA focus on how models face distribution shifts, when deployed, and on violations to the customary independent-and-identically-distributed (i.i.d.) assumption in machine learning. Yet, these benchmarks fail to faithfully represent realistic scenarios that naturally display temporal dependencies, such as how consecutive frames from a video stream likely show the same object across time. We address this shortcoming of current datasets by proposing a novel TTA benchmark we call the “Inherent Temporal Dependencies” (ITD) dataset. We ensure the instances in ITD naturally embody temporal dependencies by collecting them from tracklets-sequences of object-centric images we compile from the bounding boxes of an object-tracking dataset. We use ITD to conduct a thorough experimental analysis of current TTA methods, and shed light on the limitations of these methods when faced with the challenges of temporal dependencies. Moreover, we build upon these insights and propose a novel adversarial memory initialization strategy to improve memory-based TTA methods. We find this strategy substantially boosts the performance of various methods on our challenging benchmark.

[427] Palmistry-Informed Feature Extraction and Analysis using Machine Learning

Shweta Patil

Main category: cs.CV

TL;DR: Machine learning pipeline for automated palm feature analysis using computer vision to extract line structures, texture, and shape metrics from palm images.

DetailsMotivation: To move beyond traditional subjective interpretation of palm features by providing a data-driven, quantitative framework for studying correlations between palmar morphology and externally validated traits or conditions.

Method: Computer vision pipeline that extracts key characteristics (principal line structures, texture, shape metrics) from palm images, trained on a novel dataset of annotated palm images using machine learning models.

Result: Demonstrates feasibility for digital anthropometry and personalized user analytics, with machine learning models successfully identifying complex patterns in palm data.

Conclusion: Opens avenues for research intersecting cultural practices with computational analysis, with potential for mobile platform deployment.

Abstract: This paper explores the automated analysis of palmar features using machine learning techniques. We present a computer vision pipeline that extracts key characteristics from palm images, such as principal line structures, texture, and shape metrics. These features are used to train predictive models on a novel dataset curated from annotated palm images. Our approach moves beyond traditional subjective interpretation by providing a data-driven, quantitative framework for studying the correlations between palmar morphology and externally validated traits or conditions. The methodology demonstrates feasibility for applications in digital anthropometry and personalized user analytics, with potential for deployment on mobile platforms. Results indicate that machine learning models can identify complex patterns in palm data, opening avenues for research that intersects cultural practices with computational analysis.

[428] A Multimodal Cross-View Model for Predicting Postoperative Neck Pain in Cervical Spondylosis Patients

Jingyang Shan, Qishuai Yu, Jiacen Liu, Shaolin Zhang, Wen Shen, Yanxiao Zhao, Tianyi Wang, Xiaolin Qin, Yiheng Yin

Main category: cs.CV

TL;DR: Proposes ABPDC module and FPRAN network for multimodal cervical spondylosis analysis, achieving superior neck pain recovery prediction accuracy.

DetailsMotivation: Neck pain mechanisms in cervical spondylosis remain unclear with uncertain treatment outcomes, and multimodal feature fusion faces challenges from imaging differences and spatial mismatches.

Method: Adaptive Bidirectional Pyramid Difference Convolution (ABPDC) module for multimodal integration using difference convolution advantages, and Feature Pyramid Registration Auxiliary Network (FPRAN) to address structural misalignment.

Result: Experiments on MMCSD dataset show superior prediction accuracy of postoperative neck pain recovery compared to existing methods, with ablation studies confirming effectiveness.

Conclusion: The proposed multimodal fusion approach with ABPDC and FPRAN effectively addresses imaging challenges and improves neck pain recovery prediction in cervical spondylosis.

Abstract: Neck pain is the primary symptom of cervical spondylosis, yet its underlying mechanisms remain unclear, leading to uncertain treatment outcomes. To address the challenges of multimodal feature fusion caused by imaging differences and spatial mismatches, this paper proposes an Adaptive Bidirectional Pyramid Difference Convolution (ABPDC) module that facilitates multimodal integration by exploiting the advantages of difference convolution in texture extraction and grayscale invariance, and a Feature Pyramid Registration Auxiliary Network (FPRAN) to mitigate structural misalignment. Experiments on the MMCSD dataset demonstrate that the proposed model achieves superior prediction accuracy of postoperative neck pain recovery compared with existing methods, and ablation studies further confirm its effectiveness.

[429] DSGC-Net: A Dual-Stream Graph Convolutional Network for Crowd Counting via Feature Correlation Mining

Yihong Wu, Jinqiao Wei, Xionghui Zhao, Yidi Li, Shaoyi Du, Bin Ren, Nicu Sebe

Main category: cs.CV

TL;DR: DSGC-Net proposes a dual-stream graph convolutional network with density and representation approximation branches to improve crowd counting accuracy in complex scenarios with varying density distributions and viewpoint changes.

DetailsMotivation: Existing crowd counting models struggle with significant density distribution differences between regions and inconsistency in individual representations caused by viewpoint changes and body posture variations, which limits counting accuracy.

Method: Proposes DSGC-Net with Density Approximation (DA) branch and Representation Approximation (RA) branch. DA branch generates density distribution maps and constructs density-driven semantic graphs. RA branch establishes representation-driven semantic graphs using global representation similarity. Both use graph convolutional networks to model latent semantic relationships.

Result: Achieves MAE of 48.9 on ShanghaiTech Part A and 5.9 on ShanghaiTech Part B datasets, outperforming current state-of-the-art methods across three widely used datasets.

Conclusion: DSGC-Net effectively addresses density variation adaptation and representation inconsistency challenges in crowd counting through dual-stream graph convolutional architecture with feature correlation mining, demonstrating superior performance over existing methods.

Abstract: Deep learning-based crowd counting methods have achieved remarkable progress in recent years. However, in complex crowd scenarios, existing models still face challenges when adapting to significant density distribution differences between regions. Additionally, the inconsistency of individual representations caused by viewpoint changes and body posture differences further limits the counting accuracy of the models. To address these challenges, we propose DSGC-Net, a Dual-Stream Graph Convolutional Network based on feature correlation mining. DSGC-Net introduces a Density Approximation (DA) branch and a Representation Approximation (RA) branch. By modeling two semantic graphs, it captures the potential feature correlations in density variations and representation distributions. The DA branch incorporates a density prediction module that generates the density distribution map, and constructs a density-driven semantic graph based on density similarity. The RA branch establishes a representation-driven semantic graph by computing global representation similarity. Then, graph convolutional networks are applied to the two semantic graphs separately to model the latent semantic relationships, which enhance the model’s ability to adapt to density variations and improve counting accuracy in multi-view and multi-pose scenarios. Extensive experiments on three widely used datasets demonstrate that DSGC-Net outperforms current state-of-the-art methods. In particular, we achieve MAE of 48.9 and 5.9 in ShanghaiTech Part A and Part B datasets, respectively. The released code is available at: https://github.com/Wu-eon/CrowdCounting-DSGCNet.

[430] RS-OOD: A Vision-Language Augmented Framework for Out-of-Distribution Detection in Remote Sensing

Yingrui Ji, Jiansheng Chen, Jingbo Chen, Anzhi Yue, Chenhao Wang, Kai Li, Yao Zhu

Main category: cs.CV

TL;DR: RS-OOD is a novel framework for out-of-distribution detection in remote sensing that uses vision-language modeling with spatial feature enhancement, dual-prompt alignment, and self-training to achieve robust few-shot performance.

DetailsMotivation: Existing OOD detection methods are poorly suited for remote sensing due to data scarcity, complex multi-scale structures, and distribution shifts, creating a critical need for specialized solutions.

Method: Leverages remote sensing-specific vision-language modeling with three innovations: spatial feature enhancement, dual-prompt alignment mechanism for spatial-semantic consistency, and confidence-guided self-training loop for pseudo-label mining.

Result: Consistently outperforms existing methods across multiple remote sensing benchmarks and enables efficient adaptation with minimal labeled data.

Conclusion: Demonstrates the critical value of spatial-semantic integration for robust OOD detection in remote sensing applications.

Abstract: Out-of-distribution (OOD) detection represents a critical challenge in remote sensing applications, where reliable identification of novel or anomalous patterns is essential for autonomous monitoring, disaster response, and environmental assessment. Despite remarkable progress in OOD detection for natural images, existing methods and benchmarks remain poorly suited to remote sensing imagery due to data scarcity, complex multi-scale scene structures, and pronounced distribution shifts. To this end, we propose RS-OOD, a novel framework that leverages remote sensing-specific vision-language modeling to enable robust few-shot OOD detection. Our approach introduces three key innovations: spatial feature enhancement that improved scene discrimination, a dual-prompt alignment mechanism that cross-verifies scene context against fine-grained semantics for spatial-semantic consistency, and a confidence-guided self-training loop that dynamically mines pseudo-labels to expand training data without manual annotation. RS-OOD consistently outperforms existing methods across multiple remote sensing benchmarks and enables efficient adaptation with minimal labeled data, demonstrating the critical value of spatial-semantic integration.

[431] SynthGenNet: a self-supervised approach for test-time generalization using synthetic multi-source domain mixing of street view images

Pushpendra Dhakara, Prachi Chachodhia, Vaibhav Kumar

Main category: cs.CV

TL;DR: SynthGenNet uses self-supervised student-teacher architecture with ClassMix++ and novel loss functions to improve urban scene understanding and bridge sim-to-real domain gaps, achieving 50% mIoU on real datasets.

DetailsMotivation: Unstructured urban environments present complex challenges for scene understanding due to diverse layouts and the difficulty of generalizing from synthetic to real data.

Method: Self-supervised student-teacher architecture with ClassMix++ algorithm for blending synthetic data, Grounded Mask Consistency Loss for cross-domain alignment, and Pseudo-Label Guided Contrastive Learning for domain-invariant features.

Result: Outperforms state-of-the-art single-source methods, achieving 50% mIoU on real-world datasets like Indian Driving Dataset (IDD).

Conclusion: The approach effectively bridges sim-to-real domain gaps, reduces reliance on labeled target data, and improves generalization in complex urban environments.

Abstract: Unstructured urban environments present unique challenges for scene understanding and generalization due to their complex and diverse layouts. We introduce SynthGenNet, a self-supervised student-teacher architecture designed to enable robust test-time domain generalization using synthetic multi-source imagery. Our contributions include the novel ClassMix++ algorithm, which blends labeled data from various synthetic sources while maintaining semantic integrity, enhancing model adaptability. We further employ Grounded Mask Consistency Loss (GMC), which leverages source ground truth to improve cross-domain prediction consistency and feature alignment. The Pseudo-Label Guided Contrastive Learning (PLGCL) mechanism is integrated into the student network to facilitate domain-invariant feature learning through iterative knowledge distillation from the teacher network. This self-supervised strategy improves prediction accuracy, addresses real-world variability, bridges the sim-to-real domain gap, and reliance on labeled target data, even in complex urban areas. Outcomes show our model outperforms the state-of-the-art (relying on single source) by achieving 50% Mean Intersection-Over-Union (mIoU) value on real-world datasets like Indian Driving Dataset (IDD).

[432] Data-Driven Loss Functions for Inference-Time Optimization in Text-to-Image Generation

Sapir Esther Yiflach, Yuval Atzmon, Gal Chechik

Main category: cs.CV

TL;DR: Learn-to-Steer: A framework that learns data-driven objectives from diffusion model’s cross-attention maps to improve spatial reasoning in text-to-image generation, replacing handcrafted losses with learned classifiers.

DetailsMotivation: Text-to-image diffusion models often fail at basic spatial reasoning tasks (e.g., placing objects in correct relative positions) that humans find trivial, and existing methods use suboptimal handcrafted losses.

Method: Train lightweight classifiers to decode spatial relationships from diffusion model’s cross-attention maps, then use these classifiers as learned loss functions during inference. Uses dual-inversion strategy to prevent shortcuts and enforce geometric understanding.

Result: Dramatically improves spatial accuracy: from 0.20 to 0.61 on FLUX.1-dev and from 0.07 to 0.54 on SD2.1 across standard benchmarks. Generalizes to multiple relations with significant accuracy improvements.

Conclusion: Learning spatial objectives directly from model’s internal representations through data-driven classifiers is more effective than handcrafted losses, enabling substantial improvements in spatial reasoning for text-to-image generation.

Abstract: Text-to-image diffusion models can generate stunning visuals, yet they often fail at tasks children find trivial–like placing a dog to the right of a teddy bear rather than to the left. When combinations get more unusual–a giraffe above an airplane–these failures become even more pronounced. Existing methods attempt to fix these spatial reasoning failures through model fine-tuning or test-time optimization with handcrafted losses that are suboptimal. Rather than imposing our assumptions about spatial encoding, we propose learning these objectives directly from the model’s internal representations. We introduce Learn-to-Steer, a novel framework that learns data-driven objectives for test-time optimization rather than handcrafting them. Our key insight is to train a lightweight classifier that decodes spatial relationships from the diffusion model’s cross-attention maps, then deploy this classifier as a learned loss function during inference. Training such classifiers poses a surprising challenge: they can take shortcuts by detecting linguistic traces rather than learning true spatial patterns. We solve this with a dual-inversion strategy that enforces geometric understanding. Our method dramatically improves spatial accuracy: from 0.20 to 0.61 on FLUX.1-dev and from 0.07 to 0.54 on SD2.1 across standard benchmarks. Moreover, our approach generalizes to multiple relations and significantly improves accuracy.

[433] Hues and Cues: Human vs. CLIP

Nuria Alabau-Bosque, Jorge Vila-Tomás, Paula Daudén-Oliver, Pablo Hernández-Cámara, Jose Manuel Jaén-Lorites, Valero Laparra, Jesús Malo

Main category: cs.CV

TL;DR: Proposes using board games to evaluate AI models, specifically testing CLIP’s color perception and naming through Hues & Cues game, revealing cultural biases and abstraction inconsistencies.

DetailsMotivation: Traditional evaluation methods often miss testing human-like characteristics in AI models. Board games provide a novel way to assess capabilities like color perception and naming that challenge different human traits.

Method: Testing CLIP’s color perception and color naming capabilities by playing the board game Hues & Cues, comparing its performance against human observers.

Result: CLIP shows general alignment with human color perception but reveals cultural biases and inconsistencies when handling different abstraction levels that are hard to detect with standard benchmarks.

Conclusion: Board games offer valuable alternative evaluation methods that can expose model deficiencies not apparent through conventional testing approaches, highlighting the importance of diverse assessment strategies.

Abstract: Playing games is inherently human, and a lot of games are created to challenge different human characteristics. However, these tasks are often left out when evaluating the human-like nature of artificial models. The objective of this work is proposing a new approach to evaluate artificial models via board games. To this effect, we test the color perception and color naming capabilities of CLIP by playing the board game Hues & Cues and assess its alignment with humans. Our experiments show that CLIP is generally well aligned with human observers, but our approach brings to light certain cultural biases and inconsistencies when dealing with different abstraction levels that are hard to identify with other testing strategies. Our findings indicate that assessing models with different tasks like board games can make certain deficiencies in the models stand out in ways that are difficult to test with the commonly used benchmarks.

[434] OmniActor: A Generalist GUI and Embodied Agent for 2D&3D Worlds

Longrong Yang, Zhixiong Zeng, Yufeng Zhong, Jing Huang, Liming Zheng, Lei Chen, Haibo Qiu, Zequn Qin, Lin Ma, Xi Li

Main category: cs.CV

TL;DR: OmniActor is a generalist agent that bridges GUI and embodied environments using a Layer-heterogeneity MoE architecture inspired by human brain cerebrum-cerebellum mechanism, successfully resolving data conflicts while leveraging synergies between different modality data.

DetailsMotivation: Current multimodal agents focus on either GUI (2D virtual) or embodied (3D real) environments separately, but complex tasks require interleaved interaction with both types of environments. Initial attempts to mix GUI and embodied data resulted in performance degeneration due to data conflicts.

Method: Proposed Layer-heterogeneity MoE architecture that separates deep-layer parameters to eliminate conflict between GUI and embodied data while sharing shallow-layer parameters to leverage their synergy. Also unified action spaces and collected large-scale GUI/embodied data from various sources.

Result: OmniActor outperforms agents trained only on GUI or embodied data in both GUI and embodied tasks. The approach significantly improves performance across different scenarios, especially in GUI tasks.

Conclusion: The cerebrum-cerebellum inspired architecture successfully resolves data conflicts while maintaining synergies between different environment types, enabling effective generalist agents that can operate across both GUI and embodied environments.

Abstract: Multimodal large language models are evolving toward multimodal agents capable of proactively executing tasks. Most agent research focuses on GUI or embodied scenarios, which correspond to agents interacting with 2D virtual worlds or 3D real worlds, respectively. However, many complex tasks typically require agents to interleavely interact with these two types of environment. We initially mix GUI and embodied data to train, but find the performance degeneration brought by the data conflict. Further analysis reveals that GUI and embodied data exhibit synergy and conflict at the shallow and deep layers, respectively, which resembles the cerebrum-cerebellum mechanism in the human brain. To this end, we propose a high-performance generalist agent OmniActor, designed from both structural and data perspectives. First, we propose Layer-heterogeneity MoE to eliminate the conflict between GUI and embodied data by separating deep-layer parameters, while leverage their synergy by sharing shallow-layer parameters. By successfully leveraging the synergy and eliminating the conflict, OmniActor outperforms agents only trained by GUI or embodied data in GUI or embodied tasks. Furthermore, we unify the action spaces of GUI and embodied tasks, and collect large-scale GUI and embodied data from various sources for training. This significantly improves OmniActor under different scenarios, especially in GUI tasks. The code will be publicly available.

[435] Ordinal Adaptive Correction: A Data-Centric Approach to Ordinal Image Classification with Noisy Labels

Alireza Sedighi Moghaddam, Mohammad Reza Mohammadi

Main category: cs.CV

TL;DR: ORDAC is a novel method for detecting and correcting label noise in ordinal image classification using adaptive label distribution learning, achieving significant performance improvements on benchmark datasets.

DetailsMotivation: Label noise in ordinal classification tasks degrades model performance due to ambiguous class boundaries, requiring effective noise correction methods to improve reliability.

Method: Proposes ORDAC (ORDinal Adaptive Correction) that uses Label Distribution Learning to dynamically adjust mean and standard deviation of label distributions during training, correcting noisy samples instead of discarding them.

Result: Significant improvements on Adience (age estimation) and Diabetic Retinopathy datasets - reduced MAE from 0.86 to 0.62 and increased recall from 0.37 to 0.49 with 40% noise on Adience.

Conclusion: Adaptive label correction using label distributions effectively enhances robustness and accuracy of ordinal classification models in noisy data scenarios.

Abstract: Labeled data is a fundamental component in training supervised deep learning models for computer vision tasks. However, the labeling process, especially for ordinal image classification where class boundaries are often ambiguous, is prone to error and noise. Such label noise can significantly degrade the performance and reliability of machine learning models. This paper addresses the problem of detecting and correcting label noise in ordinal image classification tasks. To this end, a novel data-centric method called ORDinal Adaptive Correction (ORDAC) is proposed for adaptive correction of noisy labels. The proposed approach leverages the capabilities of Label Distribution Learning (LDL) to model the inherent ambiguity and uncertainty present in ordinal labels. During training, ORDAC dynamically adjusts the mean and standard deviation of the label distribution for each sample. Rather than discarding potentially noisy samples, this approach aims to correct them and make optimal use of the entire training dataset. The effectiveness of the proposed method is evaluated on benchmark datasets for age estimation (Adience) and disease severity detection (Diabetic Retinopathy) under various asymmetric Gaussian noise scenarios. Results show that ORDAC and its extended versions (ORDAC_C and ORDAC_R) lead to significant improvements in model performance. For instance, on the Adience dataset with 40% noise, ORDAC_R reduced the mean absolute error from 0.86 to 0.62 and increased the recall metric from 0.37 to 0.49. The method also demonstrated its effectiveness in correcting intrinsic noise present in the original datasets. This research indicates that adaptive label correction using label distributions is an effective strategy to enhance the robustness and accuracy of ordinal classification models in the presence of noisy data.

[436] Category-Aware 3D Object Composition with Disentangled Texture and Shape Multi-view Diffusion

Zeren Xiong, Zikun Chen, Zedong Zhang, Xiang Li, Ying Tai, Jian Yang, Jun Li

Main category: cs.CV

TL;DR: A novel 3D object synthesis method called C33D that composites 3D models with different object categories using adaptive text-image harmony and multi-view diffusion for consistent textures and accurate shapes.

DetailsMotivation: Existing 3D generation methods struggle to effectively integrate multiple content sources, resulting in inconsistent textures and inaccurate shapes when creating novel 3D models by compositing different object categories.

Method: Renders multi-view images and normal maps from input 3D model, generates novel 2D object using adaptive text-image harmony, then applies texture multi-view diffusion for consistency and shape multi-view diffusion for accuracy before reconstructing the final 3D model.

Result: Extensive experiments demonstrate effectiveness, producing impressive 3D creations like shark(3D)-crocodile(text) composites with coherent structures and consistent textures.

Conclusion: The proposed C33D approach successfully addresses the challenges of 3D object synthesis by effectively integrating multiple content sources to create novel and structurally coherent 3D models with improved texture consistency and shape accuracy.

Abstract: In this paper, we tackle a new task of 3D object synthesis, where a 3D model is composited with another object category to create a novel 3D model. However, most existing text/image/3D-to-3D methods struggle to effectively integrate multiple content sources, often resulting in inconsistent textures and inaccurate shapes. To overcome these challenges, we propose a straightforward yet powerful approach, category+3D-to-3D (C33D), for generating novel and structurally coherent 3D models. Our method begins by rendering multi-view images and normal maps from the input 3D model, then generating a novel 2D object using adaptive text-image harmony (ATIH) with the front-view image and a text description from another object category as inputs. To ensure texture consistency, we introduce texture multi-view diffusion, which refines the textures of the remaining multi-view RGB images based on the novel 2D object. For enhanced shape accuracy, we propose shape multi-view diffusion to improve the 2D shapes of both the multi-view RGB images and the normal maps, also conditioned on the novel 2D object. Finally, these outputs are used to reconstruct a complete and novel 3D model. Extensive experiments demonstrate the effectiveness of our method, yielding impressive 3D creations, such as shark(3D)-crocodile(text) in the first row of Fig. 1. A project page is available at: https://xzr52.github.io/C33D/

[437] Why Do MLLMs Struggle with Spatial Understanding? A Systematic Analysis from Data to Architecture

Wanyue Zhang, Yibin Huang, Yangbin Xu, JingJing Huang, Helu Zhi, Shuo Ren, Wang Xu, Jiajun Zhang

Main category: cs.CV

TL;DR: MLLMs struggle with spatial understanding despite recent progress. This paper presents systematic analysis across single-view, multi-view, and video scenarios, revealing data scaling limitations and architectural dependencies on visual encoder positional encoding.

DetailsMotivation: Existing research lacks comprehensive evaluation of MLLMs' spatial understanding limitations, which is essential for perception, reasoning, and planning in embodied environments.

Method: Proposed MulSeT benchmark for multi-view spatial understanding tasks. Conducted systematic analysis from data and architectural perspectives across three scenarios using designed experiments.

Result: Performance converges quickly with more training data but reaches low upper bounds, especially for spatial imagination tasks. Spatial understanding relies more on visual encoder’s positional encoding than language model’s in both cascaded and native MLLMs.

Conclusion: Merely expanding training data is insufficient. Future improvements require architectural design optimization and reasoning injection to enhance spatial reasoning capabilities in MLLMs.

Abstract: Spatial understanding is essential for Multimodal Large Language Models (MLLMs) to support perception, reasoning, and planning in embodied environments. Despite recent progress, existing studies reveal that MLLMs still struggle with spatial understanding. However, existing research lacks a comprehensive and systematic evaluation of these limitations, often restricted to isolated scenarios, such as single-view or video. In this work, we present a systematic analysis of spatial understanding from both data and architectural perspectives across three representative scenarios: single-view, multi-view, and video. We propose a benchmark named MulSeT (Multi-view Spatial Understanding Tasks), and design a series of experiments to analyze the spatial reasoning capabilities of MLLMs. From the data perspective, the performance of spatial understanding converges quickly as the training data increases, and the upper bound is relatively low, especially for tasks that require spatial imagination. This indicates that merely expanding training data is insufficient to achieve satisfactory performance. From the architectural perspective, we find that spatial understanding relies more heavily on the positional encoding within the visual encoder than within the language model, in both cascaded and native MLLMs. Moreover, we explore reasoning injection and envision future improvements through architectural design to optimize spatial understanding. These insights shed light on the limitations of current MLLMs and suggest new directions for improving spatial reasoning capabilities through data scaling and architectural tuning.

[438] MedDINOv3: How to adapt vision foundation models for medical image segmentation?

Yuheng Li, Yizhou Wu, Yuxiang Lai, Mingzhe Hu, Xiaofeng Yang

Main category: cs.CV

TL;DR: MedDINOv3 adapts DINOv3 foundation model for medical image segmentation through domain-adaptive pretraining on CT scans and architectural improvements, achieving state-of-the-art performance across multiple benchmarks.

DetailsMotivation: Current deep learning models lack generalizability across medical imaging modalities and institutions, and vision foundation models underperform specialized CNNs on medical segmentation due to domain gap between natural and medical images.

Method: Revisits plain ViTs with multi-scale token aggregation architecture, performs domain-adaptive pretraining on 3.87M CT slices using multi-stage DINOv3 recipe to learn robust dense features.

Result: Matches or exceeds state-of-the-art performance across four segmentation benchmarks, demonstrating foundation models as unified backbones for medical image segmentation.

Conclusion: MedDINOv3 provides an effective framework for adapting vision foundation models to medical imaging, overcoming domain gap challenges and achieving superior segmentation performance.

Abstract: Accurate segmentation of organs and tumors in CT and MRI scans is essential for diagnosis, treatment planning, and disease monitoring. While deep learning has advanced automated segmentation, most models remain task-specific, lacking generalizability across modalities and institutions. Vision foundation models (FMs) pretrained on billion-scale natural images offer powerful and transferable representations. However, adapting them to medical imaging faces two key challenges: (1) the ViT backbone of most foundation models still underperform specialized CNNs on medical image segmentation, and (2) the large domain gap between natural and medical images limits transferability. We introduce \textbf{MedDINOv3}, a simple and effective framework for adapting DINOv3 to medical segmentation. We first revisit plain ViTs and design a simple and effective architecture with multi-scale token aggregation. Then, we perform domain-adaptive pretraining on \textbf{CT-3M}, a curated collection of 3.87M axial CT slices, using a multi-stage DINOv3 recipe to learn robust dense features. MedDINOv3 matches or exceeds state-of-the-art performance across four segmentation benchmarks, demonstrating the potential of vision foundation models as unified backbones for medical image segmentation. The code is available at https://github.com/ricklisz/MedDINOv3.

[439] Decoupling Bidirectional Geometric Representations of 4D cost volume with 2D convolution

Xiaobao Wei, Changyong Shu, Zhaokun Yue, Chang Huang, Weiwei Liu, Shuai Yang, Lirong Yang, Peng Gao, Wenbin Zhang, Gaochao Zhu, Chengxiang Wang

Main category: cs.CV

TL;DR: DBStereo is a real-time stereo matching method that uses pure 2D convolutions for 4D cost aggregation, achieving better performance than 3D-based methods while being mobile-friendly.

DetailsMotivation: Existing high-performance stereo matching methods rely on 3D regularization that is unfriendly to mobile devices, while 2D methods struggle in ill-posed regions. There's a need for deployment-friendly yet accurate stereo matching.

Method: Proposes a 4D cost aggregation network using pure 2D convolutions with a lightweight bidirectional geometry aggregation block that captures spatial and disparity representations through decoupled learning.

Result: DBStereo outperforms all existing aggregation-based methods in both inference time and accuracy, even surpassing iterative-based method IGEV-Stereo, while achieving real-time performance.

Conclusion: Breaks the empirical design of using 3D convolutions for 4D cost volume and provides a simple yet strong baseline for decoupled aggregation paradigm in stereo matching.

Abstract: High-performance real-time stereo matching methods invariably rely on 3D regularization of the cost volume, which is unfriendly to mobile devices. And 2D regularization based methods struggle in ill-posed regions. In this paper, we present a deployment-friendly 4D cost aggregation network DBStereo, which is based on pure 2D convolutions. Specifically, we first provide a thorough analysis of the decoupling characteristics of 4D cost volume. And design a lightweight bidirectional geometry aggregation block to capture spatial and disparity representation respectively. Through decoupled learning, our approach achieves real-time performance and impressive accuracy simultaneously. Extensive experiments demonstrate that our proposed DBStereo outperforms all existing aggregation-based methods in both inference time and accuracy, even surpassing the iterative-based method IGEV-Stereo. Our study break the empirical design of using 3D convolutions for 4D cost volume and provides a simple yet strong baseline of the proposed decouple aggregation paradigm for further study. Code will be available at (\href{https://github.com/happydummy/DBStereo}{https://github.com/happydummy/DBStereo}) soon.

[440] From Noisy Labels to Intrinsic Structure: A Geometric-Structural Dual-Guided Framework for Noise-Robust Medical Image Segmentation

Tao Wang, Zhenxuan Zhang, Yuanbo Zhou, Xinlin Zhang, Yuanbin Chen, Tao Tan, Guang Yang, Tong Tong

Main category: cs.CV

TL;DR: GSD-Net is a novel network that integrates geometric and structural guidance to improve medical image segmentation performance under noisy annotations, achieving state-of-the-art results across multiple datasets.

DetailsMotivation: Medical image segmentation requires large-scale high-quality annotations that are costly and time-consuming to obtain. Even expert-labeled datasets contain noise from subjectivity and coarse delineations, which disrupts feature learning and model performance.

Method: Proposes Geometric-Structural Dual-Guided Network (GSD-Net) with: 1) Geometric Distance-Aware module that dynamically adjusts pixel-level weights using geometric features, 2) Structure-Guided Label Refinement module that refines labels with structural priors, and 3) Knowledge Transfer module to enrich supervision and improve local detail sensitivity.

Result: Achieved state-of-the-art performance under noisy annotations with improvements of 2.52% on Kvasir, 22.76% on Shenzhen, 8.87% on BU-SUC, and 4.59% on BraTS2020 under SR simulated noise. Evaluated on six public datasets with simulated and real-world noisy annotations.

Conclusion: GSD-Net effectively addresses the challenge of noisy annotations in medical image segmentation by integrating geometric and structural guidance, demonstrating robust performance across various noise types and datasets.

Abstract: The effectiveness of convolutional neural networks in medical image segmentation relies on large-scale, high-quality annotations, which are costly and time-consuming to obtain. Even expert-labeled datasets inevitably contain noise arising from subjectivity and coarse delineations, which disrupt feature learning and adversely impact model performance. To address these challenges, this study propose a Geometric-Structural Dual-Guided Network (GSD-Net), which integrates geometric and structural cues to improve robustness against noisy annotations. It incorporates a Geometric Distance-Aware module that dynamically adjusts pixel-level weights using geometric features, thereby strengthening supervision in reliable regions while suppressing noise. A Structure-Guided Label Refinement module further refines labels with structural priors, and a Knowledge Transfer module enriches supervision and improves sensitivity to local details. To comprehensively assess its effectiveness, we evaluated GSD-Net on six publicly available datasets: four containing three types of simulated label noise, and two with multi-expert annotations that reflect real-world subjectivity and labeling inconsistencies. Experimental results demonstrate that GSD-Net achieves state-of-the-art performance under noisy annotations, achieving improvements of 2.52% on Kvasir, 22.76% on Shenzhen, 8.87% on BU-SUC, and 4.59% on BraTS2020 under SR simulated noise. The codes of this study are available at https://github.com/ortonwang/GSD-Net.

[441] Faster and Better: Reinforced Collaborative Distillation and Self-Learning for Infrared-Visible Image Fusion

Yuhao Wang, Lingjuan Miao, Zhiqiang Zhou, Yajun Qiao, Lei Zhang

Main category: cs.CV

TL;DR: A reinforcement learning-driven collaborative distillation framework for infrared and visible image fusion that enables student models to learn from teachers while self-learning on challenging samples.

DetailsMotivation: To address the challenge of achieving high-quality image fusion with lightweight models by combining complementary information from different modalities more effectively.

Method: Proposes a collaborative distillation and self-learning framework using reinforcement learning. An RL agent identifies optimal training strategies, generates challenging samples for self-learning, and dynamically adjusts teacher guidance based on student performance.

Result: Experimental results show significant improvement in student performance and better fusion results compared to existing techniques.

Conclusion: The proposed framework successfully enhances image fusion quality through collaborative distillation and adaptive self-learning driven by reinforcement learning, outperforming conventional methods.

Abstract: Infrared and visible image fusion plays a critical role in enhancing scene perception by combining complementary information from different modalities. Despite recent advances, achieving high-quality image fusion with lightweight models remains a significant challenge. To bridge this gap, we propose a novel collaborative distillation and self-learning framework for image fusion driven by reinforcement learning. Unlike conventional distillation, this approach not only enables the student model to absorb image fusion knowledge from the teacher model, but more importantly, allows the student to perform self-learning on more challenging samples to enhance its capabilities. Particularly, in our framework, a reinforcement learning agent explores and identifies a more suitable training strategy for the student.The agent takes both the student’s performance and the teacher-student gap as inputs, which leads to the generation of challenging samples to facilitate the student’s self-learning. Simultaneously, it dynamically adjusts the teacher’s guidance strength based on the student’s state to optimize the knowledge transfer. Experimental results demonstrate that our method can significantly improve student performance and achieve better fusion results compared to existing techniques.

[442] Towards High-Fidelity, Identity-Preserving Real-Time Makeup Transfer: Decoupling Style Generation

Lydia Kin Ching Chau, Zhi Yu, Ruo Wei Jiang

Main category: cs.CV

TL;DR: A real-time virtual makeup try-on framework that separates makeup transfer into extraction and rendering steps, achieving high-fidelity results with temporal consistency and identity preservation.

DetailsMotivation: Existing makeup transfer methods struggle with disentangling cosmetics from skin tones, causing identity shifts and lacking real-time capabilities with temporal consistency needed for practical applications.

Method: Decouples makeup transfer into transparent makeup mask extraction and graphics-based rendering. Uses pseudo-ground-truth data from graphics rendering and unsupervised k-means clustering, with specialized training objectives including alpha-weighted reconstruction and lip color losses.

Result: Achieves robust makeup transfer across diverse poses, expressions, and skin tones while preserving temporal smoothness. Outperforms existing baselines in capturing fine details, maintaining stability, and preserving identity integrity.

Conclusion: The proposed framework successfully addresses key challenges in virtual makeup try-on by enabling real-time, high-fidelity cosmetic transfer with strong temporal consistency and identity preservation.

Abstract: We present a novel framework for real-time virtual makeup try-on that achieves high-fidelity, identity-preserving cosmetic transfer with robust temporal consistency. In live makeup transfer applications, it is critical to synthesize temporally coherent results that accurately replicate fine-grained makeup and preserve user’s identity. However, existing methods often struggle to disentangle semitransparent cosmetics from skin tones and other identify features, causing identity shifts and raising fairness concerns. Furthermore, current methods lack real-time capabilities and fail to maintain temporal consistency, limiting practical adoption. To address these challenges, we decouple makeup transfer into two steps: transparent makeup mask extraction and graphics-based mask rendering. After the makeup extraction step, the makeup rendering can be performed in real time, enabling live makeup try-on. Our makeup extraction model trained on pseudo-ground-truth data generated via two complementary methods: a graphics-based rendering pipeline and an unsupervised k-means clustering approach. To further enhance transparency estimation and color fidelity, we propose specialized training objectives, including alpha-weighted reconstruction and lip color losses. Our method achieves robust makeup transfer across diverse poses, expressions, and skin tones while preserving temporal smoothness. Extensive experiments demonstrate that our approach outperforms existing baselines in capturing fine details, maintaining temporal stability, and preserving identity integrity.

[443] RiverScope: High-Resolution River Masking Dataset

Rangel Daroya, Taylor Rowley, Jonathan Flores, Elisa Friedmann, Fiona Bennitt, Heejin An, Travis Simmons, Marissa Jean Hughes, Camryn L Kluetmeier, Solomon Kica, J. Daniel Vélez, Sarah E. Esenther, Thomas E. Howard, Yanqi Ye, Audrey Turcotte, Colin Gleason, Subhransu Maji

Main category: cs.CV

TL;DR: RiverScope is a high-resolution dataset for river and surface water monitoring that addresses limitations of low-resolution satellite data, featuring expert-labeled masks and multi-sensor co-registration to enable accurate river width estimation and hydrological modeling.

DetailsMotivation: Current satellite data poorly captures narrow or sediment-rich rivers at fine spatial and temporal scales, creating challenges for monitoring surface water dynamics that are critical for climate systems, ecosystems, agriculture, and disaster resilience.

Method: Developed through computer science-hydrology collaboration, RiverScope includes 1,145 high-resolution images with expert-labeled river masks, co-registered with Sentinel-2, SWOT, and SWORD data. Evaluated deep networks across multiple architectures, pretraining strategies, and training datasets.

Result: Achieved median error of 7.2 meters for river width estimation - significantly outperforming existing satellite-derived methods. Best models combined transfer learning with multispectral PlanetScope channels via learned adaptors.

Conclusion: RiverScope provides a valuable resource for fine-scale hydrological modeling and multi-sensor water monitoring, supporting climate adaptation and sustainable water management with unprecedented accuracy for river width estimation.

Abstract: Surface water dynamics play a critical role in Earth’s climate system, influencing ecosystems, agriculture, disaster resilience, and sustainable development. Yet monitoring rivers and surface water at fine spatial and temporal scales remains challenging – especially for narrow or sediment-rich rivers that are poorly captured by low-resolution satellite data. To address this, we introduce RiverScope, a high-resolution dataset developed through collaboration between computer science and hydrology experts. RiverScope comprises 1,145 high-resolution images (covering 2,577 square kilometers) with expert-labeled river and surface water masks, requiring over 100 hours of manual annotation. Each image is co-registered with Sentinel-2, SWOT, and the SWOT River Database (SWORD), enabling the evaluation of cost-accuracy trade-offs across sensors – a key consideration for operational water monitoring. We also establish the first global, high-resolution benchmark for river width estimation, achieving a median error of 7.2 meters – significantly outperforming existing satellite-derived methods. We extensively evaluate deep networks across multiple architectures (e.g., CNNs and transformers), pretraining strategies (e.g., supervised and self-supervised), and training datasets (e.g., ImageNet and satellite imagery). Our best-performing models combine the benefits of transfer learning with the use of all the multispectral PlanetScope channels via learned adaptors. RiverScope provides a valuable resource for fine-scale and multi-sensor hydrological modeling, supporting climate adaptation and sustainable water management.

[444] GenCompositor: Generative Video Compositing with Diffusion Transformer

Shuzhou Yang, Xiaoyu Li, Xiaodong Cun, Guangzhi Wang, Lingen Li, Ying Shan, Jian Zhang

Main category: cs.CV

TL;DR: Automated video compositing using Diffusion Transformers to inject foreground identity and motion into target videos with user control over size, trajectory, and attributes.

DetailsMotivation: Traditional video compositing requires intensive manual labor and expert collaboration, leading to long production cycles and high costs. The paper aims to automate this process using generative models.

Method: Developed a Diffusion Transformer (DiT) pipeline with background preservation branch using masked token injection, DiT fusion blocks with full self-attention, foreground augmentation, and Extended Rotary Position Embedding (ERoPE) for layout fusion. Created VideoComp dataset with 61K video sets.

Result: The method effectively achieves generative video compositing and outperforms existing solutions in both fidelity and consistency metrics.

Conclusion: The proposed generative video compositing approach successfully automates the traditionally labor-intensive process, enabling adaptive injection of foreground elements with user customization while maintaining video consistency.

Abstract: Video compositing combines live-action footage to create video production, serving as a crucial technique in video creation and film production. Traditional pipelines require intensive labor efforts and expert collaboration, resulting in lengthy production cycles and high manpower costs. To address this issue, we automate this process with generative models, called generative video compositing. This new task strives to adaptively inject identity and motion information of foreground video to the target video in an interactive manner, allowing users to customize the size, motion trajectory, and other attributes of the dynamic elements added in final video. Specifically, we designed a novel Diffusion Transformer (DiT) pipeline based on its intrinsic properties. To maintain consistency of the target video before and after editing, we revised a light-weight DiT-based background preservation branch with masked token injection. As to inherit dynamic elements from other sources, a DiT fusion block is proposed using full self-attention, along with a simple yet effective foreground augmentation for training. Besides, for fusing background and foreground videos with different layouts based on user control, we developed a novel position embedding, named Extended Rotary Position Embedding (ERoPE). Finally, we curated a dataset comprising 61K sets of videos for our new task, called VideoComp. This data includes complete dynamic elements and high-quality target videos. Experiments demonstrate that our method effectively realizes generative video compositing, outperforming existing possible solutions in fidelity and consistency.

[445] TeRA: Rethinking Text-driven Realistic 3D Avatar Generation

Yanwen Wang, Yiyu Zhuang, Jiawei Zhang, Li Wang, Yifei Zeng, Xun Cao, Xinxin Zuo, Hao Zhu

Main category: cs.CV

TL;DR: TeRA is a two-stage text-to-avatar generation framework that uses latent diffusion in a structured 3D space, eliminating slow iterative optimization and enabling efficient photorealistic avatar generation with text-based customization.

DetailsMotivation: To overcome the inefficiencies of previous SDS-based models and large 3D generative models for text-to-avatar generation, which suffer from slow iterative optimization processes.

Method: Two-stage training: 1) Distill a decoder to create structured latent space from large human reconstruction model, 2) Train text-controlled latent diffusion model to generate avatars within this latent space.

Result: Superior performance over previous text-to-avatar generative models in both subjective and objective evaluations, with faster generation and text-based partial customization capabilities.

Conclusion: TeRA provides a more efficient and effective framework for photorealistic 3D human avatar generation from text, eliminating slow optimization while enabling customization through structured 3D representation.

Abstract: In this paper, we rethink text-to-avatar generative models by proposing TeRA, a more efficient and effective framework than the previous SDS-based models and general large 3D generative models. Our approach employs a two-stage training strategy for learning a native 3D avatar generative model. Initially, we distill a decoder to derive a structured latent space from a large human reconstruction model. Subsequently, a text-controlled latent diffusion model is trained to generate photorealistic 3D human avatars within this latent space. TeRA enhances the model performance by eliminating slow iterative optimization and enables text-based partial customization through a structured 3D human representation. Experiments have proven our approach’s superiority over previous text-to-avatar generative models in subjective and objective evaluation.

[446] Anisotropic Fourier Features for Positional Encoding in Medical Imaging

Nabil Jabareen, Dongsheng Yuan, Dingming Liu, Foo-Wei Ten, Sören Lukassen

Main category: cs.CV

TL;DR: Proposes Anisotropic Fourier Feature Positional Encoding (AFPE) to address limitations of existing positional encodings in medical imaging, showing significant performance improvements in anisotropic settings.

DetailsMotivation: Standard positional encodings like Sinusoidal PEs struggle with Euclidean distance preservation in high-dimensional spaces, while Isotropic Fourier Feature PEs cannot handle anisotropy in medical images, requiring a better solution for medical imaging tasks.

Method: Developed AFPE as a generalization of IFPE that incorporates anisotropic, class-specific, and domain-specific spatial dependencies, and benchmarked it against common PEs on multi-label chest X-ray classification, CT organ classification, and echocardiography ejection fraction regression.

Result: AFPE significantly outperformed state-of-the-art positional encodings in all tested anisotropic settings, demonstrating that optimal PE choice depends on data anisotropy and structure shape.

Conclusion: For anisotropic medical images and videos, choosing an anisotropic positional encoding that fits both the data characteristics and the shape of interest is crucial for optimal model performance.

Abstract: The adoption of Transformer-based architectures in the medical domain is growing rapidly. In medical imaging, the analysis of complex shapes - such as organs, tissues, or other anatomical structures - combined with the often anisotropic nature of high-dimensional images complicates these adaptations. In this study, we critically examine the role of Positional Encodings (PEs), arguing that commonly used approaches may be suboptimal for the specific challenges of medical imaging. Sinusoidal Positional Encodings (SPEs) have proven effective in vision tasks, but they struggle to preserve Euclidean distances in higher-dimensional spaces. Isotropic Fourier Feature Positional Encodings (IFPEs) have been proposed to better preserve Euclidean distances, but they lack the ability to account for anisotropy in images. To address these limitations, we propose Anisotropic Fourier Feature Positional Encoding (AFPE), a generalization of IFPE that incorporates anisotropic, class-specific, and domain-specific spatial dependencies. We systematically benchmark AFPE against commonly used PEs on multi-label classification in chest X-rays, organ classification in CT images, and ejection fraction regression in echocardiography. Our results demonstrate that choosing the correct PE can significantly improve model performance. We show that the optimal PE depends on the shape of the structure of interest and the anisotropy of the data. Finally, our proposed AFPE significantly outperforms state-of-the-art PEs in all tested anisotropic settings. We conclude that, in anisotropic medical images and videos, it is of paramount importance to choose an anisotropic PE that fits the data and the shape of interest.

[447] Enhancing Fitness Movement Recognition with Attention Mechanism and Pre-Trained Feature Extractors

Shanjid Hasan Nishat, Srabonti Deb, Mohiuddin Ahmed

Main category: cs.CV

TL;DR: Lightweight framework combining 2D CNNs (ResNet50, EfficientNet, ViT) with LSTM and spatial attention for real-time fitness movement recognition, achieving 93.34% accuracy on UCF101 subset.

DetailsMotivation: Existing deep learning approaches for fitness activity recognition rely on computationally intensive 3D models, limiting real-time feasibility in resource-constrained settings.

Method: Integrates pre-trained 2D CNNs (ResNet50, EfficientNet, Vision Transformers) with LSTM network enhanced by spatial attention mechanism to extract spatial features and capture temporal dependencies.

Result: Achieved peak accuracy of 93.34% with ResNet50-based configuration on curated UCF101 dataset subset, outperforming several state-of-the-art HAR systems.

Conclusion: Proposed method offers scalable, real-time-capable solution for fitness activity recognition with broader applications in vision-based health and activity monitoring.

Abstract: Fitness movement recognition, a focused subdomain of human activity recognition (HAR), plays a vital role in health monitoring, rehabilitation, and personalized fitness training by enabling automated exercise classification from video data. However, many existing deep learning approaches rely on computationally intensive 3D models, limiting their feasibility in real-time or resource-constrained settings. In this paper, we present a lightweight and effective framework that integrates pre-trained 2D Convolutional Neural Networks (CNNs) such as ResNet50, EfficientNet, and Vision Transformers (ViT) with a Long Short-Term Memory (LSTM) network enhanced by spatial attention. These models efficiently extract spatial features while the LSTM captures temporal dependencies, and the attention mechanism emphasizes informative segments. We evaluate the framework on a curated subset of the UCF101 dataset, achieving a peak accuracy of 93.34% with the ResNet50-based configuration. Comparative results demonstrate the superiority of our approach over several state-of-the-art HAR systems. The proposed method offers a scalable and real-time-capable solution for fitness activity recognition with broader applications in vision-based health and activity monitoring.

[448] Mix-modal Federated Learning for MRI Image Segmentation

Guyue Hu, Siyuan Song, Jingpeng Sun, Zhe Jin, Chenglong Li, Jin Tang

Main category: cs.CV

TL;DR: Proposes a novel mix-modal federated learning framework (MDM-MixMFL) for MRI segmentation that handles client-wise modality and data heterogeneity through modality decoupling and memorizing mechanisms.

DetailsMotivation: Existing centralized multimodal MRI segmentation methods are inapplicable to non-centralized medical scenarios where distributed hospitals have diverse mixed MRI modalities and suffer from extensive client-wise heterogeneity.

Method: MDM-MixMFL framework with modality decoupling strategy (separating modality-tailored and modality-shared information) and modality memorizing mechanism (storing dynamically refreshed client-shared modality prototypes).

Result: Extensive experiments on two public MRI datasets demonstrate the effectiveness and superiority of the proposed method.

Conclusion: The proposed framework successfully addresses mix-modal federated learning challenges for MRI segmentation in distributed medical settings with heterogeneous modalities and data.

Abstract: Magnetic resonance imaging (MRI) image segmentation is crucial in diagnosing and treating many diseases, such as brain tumors. Existing MRI image segmentation methods mainly fall into a centralized multimodal paradigm, which is inapplicable in engineering non-centralized mix-modal medical scenarios. In this situation, each distributed client (hospital) processes multiple mixed MRI modalities, and the modality set and image data for each client are diverse, suffering from extensive client-wise modality heterogeneity and data heterogeneity. In this paper, we first formulate non-centralized mix-modal MRI image segmentation as a new paradigm for federated learning (FL) that involves multiple modalities, called mix-modal federated learning (MixMFL). It distinguishes from existing multimodal federating learning (MulMFL) and cross-modal federating learning (CroMFL) paradigms. Then, we proposed a novel modality decoupling and memorizing mix-modal federated learning framework (MDM-MixMFL) for MRI image segmentation, which is characterized by a modality decoupling strategy and a modality memorizing mechanism. Specifically, the modality decoupling strategy disentangles each modality into modality-tailored and modality-shared information. During mix-modal federated updating, corresponding modality encoders undergo tailored and shared updating, respectively. It facilitates stable and adaptive federating aggregation of heterogeneous data and modalities from distributed clients. Besides, the modality memorizing mechanism stores client-shared modality prototypes dynamically refreshed from every modality-tailored encoder to compensate for incomplete modalities in each local client. It further benefits modality aggregation and fusion processes during mixmodal federated learning. Extensive experiments on two public datasets for MRI image segmentation demonstrate the effectiveness and superiority of our methods.

[449] Motion-Refined DINOSAUR for Unsupervised Multi-Object Discovery

Xinrui Gong, Oliver Hahn, Christoph Reich, Krishnakant Singh, Simone Schaub-Meyer, Daniel Cremers, Stefan Roth

Main category: cs.CV

TL;DR: MR-DINOSAUR is a fully unsupervised approach for multi-object discovery that extends DINOSAUR with motion-refined pseudo labels and slot deactivation, achieving state-of-the-art results without any supervision.

DetailsMotivation: Current unsupervised MOD approaches still rely on supervision for pseudo labels to train object-centric learning models, creating a limitation that needs to be addressed with truly unsupervised methods.

Method: Extends self-supervised DINOSAUR by generating high-quality unsupervised pseudo labels from video frames without camera motion, performs motion segmentation on optical flow, refines slot representations, and trains a slot deactivation module for foreground/background assignment.

Result: Achieves strong multi-object discovery results on TRI-PD and KITTI datasets, outperforming previous state-of-the-art methods despite being fully unsupervised.

Conclusion: MR-DINOSAUR demonstrates that conceptually simple motion refinement and slot deactivation can enable fully unsupervised multi-object discovery with competitive performance.

Abstract: Unsupervised multi-object discovery (MOD) aims to detect and localize distinct object instances in visual scenes without any form of human supervision. Recent approaches leverage object-centric learning (OCL) and motion cues from video to identify individual objects. However, these approaches use supervision to generate pseudo labels to train the OCL model. We address this limitation with MR-DINOSAUR – Motion-Refined DINOSAUR – a minimalistic unsupervised approach that extends the self-supervised pre-trained OCL model, DINOSAUR, to the task of unsupervised multi-object discovery. We generate high-quality unsupervised pseudo labels by retrieving video frames without camera motion for which we perform motion segmentation of unsupervised optical flow. We refine DINOSAUR’s slot representations using these pseudo labels and train a slot deactivation module to assign slots to foreground and background. Despite its conceptual simplicity, MR-DINOSAUR achieves strong multi-object discovery results on the TRI-PD and KITTI datasets, outperforming the previous state of the art despite being fully unsupervised.

[450] FastVGGT: Training-Free Acceleration of Visual Geometry Transformer

You Shen, Zhipeng Zhang, Yansong Qu, Liujuan Cao

Main category: cs.CV

TL;DR: FastVGGT introduces token merging to accelerate VGGT, achieving 4x speedup on 1000-image inputs while maintaining reconstruction quality in 3D vision tasks.

DetailsMotivation: Scaling 3D foundation models to long-sequence images is challenging due to inference inefficiency, with identified bottlenecks and token collapse issues in VGGT.

Method: Proposes FastVGGT with training-free token merging mechanism, unique 3D-specific token partitioning strategy to eliminate redundant computation while preserving reconstruction capacity.

Result: Achieves 4x speedup over VGGT with 1000 input images while mitigating error accumulation in long-sequence scenarios across multiple 3D geometry benchmarks.

Conclusion: Token merging shows promise as a principled solution for scalable 3D vision systems, effectively addressing inference efficiency without compromising performance.

Abstract: Foundation models for 3D vision have recently demonstrated remarkable capabilities in 3D perception. However, scaling these models to long-sequence image inputs remains a significant challenge due to inference-time inefficiency. In this work, we present a detailed analysis of VGGT, a state-of-the-art feed-forward visual geometry model and identify its primary bottleneck. Visualization further reveals a token collapse phenomenon in the attention maps. Motivated by these findings, we explore the potential of token merging in the feed-forward visual geometry model. Owing to the unique architectural and task-specific properties of 3D models, directly applying existing merging techniques proves challenging. To this end, we propose FastVGGT, which, for the first time, leverages token merging in the 3D domain through a training-free mechanism for accelerating VGGT. we devise a unique token partitioning strategy tailored to 3D architectures and tasks, effectively eliminating redundant computation while preserving VGGT’s powerful reconstruction capacity. Extensive experiments on multiple 3D geometry benchmarks validate the effectiveness of our approach. Notably, with 1000 input images, FastVGGT achieves a 4x speedup over VGGT while mitigating error accumulation in long-sequence scenarios. These findings underscore the potential of token merging as a principled solution for scalable 3D vision systems. Code is available at: https://mystorm16.github.io/fastvggt/.

[451] Net2Brain: A Toolbox to compare artificial vision models with human brain responses

Domenic Bersch, Kshitij Dwivedi, Martina Vilas, Radoslaw M. Cichy, Gemma Roig

Main category: cs.CV

TL;DR: Net2Brain is a toolbox that compares artificial neural networks and human brain recordings using representational similarity analysis, supporting over 600 DNNs across various vision tasks.

DetailsMotivation: Existing toolboxes have limited functionality and focus only on small subsets of supervised image classification models, lacking comprehensive comparison capabilities between diverse DNNs and brain data.

Method: The toolbox extracts activations from 600+ DNNs trained on various vision tasks, computes representational dissimilarity matrices (RDMs), and compares them to brain recordings using RSA and weighted RSA in specific ROIs and with searchlight search.

Result: Net2Brain enables comprehensive comparison of DNN representational spaces with human brain data, supporting addition of new stimulus datasets and brain recordings for evaluation.

Conclusion: The toolbox provides a powerful interface for testing cognitive computational neuroscience hypotheses by facilitating detailed comparison between artificial neural networks and human brain representations across diverse vision tasks.

Abstract: We introduce Net2Brain, a graphical and command-line user interface toolbox for comparing the representational spaces of artificial deep neural networks (DNNs) and human brain recordings. While different toolboxes facilitate only single functionalities or only focus on a small subset of supervised image classification models, Net2Brain allows the extraction of activations of more than 600 DNNs trained to perform a diverse range of vision-related tasks (e.g semantic segmentation, depth estimation, action recognition, etc.), over both image and video datasets. The toolbox computes the representational dissimilarity matrices (RDMs) over those activations and compares them to brain recordings using representational similarity analysis (RSA), weighted RSA, both in specific ROIs and with searchlight search. In addition, it is possible to add a new data set of stimuli and brain recordings to the toolbox for evaluation. We demonstrate the functionality and advantages of Net2Brain with an example showcasing how it can be used to test hypotheses of cognitive computational neuroscience.

[452] Cross-Modal Adapter for Vision-Language Retrieval

Haojun Jiang, Jianke Zhang, Rui Huang, Chunjiang Ge, Zanlin Ni, Shiji Song, Gao Huang

Main category: cs.CV

TL;DR: A parameter-efficient cross-modal adapter for vision-language retrieval that reduces fine-tuning parameters while maintaining performance across multiple datasets.

DetailsMotivation: Pre-trained models like CLIP risk overfitting when fully fine-tuned on downstream retrieval tasks, and storing large models for each task is costly.

Method: Adapter-based approach designed for multi-modal domain with encoder-level implicit cross-modal interactions between vision and language encoders, using few parameterization layers.

Result: Outperforms adapter-based methods on image-text retrieval (MSCOCO, Flickr30K) and video-text retrieval (MSR-VTT, DiDeMo, ActivityNet) datasets while reducing parameters and training time.

Conclusion: The proposed cross-modal adapter enables efficient transfer learning by fixing pre-trained parameters, allowing model sharing across datasets while maintaining strong retrieval performance.

Abstract: Vision-language retrieval is an important multi-modal learning topic, where the goal is to retrieve the most relevant visual candidate for a given text query. Recently, pre-trained models, e.g., CLIP, show great potential on retrieval tasks. However, as pre-trained models are scaling up, fully fine-tuning them on donwstream retrieval datasets has a high risk of overfitting. Moreover, in practice, it would be costly to train and store a large model for each task. To overcome the above issues, we present a novel Cross-Modal Adapter for parameter-efficient transfer learning. Inspired by adapter-based methods, we adjust the pre-trained model with a few parameterization layers. However, there are two notable differences. First, our method is designed for the multi-modal domain. Secondly, it allows encoder-level implicit cross-modal interactions between vision and language encoders. Although surprisingly simple, our approach has three notable benefits: (1) reduces the vast majority of fine-tuned parameters, (2) saves training time, and (3) allows all the pre-trained parameters to be fixed, enabling the pre-trained model to be shared across datasets. Extensive experiments demonstrate that, without bells and whistles, our approach outperforms adapter-based methods on image-text retrieval datasets (MSCOCO, Flickr30K) and video-text retrieval datasets (MSR-VTT, DiDeMo, and ActivityNet).

[453] OA-DET3D: Embedding Object Awareness as a General Plug-in for Multi-Camera 3D Object Detection

Xiaomeng Chu, Jiajun Deng, Jianmin Ji, Yu Zhang, Houqiang Li, Yanyong Zhang

Main category: cs.CV

TL;DR: OA-DET3D is a plug-in module that enhances 3D object detection by incorporating object awareness through object-centric depth learning and foreground pseudo points, improving feature representation in BEV-based systems.

DetailsMotivation: The transformation from image-plane view to 3D space causes feature clutter and distortion, making objects blur into the background. The paper aims to incorporate supplementary cues for better object differentiation.

Method: Uses object-level supervision from 3D bounding boxes to guide depth distribution learning, selects foreground pixels with 2D object detector and projects them into 3D space for pseudo-voxel feature encoding, then incorporates these features into BEV representation using deformable attention mechanism.

Result: Achieves consistent improvements over BEV-based baselines on nuScenes and Argoverse 2 datasets in terms of average precision and comprehensive detection score.

Conclusion: OA-DET3D effectively enhances 3D object detection performance by introducing object awareness through depth learning and pseudo-point features, demonstrating general applicability to various existing 3D detection pipelines.

Abstract: The recent advance in multi-camera 3D object detection is featured by bird’s-eye view (BEV) representation or object queries. However, the ill-posed transformation from image-plane view to 3D space inevitably causes feature clutter and distortion, making the objects blur into the background. To this end, we explore how to incorporate supplementary cues for differentiating objects in the transformed feature representation. Formally, we introduce OA-DET3D, a general plug-in module that improves 3D object detection by bringing object awareness into a variety of existing 3D object detection pipelines. Specifically, OA-DET3D boosts the representation of objects by leveraging object-centric depth information and foreground pseudo points. First, we use object-level supervision from the properties of each 3D bounding box to guide the network in learning the depth distribution. Next, we select foreground pixels using a 2D object detector and project them into 3D space for pseudo-voxel feature encoding. Finally, the object-aware depth features and pseudo-voxel features are incorporated into the BEV representation or query feature from the baseline model with a deformable attention mechanism. We conduct extensive experiments on the nuScenes dataset and Argoverse 2 dataset to validate the merits of OA-DET3D. Our method achieves consistent improvements over the BEV-based baselines in terms of both average precision and comprehensive detection score.

[454] Multiscale Feature Learning Using Co-Tuplet Loss for Offline Handwritten Signature Verification

Fu-Hsien Huang, Hsin-Min Lu

Main category: cs.CV

TL;DR: MS-SigNet with co-tuplet loss for offline handwritten signature verification, learning multi-scale features to distinguish genuine signatures from skilled forgeries, tested on multiple language datasets including new Chinese HanSig dataset.

DetailsMotivation: Address challenges in handwritten signature verification including inter-writer similarity, intra-writer variations, and limited signature samples for robust authentication in legal and financial applications.

Method: MultiScale Signature feature learning Network (MS-SigNet) with novel co-tuplet loss that learns global and regional features from multiple spatial scales, focusing on multiple positive and negative examples to overcome metric learning limitations.

Result: Promising performance demonstrated on four benchmark datasets across different languages, outperforming state-of-the-art approaches in distinguishing genuine signatures from skilled forgeries.

Conclusion: The proposed MS-SigNet with co-tuplet loss effectively addresses signature verification challenges and shows strong cross-language performance, supported by a new large-scale Chinese signature dataset (HanSig) for future research.

Abstract: Handwritten signature verification, crucial for legal and financial institutions, faces challenges including inter-writer similarity, intra-writer variations, and limited signature samples. To address these, we introduce the MultiScale Signature feature learning Network (MS-SigNet) with the co-tuplet loss, a novel metric learning loss designed for offline handwritten signature verification. MS-SigNet learns both global and regional signature features from multiple spatial scales, enhancing feature discrimination. This approach effectively distinguishes genuine signatures from skilled forgeries by capturing overall strokes and detailed local differences. The co-tuplet loss, focusing on multiple positive and negative examples, overcomes the limitations of typical metric learning losses by addressing inter-writer similarity and intra-writer variations and emphasizing informative examples. The code is available at https://github.com/ashleyfhh/MS-SigNet. We also present HanSig, a large-scale Chinese signature dataset to support robust system development for this language. The dataset is accessible at https://github.com/hsinmin/HanSig. Experimental results on four benchmark datasets in different languages demonstrate the promising performance of our method in comparison to state-of-the-art approaches.

[455] Vehicle-to-Everything Cooperative Perception for Autonomous Driving

Tao Huang, Jianan Liu, Xi Zhou, Dinh C. Nguyen, Mostafa Rahimi Azghadi, Yuxuan Xia, Qing-Long Han, Sumei Sun

Main category: cs.CV

TL;DR: A comprehensive survey of vehicle-to-everything (V2X) cooperative perception techniques for autonomous driving, covering mathematical models, key enabling techniques, current challenges, and future research directions.

DetailsMotivation: To enhance autonomous driving safety and efficiency by overcoming individual vehicle sensing limitations through cooperative perception that extends perception range, increases detection accuracy, and supports robust decision-making in complex environments.

Method: Provides a comprehensive survey of recent developments, introduces mathematical models for different collaboration strategies, examines key techniques including agent selection, data alignment, and feature fusion, and discusses major challenges and research directions.

Result: The paper systematically organizes and analyzes the state-of-the-art in V2X cooperative perception, identifying technical approaches for reliable perception sharing and highlighting critical challenges that need to be addressed.

Conclusion: V2X cooperative perception is crucial for autonomous driving advancement, with promising future directions including privacy-preserving AI methods, collaborative intelligence, and integrated sensing frameworks to overcome current limitations and support further development.

Abstract: Achieving fully autonomous driving with enhanced safety and efficiency relies on vehicle-to-everything cooperative perception, which enables vehicles to share perception data, thereby enhancing situational awareness and overcoming the limitations of the sensing ability of individual vehicles. Vehicle-to-everything cooperative perception plays a crucial role in extending the perception range, increasing detection accuracy, and supporting more robust decision-making and control in complex environments. This paper provides a comprehensive survey of recent developments in vehicle-to-everything cooperative perception, introducing mathematical models that characterize the perception process under different collaboration strategies. Key techniques for enabling reliable perception sharing, such as agent selection, data alignment, and feature fusion, are examined in detail. In addition, major challenges are discussed, including differences in agents and models, uncertainty in perception outputs, and the impact of communication constraints such as transmission delay and data loss. The paper concludes by outlining promising research directions, including privacy-preserving artificial intelligence methods, collaborative intelligence, and integrated sensing frameworks to support future advancements in vehicle-to-everything cooperative perception.

[456] LvBench: A Benchmark for Long-form Video Understanding with Versatile Multi-modal Question Answering

Hongjie Zhang, Lu Dong, Yi Liu, Yifei Huang, Yali Wang, Limin Wang, Yu Qiao

Main category: cs.CV

TL;DR: LvBench is a new long-form video understanding benchmark with extended video durations (70s-4h), diverse question types using both video frames and subtitles, and high-quality manual annotations from 100 movies.

DetailsMotivation: Existing long-form VideoQA datasets fail to meet genuine long-form understanding criteria due to short videos, limited sub-clips as clues, and restricted question types/modalities.

Method: Created LvBench with 20,061 QA pairs from 100 movies across genres, featuring extended temporal durations (single-scene to full-scene contexts), six diverse question types using video frames and subtitles, and rigorous human annotation.

Result: Analysis shows all existing methods’ performance significantly deteriorates as video and clue length increases, demonstrating the challenge of genuine long-form video understanding.

Conclusion: LvBench serves as a valuable benchmark for future long-form video understanding research, addressing limitations of previous datasets through extended durations, diverse modalities, and high-quality annotations.

Abstract: Despite remarkable recent progress, existing long-form VideoQA datasets fall short of meeting the criteria for genuine long-form video understanding. This is primarily due to the use of short videos for question curation, and the reliance on limited-length sub-clips as clues to answer those questions. Meanwhile, previous datasets have limited focus on question type and modality. To remedy this, we introduce LvBench, a Long-form video understanding benchmark for versatile multi-modal question-answering. Our LvBench stands out from existing long-form VideoQA datasets through three key characteristics: 1) Extended temporal durations: We consider videos ranging from 70 seconds to 4 hours, covering single-scene, multi-scene, and full-scene contexts. This design accounts for both video and clue lengths, capturing diverse contextual dynamics. 2) Diverse question types and modalities: LvBench introduces six distinct question types that evaluate various perceptual and cognitive capabilities, utilizing both video frames and subtitles. 3) High-quality annotations: We employ rigorous manual labeling by human annotators. Our dataset comprises 20,061 question-answer pairs sourced from 100 carefully selected movies across diverse genres, annotated collaboratively by multiple individuals. Analysis involving various baselines reveals a consistent trend: the performance of all existing methods significantly deteriorates when video and clue length increases. We expect LvBench to serve as a valuable resource for future works on long-form video understanding.

[457] Supervised Embedded Methods for Hyperspectral Band Selection

Yaniv Zimmer, Ofir Lindenbaum, Oren Glickman

Main category: cs.CV

TL;DR: Novel supervised embedded methods for hyperspectral imaging band selection that integrate directly into deep learning models, eliminating separate preprocessing and achieving state-of-the-art performance with minimal bands.

DetailsMotivation: Hyperspectral imaging's high dimensionality poses computational challenges for real-time applications, and existing band selection methods lack integration with downstream tasks and require separate preprocessing.

Method: Two novel supervised, embedded methods that integrate band selection directly into deep learning training pipelines, ensuring alignment with target tasks without separate preprocessing steps.

Result: Extensive experiments on three remote sensing benchmarks and an autonomous driving dataset show state-of-the-art performance while selecting only a minimal number of bands.

Conclusion: The proposed methods demonstrate the potential of efficient, task-specific hyperspectral imaging pipelines for practical deployment in resource-constrained settings.

Abstract: Hyperspectral Imaging (HSI) captures rich spectral information across contiguous wavelength bands, supporting applications in precision agriculture, environmental monitoring, and autonomous driving. However, its high dimensionality poses computational challenges, particularly in real-time or resource-constrained settings. While prior band selection methods attempt to reduce complexity, they often rely on separate preprocessing steps and lack alignment with downstream tasks. We propose two novel supervised, embedded methods for task-specific HSI band selection that integrate directly into deep learning models. By embedding band selection within the training pipeline, our methods eliminate the need for separate preprocessing and ensure alignment with the target task. Extensive experiments on three remote sensing benchmarks and an autonomous driving dataset show that our methods achieve state-of-the-art performance while selecting only a minimal number of bands. These results highlight the potential of efficient, task-specific HSI pipelines for practical deployment.

[458] DivCon: Divide and Conquer for Complex Numerical and Spatial Reasoning in Text-to-Image Generation

Yuhao Jia, Wenhan Tan

Main category: cs.CV

TL;DR: A divide-and-conquer approach for text-to-image generation that decouples layout prediction into numerical/spatial reasoning and visual planning, enabling lightweight LLMs to achieve comparable accuracy to large models while handling complex multi-object prompts.

DetailsMotivation: Existing T2I methods rely on closed-source large LLMs for layout prediction, limiting accessibility and scalability. They struggle with complex prompts containing multiple objects and spatial relationships.

Method: Divides generation into subtasks: 1) Layout prediction split into numerical/spatial reasoning and bounding box visual planning, 2) Layout-to-image generation synthesizes objects from easy to difficult ones in two steps.

Result: Outperforms previous approaches on HRS and NSR-1K benchmarks with notable margins. Visual results and user studies show significant improvement in perceptual quality for complex multi-object prompts.

Conclusion: The divide-and-conquer approach enables lightweight LLMs to achieve layout accuracy comparable to large-scale models while significantly improving performance on complex text-to-image generation tasks with multiple objects.

Abstract: Diffusion-driven text-to-image (T2I) generation has achieved remarkable advancements in recent years. To further improve T2I models’ capability in numerical and spatial reasoning, layout is employed as an intermedium to bridge large language models and layout-based diffusion models. However, these methods often rely on closed-source, large-scale LLMs for layout prediction, limiting accessibility and scalability. They also struggle with generating images from prompts with multiple objects and complicated spatial relationships. To tackle these challenges, we introduce a divide-and-conquer approach which decouples the generation task into multiple subtasks. First, the layout prediction stage is divided into numerical & spatial reasoning and bounding box visual planning, enabling even lightweight LLMs to achieve layout accuracy comparable to large-scale models. Second, the layout-to-image generation stage is divided into two steps to synthesize objects from easy ones to difficult ones. Experiments are conducted on the HRS and NSR-1K benchmarks and our method outperforms previous approaches with notable margins. In addition, visual results and user study demonstrate that our approach significantly improves the perceptual quality, especially when generating multiple objects from complex textural prompts.

[459] SC-Diff: 3D Shape Completion with Latent Diffusion Models

Simon Schaefer, Juan D. Galvis, Xingxing Zuo, Stefan Leutengger

Main category: cs.CV

TL;DR: A 3D shape completion framework using latent diffusion with multimodal conditioning from 2D images and 3D partial scans, achieving efficient high-resolution processing and superior reconstruction quality.

DetailsMotivation: To address the limitations of existing 3D shape completion methods by unifying 2D and 3D information through multimodal conditioning, enabling more robust and efficient completion in real-world scenarios.

Method: Uses Truncated Signed Distance Functions (TSDFs) encoded into a discrete latent space with joint 2D/3D supervision, employs latent diffusion model with multimodal conditioning, and trains with simulated partial observations.

Result: Reduces GPU memory usage by 30%, outperforms class-specific models by 12% and class-agnostic models by 47% in reconstruction error, produces more diverse and high-fidelity completions.

Conclusion: The framework successfully unifies multimodal conditioning for 3D shape completion, demonstrating significant improvements in efficiency, generalization, and reconstruction quality over existing approaches.

Abstract: We present a novel 3D shape completion framework that unifies multimodal conditioning, leveraging both 2D images and 3D partial scans through a latent diffusion model. Shapes are represented as Truncated Signed Distance Functions (TSDFs) and encoded into a discrete latent space jointly supervised by 2D and 3D cues, enabling efficient high-resolution processing while reducing GPU memory usage by 30% compared to state-of-the-art methods. Our approach guides the generation process with flexible multimodal conditioning, ensuring consistent integration of 2D and 3D information from encoding to reconstruction. Our training strategy simulates realistic partial observations, avoiding assumptions about input structure and improving robustness in real-world scenarios. Leveraging our efficient latent space and multimodal conditioning, our model generalizes across object categories, outperforming class-specific models by 12% and class-agnostic models by 47% in $l_1$ reconstruction error, while producing more diverse, realistic, and high-fidelity completions than prior approaches.

[460] TwinLiteNet+: An Enhanced Multi-Task Segmentation Model for Autonomous Driving

Quang-Huy Che, Duc-Tri Le, Minh-Quan Pham, Vinh-Tiep Nguyen, Duc-Khai Lam

Main category: cs.CV

TL;DR: TwinLiteNet+ is an efficient multi-task segmentation model for real-time drivable area and lane segmentation in autonomous driving, achieving SOTA performance with significantly reduced computational costs.

DetailsMotivation: Most state-of-the-art semantic segmentation models are computationally intensive and unsuitable for real-time deployment on resource-constrained embedded devices in autonomous driving systems.

Method: Uses hybrid encoder architecture with stride-based dilated convolutions and depthwise separable dilated convolutions, lightweight upsampling modules (UCB and USB), and Partial Class Activation Attention (PCAA) mechanism. Available in four configurations from ultra-compact to high-performance.

Result: TwinLiteNet+_Large achieves 92.9% mIoU for drivable area segmentation and 34.2% IoU for lane segmentation on BDD100K dataset, surpassing existing models with 11x fewer FLOPs. Demonstrates superior inference speed, quantization robustness, and energy efficiency on embedded devices.

Conclusion: TwinLiteNet+ is a compelling solution for real-world autonomous driving systems, offering high efficiency and performance suitable for resource-constrained embedded deployment.

Abstract: Semantic segmentation is a fundamental perception task in autonomous driving, particularly for identifying drivable areas and lane markings to enable safe navigation. However, most state-of-the-art (SOTA) models are computationally intensive and unsuitable for real-time deployment on resource-constrained embedded devices. In this paper, we introduce TwinLiteNet+, an enhanced multi-task segmentation model designed for real-time drivable area and lane segmentation with high efficiency. TwinLiteNet+ employs a hybrid encoder architecture that integrates stride-based dilated convolutions and depthwise separable dilated convolutions, balancing representational capacity and computational cost. To improve task-specific decoding, we propose two lightweight upsampling modules-Upper Convolution Block (UCB) and Upper Simple Block (USB)-alongside a Partial Class Activation Attention (PCAA) mechanism that enhances segmentation precision. The model is available in four configurations, ranging from the ultra-compact TwinLiteNet+{Nano} (34K parameters) to the high-performance TwinLiteNet+{Large} (1.94M parameters). On the BDD100K dataset, TwinLiteNet+_{Large} achieves 92.9% mIoU for drivable area segmentation and 34.2% IoU for lane segmentation-surpassing existing state-of-the-art models while requiring 11x fewer floating-point operations (FLOPs) for computation. Extensive evaluations on embedded devices demonstrate superior inference speed, quantization robustness (INT8/FP16), and energy efficiency, validating TwinLiteNet+ as a compelling solution for real-world autonomous driving systems. Code is available at https://github.com/chequanghuy/TwinLiteNetPlus.

[461] StylizedGS: Controllable Stylization for 3D Gaussian Splatting

Dingxi Zhang, Yu-Jie Yuan, Zhuoxun Chen, Fang-Lue Zhang, Zhenliang He, Shiguang Shan, Lin Gao

Main category: cs.CV

TL;DR: StylizedGS is an efficient 3D neural style transfer framework using 3D Gaussian Splatting that addresses efficiency issues and geometric pattern limitations of NeRF-based methods while providing flexible artistic control.

DetailsMotivation: Current NeRF-based 3D stylization methods suffer from efficiency problems affecting user experience, cannot accurately transfer geometric pattern styles due to their implicit nature, and lack flexible control capabilities needed for creative exploration.

Method: Uses 3D Gaussian Splatting representation with filter-based refinement to eliminate floaters, nearest neighbor-based style loss for stylization by fine-tuning geometry and color parameters, depth preservation loss to maintain geometry integrity, and specially designed losses for color, scale, and region control.

Result: Achieves high-quality stylization results with faithful brushstrokes and geometric consistency, demonstrating effectiveness and efficiency across various scenes and styles in both quality and inference speed.

Conclusion: StylizedGS provides an efficient and flexible solution for 3D neural style transfer that overcomes limitations of previous methods while enabling customizable artistic control over stylized scenes.

Abstract: As XR technology continues to advance rapidly, 3D generation and editing are increasingly crucial. Among these, stylization plays a key role in enhancing the appearance of 3D models. By utilizing stylization, users can achieve consistent artistic effects in 3D editing using a single reference style image, making it a user-friendly editing method. However, recent NeRF-based 3D stylization methods encounter efficiency issues that impact the user experience, and their implicit nature limits their ability to accurately transfer geometric pattern styles. Additionally, the ability for artists to apply flexible control over stylized scenes is considered highly desirable to foster an environment conducive to creative exploration. To address the above issues, we introduce StylizedGS, an efficient 3D neural style transfer framework with adaptable control over perceptual factors based on 3D Gaussian Splatting representation. We propose a filter-based refinement to eliminate floaters that affect the stylization effects in the scene reconstruction process. The nearest neighbor-based style loss is introduced to achieve stylization by fine-tuning the geometry and color parameters of 3DGS, while a depth preservation loss with other regularizations is proposed to prevent the tampering of geometry content. Moreover, facilitated by specially designed losses, StylizedGS enables users to control color, stylized scale, and regions during the stylization to possess customization capabilities. Our method achieves high-quality stylization results characterized by faithful brushstrokes and geometric consistency with flexible controls. Extensive experiments across various scenes and styles demonstrate the effectiveness and efficiency of our method concerning both stylization quality and inference speed.

[462] BRACTIVE: A Brain Activation Approach to Human Visual Brain Learning

Xuan-Bac Nguyen, Hojin Jang, Xin Li, Samee U. Khan, Pawan Sinha, Khoa Luu

Main category: cs.CV

TL;DR: BRACTIVE is a transformer-based framework that aligns visual features with fMRI brain signals to identify person-specific Regions of Interest across multiple subjects, improving both neuroscience research and deep neural network performance.

DetailsMotivation: Understanding how the human brain processes visual information can inspire better machine learning algorithms and architectures. Current brain research methods are limited to single subjects and cannot efficiently identify Regions of Interest across multiple individuals.

Method: Transformer-based framework called BRACTIVE that aligns visual features with functional MRI signals to automatically identify brain Regions of Interest across multiple subjects and various object categories.

Result: BRACTIVE effectively identifies person-specific regions like face and body-selective areas, aligns with neuroscience findings, and enhances deep neural network performance across various benchmarks when guided by human visual brain activity.

Conclusion: BRACTIVE shows strong potential for both neuroscience research and machine intelligence studies by enabling scalable ROI identification across multiple subjects and demonstrating that brain activity guidance improves neural network performance.

Abstract: The human brain is a highly efficient processing unit, and understanding how it works can inspire new algorithms and architectures in machine learning. In this work, we introduce a novel framework named Brain Activation Network (BRACTIVE), a transformer-based approach to studying the human visual brain. The primary objective of BRACTIVE is to align the visual features of subjects with their corresponding brain representations using functional Magnetic Resonance Imaging (fMRI) signals. It enables us to identify the brain’s Regions of Interest (ROIs) in the subjects. Unlike previous brain research methods, which can only identify ROIs for one subject at a time and are limited by the number of subjects, BRACTIVE automatically extends this identification to multiple subjects and ROIs. Our experiments demonstrate that BRACTIVE effectively identifies person-specific regions of interest, such as face and body-selective areas, aligning with neuroscience findings and indicating potential applicability to various object categories. More importantly, we found that leveraging human visual brain activity to guide deep neural networks enhances performance across various benchmarks. It encourages the potential of BRACTIVE in both neuroscience and machine intelligence studies.

[463] TE-NeXt: A LiDAR-Based 3D Sparse Convolutional Network for Traversability Estimation

Antonio Santo, Juan J. Cabrera, David Valiente, Carlos Viegas, Arturo Gil

Main category: cs.CV

TL;DR: TE-NeXt is a novel architecture for traversability estimation from sparse LiDAR point clouds using residual convolution blocks that combines attention mechanisms and 3D sparse convolutions, achieving state-of-the-art performance in semantic segmentation across urban and natural environments.

DetailsMotivation: To develop an efficient and generalizable architecture for traversability estimation that can perform well in both structured urban environments and unstructured natural terrains using sparse LiDAR data.

Method: Proposes TE-NeXt block that fuses attention mechanisms with 3D sparse convolutions in a residual convolution framework, trained and evaluated on SemanticKITTI, Rellis-3D, and SemanticUSL datasets.

Result: Outperforms state-of-the-art methods in semantic segmentation, demonstrating better performance in unstructured environments while maintaining high reliability and robustness in urban settings.

Conclusion: TE-NeXt provides superior abstraction capabilities for traversability estimation and the implementation is made publicly available to ensure reproducibility of results for the scientific community.

Abstract: This paper presents TE-NeXt, a novel and efficient architecture for Traversability Estimation (TE) from sparse LiDAR point clouds based on a residual convolution block. TE-NeXt block fuses notions of current trends such as attention mechanisms and 3D sparse convolutions. TE-NeXt aims to demonstrate high capacity for generalisation in a variety of urban and natural environments, using well-known and accessible datasets such as SemanticKITTI, Rellis-3D and SemanticUSL. Thus, the designed architecture ouperforms state-of-the-art methods in the problem of semantic segmentation, demonstrating better results in unstructured environments and maintaining high reliability and robustness in urbans environments, which leads to better abstraction. Implementation is available in a open repository to the scientific community with the aim of ensuring the reproducibility of results.

[464] Story Generation from Visual Inputs: Techniques, Related Tasks, and Challenges

Daniel A. P. Oliveira, Eugénio Ribeiro, David Martins de Matos

Main category: cs.CV

TL;DR: Survey paper on visual story generation methodologies, covering principles, strengths, limitations, related tasks, datasets, and evaluation metrics.

DetailsMotivation: To address the need for automated narrative creation from visual data for digital media consumption, assistive technologies, and interactive entertainment.

Method: Comprehensive survey analysis of existing methodologies, including examination of related tasks like image/video captioning, visual question answering, and non-visual story generation techniques.

Result: Provides critical analysis of current approaches, datasets, and evaluation metrics, highlighting their limitations and common challenges in the field.

Conclusion: The survey offers a systematic overview of visual story generation techniques and identifies key challenges and limitations that need to be addressed for future advancements in the field.

Abstract: Creating engaging narratives from visual data is crucial for automated digital media consumption, assistive technologies, and interactive entertainment. This survey covers methodologies used in the generation of these narratives, focusing on their principles, strengths, and limitations. The survey also covers tasks related to automatic story generation, such as image and video captioning, and visual question answering, as well as story generation without visual inputs. These tasks share common challenges with visual story generation and have served as inspiration for the techniques used in the field. We analyze the main datasets and evaluation metrics, providing a critical perspective on their limitations.

[465] Towards Human-Level 3D Relative Pose Estimation: Generalizable, Training-Free, with Single Reference

Yuan Gao, Yajing Luo, Junhong Wang, Kui Jia, Gui-Song Xia

Main category: cs.CV

TL;DR: A training-free method for 3D relative pose estimation that uses RGB-D reference images, differentiable rendering, and semantic cues from DINOv2 to estimate object poses without requiring labeled data or model training.

DetailsMotivation: Humans can easily deduce relative object poses from single image pairs without training, leveraging 3D shape perception, render-and-compare simulation, and semantic awareness. The paper aims to replicate this capability computationally.

Method: Uses 2.5D shape from RGB-D reference, differentiable renderer to generate rotated views, and semantic maps from DINOv2. Compares rendered RGB and semantic maps with query images to refine 3D relative pose through gradient backpropagation.

Result: Outperforms state-of-the-art supervised methods on LineMOD, LM-O, and YCB-V datasets, especially under rigorous Acc@5/10/15° metrics and challenging cross-dataset settings.

Conclusion: The method enables zero-shot relative pose estimation for unseen objects using only a single RGB-D reference, demonstrating superior performance without requiring training or labeling.

Abstract: Humans can easily deduce the relative pose of a previously unseen object, without labeling or training, given only a single query-reference image pair. This is arguably achieved by incorporating i) 3D/2.5D shape perception from a single image, ii) render-and-compare simulation, and iii) rich semantic cue awareness to furnish (coarse) reference-query correspondence. Motivated by this, we propose a novel 3D generalizable relative pose estimation method by elaborating 3D/2.5D shape perception with a 2.5D shape from an RGB-D reference, fulfilling the render-and-compare paradigm with an off-the-shelf differentiable renderer, and leveraging the semantic cues from a pretrained model like DINOv2. Specifically, our differentiable renderer takes the 2.5D rotatable mesh textured by the RGB and the semantic maps (obtained by DINOv2 from the RGB input), then renders new RGB and semantic maps (with back-surface culling) under a novel rotated view. The refinement loss comes from comparing the rendered RGB and semantic maps with the query ones, back-propagating the gradients through the differentiable renderer to refine the 3D relative pose. As a result, \emph{our method can be readily applied to unseen objects, given only a single RGB-D reference, without labeling or training}. Extensive experiments on LineMOD, LM-O, and YCB-V show that our training-free method significantly outperforms the state-of-the-art supervised methods, especially under the rigorous \texttt{Acc@5/10/15}$^\circ$ metrics and the challenging cross-dataset settings.

[466] Multimodal Conditional 3D Face Geometry Generation

Christopher Otto, Prashanth Chandran, Sebastian Weiss, Markus Gross, Gaspard Zoss, Derek Bradley

Main category: cs.CV

TL;DR: A multimodal 3D face generation method that accepts various input types (sketches, photos, edges, parameters, landmarks, text) to produce topology-consistent 3D face geometry through diffusion in UV space with cross-attention conditioning.

DetailsMotivation: To create a user-friendly 3D face generation tool that provides fine-grained control over identity and expression through multiple conditioning signals, making 3D face modeling more accessible and versatile.

Method: Uses a diffusion process in 2D parameterized UV domain with IP-Adapter cross-attention layers for each conditioning signal type (sketches, photos, edges, FLAME parameters, landmarks, text).

Result: Produces topology-consistent, high-quality 3D face geometry from diverse input modalities within a single unified model.

Conclusion: The approach enables easy-to-use 3D face generation with multiple input options and fine control, demonstrating versatility across artistic, photographic, parametric, and textual inputs.

Abstract: We present a new method for multimodal conditional 3D face geometry generation that allows user-friendly control over the output identity and expression via a number of different conditioning signals. Within a single model, we demonstrate 3D faces generated from artistic sketches, portrait photos, Canny edges, FLAME face model parameters, 2D face landmarks, or text prompts. Our approach is based on a diffusion process that generates 3D geometry in a 2D parameterized UV domain. Geometry generation passes each conditioning signal through a set of cross-attention layers (IP-Adapter), one set for each user-defined conditioning signal. The result is an easy-to-use 3D face generation tool that produces topology-consistent, high-quality geometry with fine-grain user control.

[467] FeatureSORT: Essential Features for Effective Tracking

Hamidreza Hashempoor, Rosemary Koikara, Yu Dong Hwang

Main category: cs.CV

TL;DR: FeatureSORT enhances DeepSORT with multi-attribute detection (clothing color/style, motion direction) and stronger post-processing, achieving state-of-the-art MOT performance.

DetailsMotivation: To improve online multiple object tracking by addressing limitations of conventional detectors that only provide bounding boxes, and to handle occlusions better while reducing identity switches.

Method: Modified YOLOX detector outputs multiple appearance attributes + ReID network for complementary embeddings. Uses joint distance function considering IoU, direction, color, style, and ReID similarity. Incorporates global linking and Gaussian Smoothing Process interpolation.

Result: Achieves MOTA scores: 79.7 on MOT16, 80.6 on MOT17, 77.9 on MOT20, and 92.2 on DanceTrack - state-of-the-art online performance.

Conclusion: Feature-enriched detection and modular post-processing significantly advance multi-object tracking, maintaining consistent identities through occlusions with reduced identity switches.

Abstract: We introduce FeatureSORT, a simple yet effective online multiple object tracker that reinforces the DeepSORT baseline with a redesigned detector and additional feature cues. In contrast to conventional detectors that only provide bounding boxes, our modified YOLOX architecture is extended to output multiple appearance attributes, including clothing color, clothing style, and motion direction, alongside the bounding boxes. These feature cues, together with a ReID network, form complementary embeddings that substantially improve association accuracy. Furthermore, we incorporate stronger post-processing strategies, such as global linking and Gaussian Smoothing Process interpolation, to handle missing associations and detections. During online tracking, we define a measurement-to-track distance function that jointly considers IoU, direction, color, style, and ReID similarity. This design enables FeatureSORT to maintain consistent identities through longer occlusions while reducing identity switches. Extensive experiments on standard MOT benchmarks demonstrate that FeatureSORT achieves state-of-the-art online performance, with MOTA scores of 79.7 on MOT16, 80.6 on MOT17, 77.9 on MOT20, and 92.2 on DanceTrack, underscoring the effectiveness of feature-enriched detection and modular post processing in advancing multi-object tracking.

[468] Muzzle-Based Cattle Identification System Using Artificial Intelligence (AI)

Hasan Zohirul Islam, Safayet Khan, Sanjib Kumar Paul, Sheikh Imtiaz Rahi, Fahim Hossain Sifat, Md. Mahadi Hasan Sany, Md. Shahjahan Ali Sarker, Tareq Anam, Ismail Hossain Polas

Main category: cs.CV

TL;DR: Developed a muzzle-based cattle identification system using machine learning to enable livestock insurance by providing tamper-proof identification, achieving 96.5% accuracy.

DetailsMotivation: Absence of reliable cattle identification technology prevented insurance companies from offering livestock insurance, causing financial hardship for marginal farmers who couldn't claim compensation for cattle deaths.

Method: Used YOLO algorithm for muzzle detection and FaceNet architecture for learning embeddings from muzzle images. Applied CLAHE with sharpening filters for preprocessing. Collected 32,374 images from 826 cattle.

Result: Achieved 96.489% accuracy, 97.334% F1 score, 87.993% true positive rate at extremely low false positive rate of 0.098%.

Conclusion: The system provides reliable and efficient cattle identification that can significantly advance livestock insurance and precision farming.

Abstract: Absence of tamper-proof cattle identification technology was a significant problem preventing insurance companies from providing livestock insurance. This lack of technology had devastating financial consequences for marginal farmers as they did not have the opportunity to claim compensation for any unexpected events such as the accidental death of cattle in Bangladesh. Using machine learning and deep learning algorithms, we have solved the bottleneck of cattle identification by developing and introducing a muzzle-based cattle identification system. The uniqueness of cattle muzzles has been scientifically established, which resembles human fingerprints. This is the fundamental premise that prompted us to develop a cattle identification system that extracts the uniqueness of cattle muzzles. For this purpose, we collected 32,374 images from 826 cattle. Contrast-limited adaptive histogram equalization (CLAHE) with sharpening filters was applied in the preprocessing steps to remove noise from images. We used the YOLO algorithm for cattle muzzle detection in the image and the FaceNet architecture to learn unified embeddings from muzzle images using squared $L_2$ distances. Our system performs with an accuracy of $96.489%$, $F_1$ score of $97.334%$, and a true positive rate (tpr) of $87.993%$ at a remarkably low false positive rate (fpr) of $0.098%$. This reliable and efficient system for identifying cattle can significantly advance livestock insurance and precision farming.

[469] Scalable Frame Sampling for Video Classification: A Semi-Optimal Policy Approach with Reduced Search Space

Junho Lee, Jeongwoo Shin, Seung Woo Ko, Seongsu Ha, Joonseok Lee

Main category: cs.CV

TL;DR: Proposes a semi-optimal frame sampling policy that reduces search space from O(T^N) to O(T) by selecting top N frames based on independently estimated per-frame confidence values.

DetailsMotivation: Existing frame sampling methods suffer from vast search space of binom(T,N) when selecting N frames from T total frames, especially when N is large, making brute-force search and other approaches computationally infeasible.

Method: Introduces a semi-optimal policy that independently estimates the value of each frame using per-frame confidence scores, then selects the top N frames based on these individual estimates rather than exploring the entire combinatorial space.

Result: The method efficiently approximates the optimal policy, particularly in practical settings, and demonstrates stable high performance across various datasets and model architectures regardless of N and T sizes.

Conclusion: The proposed semi-optimal frame sampling policy provides a computationally efficient solution that maintains performance while dramatically reducing the search complexity from exponential to linear in the number of frames.

Abstract: Given a video with $T$ frames, frame sampling is a task to select $N \ll T$ frames, so as to maximize the performance of a fixed video classifier. Not just brute-force search, but most existing methods suffer from its vast search space of $\binom{T}{N}$, especially when $N$ gets large. To address this challenge, we introduce a novel perspective of reducing the search space from $O(T^N)$ to $O(T)$. Instead of exploring the entire $O(T^N)$ space, our proposed semi-optimal policy selects the top $N$ frames based on the independently estimated value of each frame using per-frame confidence, significantly reducing the computational complexity. We verify that our semi-optimal policy can efficiently approximate the optimal policy, particularly under practical settings. Additionally, through extensive experiments on various datasets and model architectures, we demonstrate that learning our semi-optimal policy ensures stable and high performance regardless of the size of $N$ and $T$.

[470] DivScene: Towards Open-Vocabulary Object Navigation with Large Vision Language Models in Diverse Scenes

Zhaowei Wang, Hongming Zhang, Tianqing Fang, Ye Tian, Yue Yang, Kaixin Ma, Xiaoman Pan, Yangqiu Song, Dong Yu

Main category: cs.CV

TL;DR: LVLMs show promise for embodied navigation but lack open-vocabulary object navigation capabilities. The paper introduces DivScene dataset with 4,614 houses and 5,707 object types, then fine-tunes LVLMs with CoT explanations using BFS-generated paths, achieving >20% improvement over GPT-4o.

DetailsMotivation: Large Vision-Language Models have advanced in visual tasks but their ability to comprehend embodied environments and navigate within them remains underexplored, particularly for open-vocabulary object navigation.

Method: Introduced DivScene dataset with 4,614 houses across 81 scene types and 5,707 target objects. Fine-tuned LVLMs to predict next actions with Chain-of-Thought explanations using BFS-generated shortest paths without human supervision.

Result: Current models fall short on open-vocab object navigation. Fine-tuned LVLMs achieved substantial improvement, surpassing GPT-4o by over 20% in success rates using only BFS-generated paths.

Conclusion: LVLMs’ navigation capabilities can be significantly enhanced through fine-tuning with automatically generated paths, demonstrating strong potential for embodied AI applications without requiring human supervision.

Abstract: Large Vision-Language Models (LVLMs) have achieved significant progress in tasks like visual question answering and document understanding. However, their potential to comprehend embodied environments and navigate within them remains underexplored. In this work, we first study the challenge of open-vocabulary object navigation by introducing DivScene, a large-scale dataset with 4,614 houses across 81 scene types and 5,707 kinds of target objects. Our dataset provides a much greater diversity of target objects and scene types than existing datasets, enabling a comprehensive task evaluation. We evaluated various methods with LVLMs and LLMs on our dataset and found that current models still fall short of open-vocab object navigation ability. Then, we fine-tuned LVLMs to predict the next action with CoT explanations. We observe that LVLM’s navigation ability can be improved substantially with only BFS-generated shortest paths without any human supervision, surpassing GPT-4o by over 20% in success rates.

[471] A low complexity contextual stacked ensemble-learning approach for pedestrian intent prediction

Chia-Yen Chiang, Yasmin Fathy, Gregory Slabaugh, Mona Jaber

Main category: cs.CV

TL;DR: A low-complexity ensemble learning approach using skeletonization and contextual data achieves state-of-the-art pedestrian crossing intent prediction with 99.7% reduced computational complexity.

DetailsMotivation: Accurate pedestrian crossing intention prediction is crucial for autonomous and advanced driver-assisted vehicles to avoid collisions, but current computer vision methods require high computation power for reliable results.

Method: Proposes an ensemble-learning approach that first detects pedestrians, compresses images using skeletonization, adds contextual information, and uses a stacked ensemble-learning model for intent prediction.

Result: Achieves similar pedestrian intent prediction performance as state-of-the-art approaches while reducing computational complexity by 99.7% across different datasets.

Conclusion: The proposed low-complexity method effectively balances prediction accuracy and computational efficiency, making it suitable for real-time applications in autonomous vehicles.

Abstract: Walking as a form of active travel is essential in promoting sustainable transport. It is thus crucial to accurately predict pedestrian crossing intention and avoid collisions, especially with the advent of autonomous and advanced driver-assisted vehicles. Current research leverages computer vision and machine learning advances to predict near-misses; however, this often requires high computation power to yield reliable results. In contrast, this work proposes a low-complexity ensemble-learning approach that employs contextual data for predicting the pedestrian’s intent for crossing. The pedestrian is first detected, and their image is then compressed using skeleton-ization, and contextual information is added into a stacked ensemble-learning approach. Our experiments on different datasets achieve similar pedestrian intent prediction performance as the state-of-the-art approaches with 99.7% reduction in computational complexity. Our source code and trained models will be released upon paper acceptance

[472] Synergistic Dual Spatial-aware Generation of Image-to-Text and Text-to-Image

Yu Zhao, Hao Fei, Xiangtai Li, Libo Qin, Jiayi Ji, Hongyuan Zhu, Meishan Zhang, Min Zhang, Jianguo Wei

Main category: cs.CV

TL;DR: Proposes Spatial Dual Discrete Diffusion (SD^3) framework that jointly models spatial image-to-text (SI2T) and spatial text-to-image (ST2I) tasks using 3D scene graphs and dual learning to improve 3D spatial understanding.

DetailsMotivation: Existing standalone methods for SI2T or ST2I perform poorly in spatial understanding due to difficulty in 3D spatial feature modeling. The authors recognize that these tasks appear in dual form and should benefit from joint modeling.

Method: Uses a dual learning framework with shared 3D scene graph (3DSG) representation. Proposes SD^3 framework that leverages intermediate features from easier 3D→image and 3D→text processes to guide the harder image→3D and text→3D processes.

Result: Outperforms mainstream T2I and I2T methods significantly on the VSD dataset. In-depth analysis shows the dual learning strategy advances performance.

Conclusion: Joint modeling of SI2T and ST2I tasks with shared 3D scene graph representation and dual diffusion framework significantly improves spatial understanding performance compared to standalone approaches.

Abstract: In the visual spatial understanding (VSU) area, spatial image-to-text (SI2T) and spatial text-to-image (ST2I) are two fundamental tasks that appear in dual form. Existing methods for standalone SI2T or ST2I perform imperfectly in spatial understanding, due to the difficulty of 3D-wise spatial feature modeling. In this work, we consider modeling the SI2T and ST2I together under a dual learning framework. During the dual framework, we then propose to represent the 3D spatial scene features with a novel 3D scene graph (3DSG) representation that can be shared and beneficial to both tasks. Further, inspired by the intuition that the easier 3D$\to$image and 3D$\to$text processes also exist symmetrically in the ST2I and SI2T, respectively, we propose the Spatial Dual Discrete Diffusion (SD$^3$) framework, which utilizes the intermediate features of the 3D$\to$X processes to guide the hard X$\to$3D processes, such that the overall ST2I and SI2T will benefit each other. On the visual spatial understanding dataset VSD, our system outperforms the mainstream T2I and I2T methods significantly. Further in-depth analysis reveals how our dual learning strategy advances.

[473] OSMLoc: Single Image-Based Visual Localization in OpenStreetMap with Fused Geometric and Semantic Guidance

Youqi Liao, Xieyuanli Chen, Shuhao Kang, Jianping Li, Zhen Dong, Hongchao Fan, Bisheng Yang

Main category: cs.CV

TL;DR: OSMLoc is a brain-inspired visual localization method that combines semantic and geometric guidance to match first-person camera images with OpenStreetMap data, achieving superior accuracy and robustness across diverse conditions.

DetailsMotivation: The disparity between camera imagery and vectorized map data limits effective matching for visual localization. Inspired by human brain's fusion of geometric and semantic understanding for spatial tasks, the authors aim to bridge this modality gap.

Method: 1) Uses visual foundational models for image feature extraction 2) Geometry-guided depth distribution adapter bridges monocular depth estimation and camera-to-BEV transform 3) Leverages semantic embeddings from OSM as auxiliary guidance for feature matching

Result: Superior performance demonstrated on MGL dataset, cross-area/cross-condition benchmark, and KITTI dataset. The method shows improved accuracy, robustness, and generalization capability.

Conclusion: OSMLoc effectively integrates semantic and geometric guidance to overcome modality disparities between visual observations and map data, enabling more accurate and robust visual localization using OpenStreetMap.

Abstract: OpenStreetMap (OSM), a rich and versatile source of volunteered geographic information (VGI), facilitates human self-localization and scene understanding by integrating nearby visual observations with vectorized map data. However, the disparity in modalities and perspectives poses a major challenge for effectively matching camera imagery with compact map representations, thereby limiting the full potential of VGI data in real-world localization applications. Inspired by the fact that the human brain relies on the fusion of geometric and semantic understanding for spatial localization tasks, we propose the OSMLoc in this paper. OSMLoc is a brain-inspired visual localization approach based on first-person-view images against the OSM maps. It integrates semantic and geometric guidance to significantly improve accuracy, robustness, and generalization capability. First, we equip the OSMLoc with the visual foundational model to extract powerful image features. Second, a geometry-guided depth distribution adapter is proposed to bridge the monocular depth estimation and camera-to-BEV transform. Thirdly, the semantic embeddings from the OSM data are utilized as auxiliary guidance for image-to-OSM feature matching. To validate the proposed OSMLoc, we collect a worldwide cross-area and cross-condition (CC) benchmark for extensive evaluation. Experiments on the MGL dataset, CC validation benchmark, and KITTI dataset have demonstrated the superiority of our method. Code, pre-trained models, CC validation benchmark, and additional results are available at: https://github.com/WHU-USI3DV/OSMLoc.

[474] ZoomEye: Enhancing Multimodal LLMs with Human-Like Zooming Capabilities through Tree-Based Image Exploration

Haozhan Shen, Kangjia Zhao, Tiancheng Zhao, Ruochen Xu, Zilun Zhang, Mingwei Zhu, Jianwei Yin

Main category: cs.CV

TL;DR: Zoom Eye is a training-free tree search algorithm that enables MLLMs to perform vision-level reasoning by dynamically zooming into specific image regions, significantly improving performance on high-resolution benchmarks.

DetailsMotivation: Existing MLLM reasoning approaches are text-level and keep visual input fixed, limiting their ability to exploit rich visual information, especially for images with fine-grained elements where vision-level reasoning is crucial.

Method: Proposes Zoom Eye, a model-agnostic tree search algorithm that treats images as hierarchical trees where child nodes represent zoomed-in sub-regions. MLLMs navigate from root to leaf nodes to find task-relevant visual evidence, simulating human-like zooming behavior.

Result: Zoom Eye consistently improves multiple MLLMs by large margins (e.g., InternVL2.5-8B increases by 15.71% and 17.69% on HR-Bench) and enables small 3-8B MLLMs to outperform strong large models like GPT-4o.

Conclusion: The proposed vision-level reasoning approach through hierarchical zooming significantly enhances MLLM performance on high-resolution visual tasks, demonstrating the importance of dynamic visual exploration over static text-level reasoning.

Abstract: Multimodal Large Language Models (MLLMs) have demonstrated impressive capabilities in vision-language understanding. Recently, with the integration of test-time scaling techniques, these models have also shown strong potential in visual reasoning. However, most existing reasoning approaches remain text-level in nature: MLLMs are prompted to explore various combinations of textual tokens via their underlying language model, while the visual input remains fixed throughout the reasoning process. This paradigm limits the model’s ability to fully exploit rich visual information, particularly when dealing with images containing numerous fine-grained elements. In such cases, vision-level reasoning becomes crucial - where models dynamically zoom into specific regions of the image to gather detailed visual cues necessary for accurate decision-making. In this paper, we propose Zoom Eye, a training-free, model-agnostic tree search algorithm tailored for vision-level reasoning. Zoom Eye treats an image as a hierarchical tree structure, where each child node represents a zoomed-in sub-region of its parent, and the root corresponds to the full image. The algorithm enables MLLMs to simulate human-like zooming behavior by navigating from root to leaf nodes in search of task-relevant visual evidence. We experiment on a series of high-resolution benchmarks and the results demonstrate that Zoom Eye consistently improves the performance of multiple MLLMs by a large margin (e.g., InternVL2.5-8B increases by 15.71% and 17.69% on HR-Bench) and also enables small 3-8B MLLMs to outperform strong large models such as GPT-4o. Code: https://github.com/om-ai-lab/ZoomEye

[475] EFTViT: Efficient Federated Training of Vision Transformers with Masked Images on Resource-Constrained Clients

Meihan Wu, Tao Chang, Cui Miao, Jie Zhou, Chun Li, Xiangyu Xu, Ming Li, Xiaodong Wang

Main category: cs.CV

TL;DR: EFT-ViT enables efficient full-parameter Vision Transformer training on resource-constrained edge devices through masked image patches and hierarchical federated learning, achieving significant accuracy improvements while reducing computational costs and enhancing privacy.

DetailsMotivation: Vision Transformers (ViTs) outperform CNNs but require more computational resources, making efficient federated training on edge devices challenging. Current research lacks solutions for resource-constrained ViT training in federated settings.

Method: Proposes EFTViT framework with masked image patches - randomly excluding patches during training to reduce computation. Uses hierarchical structure with lightweight local modules on clients and larger global module on server. Implements median sampling strategy to balance intermediate features and protect data privacy.

Result: Achieves up to 28.17% accuracy improvement, reduces local training computational cost by 2.8x, and cuts local training time by 4.4x compared to existing methods on popular benchmarks.

Conclusion: EFTViT successfully enables efficient full-parameter ViT training on resource-constrained devices while maintaining performance, reducing computational overhead, and enhancing privacy protection through masked patch training and hierarchical federated architecture.

Abstract: Federated learning research has recently shifted from Convolutional Neural Networks (CNNs) to Vision Transformers (ViTs) due to their superior capacity. ViTs training demands higher computational resources due to the lack of 2D inductive biases inherent in CNNs. However, efficient federated training of ViTs on resource-constrained edge devices remains unexplored in the community. In this paper, we propose EFTViT, a hierarchical federated framework that leverages masked images to enable efficient, full-parameter training on resource-constrained edge devices, offering substantial benefits for learning on heterogeneous data. In general, we patchify images and randomly mask a portion of the patches, observing that excluding them from training has minimal impact on performance while substantially reducing computation costs and enhancing data content privacy protection. Specifically, EFTViT comprises a series of lightweight local modules and a larger global module, updated independently on clients and the central server, respectively. The local modules are trained on masked image patches, while the global module is trained on intermediate patch features uploaded from the local client, balanced through a proposed median sampling strategy to erase client data distribution privacy. We analyze the computational complexity and privacy protection of EFTViT. Extensive experiments on popular benchmarks show that EFTViT achieves up to 28.17% accuracy improvement, reduces local training computational cost by up to 2.8$\times$, and cuts local training time by up to 4.4$\times$ compared to existing methods.

[476] OCR Hinders RAG: Evaluating the Cascading Impact of OCR on Retrieval-Augmented Generation

Junyuan Zhang, Qintong Zhang, Bin Wang, Linke Ouyang, Zichen Wen, Ying Li, Ka-Ho Chow, Conghui He, Wentao Zhang

Main category: cs.CV

TL;DR: OHRBench is the first benchmark for evaluating OCR noise impact on RAG systems, featuring 8,561 document images and 8,498 Q&A pairs across 7 domains, revealing current OCR solutions are inadequate for high-quality RAG knowledge bases.

DetailsMotivation: RAG systems rely on OCR to extract structured data from PDFs for knowledge bases, but OCR imperfections introduce noise that affects RAG performance. There was no dedicated benchmark to measure this cascading impact.

Method: Created OHRBench with carefully selected document images and multimodal Q&A pairs. Identified two OCR noise types (Semantic and Formatting) and applied perturbations to generate structured data with varying noise levels. Evaluated current OCR solutions and systematically assessed noise impact on RAG.

Result: Current OCR solutions are not competent for constructing high-quality RAG knowledge bases. Both Semantic and Formatting noise types significantly impact RAG performance, with clear trend relationships between noise degree and performance degradation.

Conclusion: OHRBench provides the first comprehensive benchmark to understand OCR’s cascading impact on RAG systems, revealing critical limitations in current OCR technology and establishing the need for improved OCR solutions tailored for RAG applications.

Abstract: Retrieval-augmented Generation (RAG) enhances Large Language Models (LLMs) by integrating external knowledge to reduce hallucinations and incorporate up-to-date information without retraining. As an essential part of RAG, external knowledge bases are commonly built by extracting structured data from unstructured PDF documents using Optical Character Recognition (OCR). However, given the imperfect prediction of OCR and the inherent non-uniform representation of structured data, knowledge bases inevitably contain various OCR noises. In this paper, we introduce OHRBench, the first benchmark for understanding the cascading impact of OCR on RAG systems. OHRBench includes 8,561 carefully selected unstructured document images from seven real-world RAG application domains, along with 8,498 Q&A pairs derived from multimodal elements in documents, challenging existing OCR solutions used for RAG. To better understand OCR’s impact on RAG systems, we identify two primary types of OCR noise: Semantic Noise and Formatting Noise and apply perturbation to generate a set of structured data with varying degrees of each OCR noise. Using OHRBench, we first conduct a comprehensive evaluation of current OCR solutions and reveal that none is competent for constructing high-quality knowledge bases for RAG systems. We then systematically evaluate the impact of these two noise types and demonstrate the trend relationship between the degree of OCR noise and RAG performance. Our OHRBench, including PDF documents, Q&As, and the ground truth structured data are released at: https://github.com/opendatalab/OHR-Bench

[477] Street Gaussians without 3D Object Tracker

Ruida Zhang, Chengxi Li, Chenyangguang Zhang, Xingyu Liu, Haili Yuan, Yanyan Li, Xiangyang Ji, Gim Hee Lee

Main category: cs.CV

TL;DR: Proposes a method for realistic scene reconstruction in driving scenarios using 2D foundation models instead of 3D trackers, with motion learning to correct tracking errors.

DetailsMotivation: Existing methods rely on manual labeling or limited 3D trackers for dynamic object reconstruction, which lack generalization and robustness in real-world settings.

Method: Uses 2D deep trackers within a 3D object fusion strategy and introduces motion learning in implicit feature space to autonomously correct trajectory errors and recover missed detections.

Result: Outperforms existing approaches on Waymo-NOTR and KITTI datasets.

Conclusion: The method eliminates reliance on 3D trackers and enhances robustness across diverse environments by leveraging 2D foundation models with error correction capabilities.

Abstract: Realistic scene reconstruction in driving scenarios poses significant challenges due to fast-moving objects. Most existing methods rely on labor-intensive manual labeling of object poses to reconstruct dynamic objects in canonical space and move them based on these poses during rendering. While some approaches attempt to use 3D object trackers to replace manual annotations, the limited generalization of 3D trackers – caused by the scarcity of large-scale 3D datasets – results in inferior reconstructions in real-world settings. In contrast, 2D foundation models demonstrate strong generalization capabilities. To eliminate the reliance on 3D trackers and enhance robustness across diverse environments, we propose a stable object tracking module by leveraging associations from 2D deep trackers within a 3D object fusion strategy. We address inevitable tracking errors by further introducing a motion learning strategy in an implicit feature space that autonomously corrects trajectory errors and recovers missed detections. Experimental results on Waymo-NOTR and KITTI show that our method outperforms existing approaches. Our code will be released on https://lolrudy.github.io/No3DTrackSG/.

[478] Temporal Preference Optimization for Long-Form Video Understanding

Rui Li, Xiaohan Wang, Yuhui Zhang, Orr Zohar, Zeyu Wang, Serena Yeung-Levy

Main category: cs.CV

TL;DR: TPO is a post-training framework that enhances video-LMMs’ temporal grounding through preference learning, achieving state-of-the-art performance on long-form video benchmarks.

DetailsMotivation: Existing video-LMMs struggle with effective temporal grounding in long-form videos, creating a need for improved temporal understanding capabilities.

Method: Temporal Preference Optimization (TPO) uses self-training with curated preference datasets at two granularities: localized temporal grounding and comprehensive temporal grounding to differentiate between well-grounded and inaccurate responses.

Result: TPO significantly enhances temporal understanding across three benchmarks (LongVideoBench, MLVU, Video-MME) and establishes LLaVA-Video-TPO as the leading 7B model on Video-MME.

Conclusion: TPO provides a scalable and efficient solution for advancing temporal reasoning in long-form video understanding while reducing reliance on manual annotations.

Abstract: Despite significant advancements in video large multimodal models (video-LMMs), achieving effective temporal grounding in long-form videos remains a challenge for existing models. To address this limitation, we propose Temporal Preference Optimization (TPO), a novel post-training framework designed to enhance the temporal grounding capabilities of video-LMMs through preference learning. TPO adopts a self-training approach that enables models to differentiate between well-grounded and less accurate temporal responses by leveraging curated preference datasets at two granularities: localized temporal grounding, which focuses on specific video segments, and comprehensive temporal grounding, which captures extended temporal dependencies across entire video sequences. By optimizing on these preference datasets, TPO significantly enhances temporal understanding while reducing reliance on manually annotated data. Extensive experiments on three long-form video understanding benchmarks–LongVideoBench, MLVU, and Video-MME–demonstrate the effectiveness of TPO across two state-of-the-art video-LMMs. Notably, LLaVA-Video-TPO establishes itself as the leading 7B model on the Video-MME benchmark, underscoring the potential of TPO as a scalable and efficient solution for advancing temporal reasoning in long-form video understanding. Project page: https://ruili33.github.io/tpo_website.

[479] A Data-Free Analytical Quantization Scheme for Deep Learning Models

Ahmed Luqman, Khuzemah Qazi, Murray Patterson, Malik Jehan Khan, Imdadullah Khan

Main category: cs.CV

TL;DR: Novel post-training quantization method that finds optimal clipping thresholds and scaling factors with mathematical guarantees to minimize quantization noise, reducing model size and computational requirements while preserving accuracy.

DetailsMotivation: CNN models have large computational and storage demands that challenge deployment on resource-constrained devices. Quantization can reduce these requirements by using lower-bit representations.

Method: Post-training quantization method that finds optimal clipping thresholds and scaling factors with mathematical guarantees to minimize quantization noise.

Result: Empirical results show significant reduction in model size and computational requirements while preserving model accuracy on real-world datasets.

Conclusion: The proposed quantization method effectively addresses deployment challenges of CNN models on resource-constrained devices by reducing storage and computational demands without sacrificing accuracy.

Abstract: Despite the success of CNN models on a variety of Image classification and segmentation tasks, their extensive computational and storage demands pose considerable challenges for real-world deployment on resource-constrained devices. Quantization is one technique that aims to alleviate these large storage requirements and speed up the inference process by reducing the precision of model parameters to lower-bit representations. In this paper, we introduce a novel post-training quantization method for model weights. Our method finds optimal clipping thresholds and scaling factors along with mathematical guarantees that our method minimizes quantization noise. Empirical results on real-world datasets demonstrate that our quantization scheme significantly reduces model size and computational requirements while preserving model accuracy.

[480] FreeSplatter: Pose-free Gaussian Splatting for Sparse-view 3D Reconstruction

Jiale Xu, Shenghua Gao, Ying Shan

Main category: cs.CV

TL;DR: FreeSplatter is a feed-forward framework that generates high-quality 3D Gaussians from uncalibrated sparse-view images while simultaneously estimating camera parameters within seconds, outperforming pose-dependent methods.

DetailsMotivation: Sparse-view reconstruction typically requires precise camera poses, but obtaining these parameters from sparse-view images remains challenging. Existing methods struggle with pose estimation from limited views.

Method: Uses a streamlined transformer architecture with self-attention blocks to exchange information among multi-view image tokens, decoding them into pixel-aligned 3D Gaussian primitives in a unified reference frame. Developed two specialized variants for object-centric and scene-level reconstruction.

Result: Outperforms several pose-dependent Large Reconstruction Models by a notable margin while achieving comparable or better pose estimation accuracy than state-of-the-art pose-free approach MASt3R in challenging benchmarks.

Conclusion: FreeSplatter eliminates camera pose management complexity while delivering exceptional visual fidelity, streamlining text/image-to-3D content creation pipelines without requiring pre-calibrated camera parameters.

Abstract: Sparse-view reconstruction models typically require precise camera poses, yet obtaining these parameters from sparse-view images remains challenging. We introduce FreeSplatter, a scalable feed-forward framework that generates high-quality 3D Gaussians from uncalibrated sparse-view images while estimating camera parameters within seconds. Our approach employs a streamlined transformer architecture where self-attention blocks facilitate information exchange among multi-view image tokens, decoding them into pixel-aligned 3D Gaussian primitives within a unified reference frame. This representation enables both high-fidelity 3D modeling and efficient camera parameter estimation using off-the-shelf solvers. We develop two specialized variants–for object-centric and scene-level reconstruction–trained on comprehensive datasets. Remarkably, FreeSplatter outperforms several pose-dependent Large Reconstruction Models (LRMs) by a notable margin while achieving comparable or even better pose estimation accuracy compared to state-of-the-art pose-free reconstruction approach MASt3R in challenging benchmarks. Beyond technical benchmarks, FreeSplatter streamlines text/image-to-3D content creation pipelines, eliminating the complexity of camera pose management while delivering exceptional visual fidelity.

[481] Monocular Facial Appearance Capture in the Wild

Yingyan Xu, Kate Gadola, Prashanth Chandran, Sebastian Weiss, Markus Gross, Gaspard Zoss, Derek Bradley

Main category: cs.CV

TL;DR: Novel method for high-fidelity facial appearance reconstruction from monocular video with simple head rotation, handling complex lighting and occlusions without studio equipment.

DetailsMotivation: To enable high-quality facial appearance capture (geometry, albedo, specular properties) from simple in-the-wild videos rather than complex studio setups, making professional-quality facial reconstruction more accessible.

Method: Uses monocular video with simple head rotation, explicitly accounts for environment lighting, visibility and occlusions without simplifying assumptions, reconstructs surface geometry, diffuse albedo, specular intensity and roughness.

Result: Produces facial appearance maps approaching studio-based multi-view capture fidelity, but with much simpler and cheaper capture procedure in unconstrained environments.

Conclusion: Demonstrates that high-fidelity facial appearance reconstruction is achievable from lightweight in-the-wild capture, potentially democratizing professional-quality facial scanning.

Abstract: We present a new method for reconstructing the appearance properties of human faces from a lightweight capture procedure in an unconstrained environment. Our method recovers the surface geometry, diffuse albedo, specular intensity and specular roughness from a monocular video containing a simple head rotation in-the-wild. Notably, we make no simplifying assumptions on the environment lighting, and we explicitly take visibility and occlusions into account. As a result, our method can produce facial appearance maps that approach the fidelity of studio-based multi-view captures, but with a far easier and cheaper procedure.

[482] Does Acceleration Cause Hidden Instability in Vision Language Models? Uncovering Instance-Level Divergence Through a Large-Scale Empirical Study

Yizheng Sun, Hao Li, Chang Xu, Hongpeng Zhou, Chenghua Lin, Riza Batista-Navarro, Jingyuan Sun

Main category: cs.CV

TL;DR: Accelerated vision-language models show significant answer instability despite minimal overall performance degradation, with up to 20% answer changes and 6.5% correct-to-incorrect conversions.

DetailsMotivation: To investigate whether accelerated VLMs maintain answer consistency for the same questions after acceleration, which is crucial for stability-critical applications like medical diagnosis.

Method: Systematic testing of four leading VLMs (LLaVA-1.5, LLaVA-Next, Qwen2-VL, Qwen2.5-VL) with eight acceleration methods across ten multi-modal benchmarks, including case studies with medical VLM LLaVA-Med.

Result: Accelerated models changed original answers up to 20% of the time, with up to 6.5% converting correct answers to incorrect. Input perturbations amplified these inconsistencies.

Conclusion: Current VLM acceleration evaluations overlook instance-level stability, revealing a critical need for stability checks to ensure trustworthy real-world deployment in sensitive applications.

Abstract: Vision-Language Models (VLMs) are powerful yet computationally intensive for widespread practical deployments. To address such challenge without costly re-training, post-training acceleration techniques like quantization and token reduction are extensively explored. However, current acceleration evaluations primarily target minimal overall performance degradation, overlooking a crucial question: does the accelerated model still give the same answers to the same questions as it did before acceleration? This is vital for stability-centered industrial applications where consistently correct answers for specific, known situations are paramount, such as in AI-based disease diagnosis. We systematically investigate this for accelerated VLMs, testing four leading models (LLaVA-1.5, LLaVA-Next, Qwen2-VL, Qwen2.5-VL) with eight acceleration methods on ten multi-modal benchmarks. Our findings are stark: despite minimal aggregate performance drops, accelerated models changed original answers up to 20% of the time. Critically, up to 6.5% of these changes converted correct answers to incorrect. Input perturbations magnified these inconsistencies, and the trend is confirmed by case studies with the medical VLM LLaVA-Med. This research reveals a significant oversight in VLM acceleration, stressing an urgent need for instance-level stability checks to ensure trustworthy real-world deployment.

[483] VLM-AD: End-to-End Autonomous Driving through Vision-Language Model Supervision

Yi Xu, Yuxin Hu, Zaiwei Zhang, Gregory P. Meyer, Siva Karthik Mustikovela, Siddhartha Srinivasa, Eric M. Wolff, Xin Huang

Main category: cs.CV

TL;DR: VLM-AD uses vision-language models as teachers to provide reasoning-based supervision for autonomous driving models, improving planning accuracy and reducing collisions without requiring VLMs during inference.

DetailsMotivation: Existing end-to-end autonomous driving models mimic driving patterns but lack underlying reasoning processes, limiting their ability to handle challenging scenarios.

Method: Leverages vision-language models as teachers to provide additional supervision with unstructured reasoning information and structured action labels during training.

Result: Achieves significant improvements in planning accuracy, reduced collision rates on nuScenes dataset, and better route completion and driving scores in closed-loop evaluation.

Conclusion: VLM-AD enhances feature representations to capture driving rationale, demonstrates effectiveness in long-horizon interactive scenarios, and shows potential for safe real-world deployment.

Abstract: Human drivers rely on commonsense reasoning to navigate diverse and dynamic real-world scenarios. Existing end-to-end (E2E) autonomous driving (AD) models are typically optimized to mimic driving patterns observed in data, without capturing the underlying reasoning processes. This limitation constrains their ability to handle challenging driving scenarios. To close this gap, we propose VLM-AD, a method that leverages vision-language models (VLMs) as teachers to enhance training by providing additional supervision that incorporates unstructured reasoning information and structured action labels. Such supervision enhances the model’s ability to learn richer feature representations that capture the rationale behind driving patterns. Importantly, our method does not require a VLM during inference, making it practical for real-time deployment. When integrated with state-of-the-art methods, VLM-AD achieves significant improvements in planning accuracy and reduced collision rates on the nuScenes dataset. It further improves route completion and driving scores under closed-loop evaluation, demonstrating its effectiveness in long-horizon, interactive driving scenarios and its potential for safe and reliable real-world deployment.

[484] VPO: Aligning Text-to-Video Generation Models with Prompt Optimization

Jiale Cheng, Ruiliang Lyu, Xiaotao Gu, Xiao Liu, Jiazheng Xu, Yida Lu, Jiayan Teng, Zhuoyi Yang, Yuxiao Dong, Jie Tang, Hongning Wang, Minlie Huang

Main category: cs.CV

TL;DR: VPO is a prompt optimization framework for video generation that improves safety, alignment, and video quality by preserving user intent through a two-stage approach with SFT and preference learning.

DetailsMotivation: There's a gap between detailed training descriptions and real-world user inputs which are often concise and vague, leading to suboptimal video generation. Current LLM-based methods may distort intent, omit details, or introduce safety risks.

Method: Two-stage optimization: 1) Construct and refine SFT dataset based on safety and alignment principles, 2) Use text-level and video-level feedback for preference learning to optimize the SFT model.

Result: VPO significantly improves safety, alignment, and video quality compared to baselines, shows strong generalization across video models, and can outperform/combine with RLHF methods.

Conclusion: VPO provides an effective framework for aligning video generation models through principled prompt optimization that preserves user intent while enhancing safety and quality.

Abstract: Video generation models have achieved remarkable progress in text-to-video tasks. These models are typically trained on text-video pairs with highly detailed and carefully crafted descriptions, while real-world user inputs during inference are often concise, vague, or poorly structured. This gap makes prompt optimization crucial for generating high-quality videos. Current methods often rely on large language models (LLMs) to refine prompts through in-context learning, but suffer from several limitations: they may distort user intent, omit critical details, or introduce safety risks. Moreover, they optimize prompts without considering the impact on the final video quality, which can lead to suboptimal results. To address these issues, we introduce VPO, a principled framework that optimizes prompts based on three core principles: harmlessness, accuracy, and helpfulness. The generated prompts faithfully preserve user intents and, more importantly, enhance the safety and quality of generated videos. To achieve this, VPO employs a two-stage optimization approach. First, we construct and refine a supervised fine-tuning (SFT) dataset based on principles of safety and alignment. Second, we introduce both text-level and video-level feedback to further optimize the SFT model with preference learning. Our extensive experiments demonstrate that VPO significantly improves safety, alignment, and video quality compared to baseline methods. Moreover, VPO shows strong generalization across video generation models. Furthermore, we demonstrate that VPO could outperform and be combined with RLHF methods on video generation models, underscoring the effectiveness of VPO in aligning video generation models. Our code and data are publicly available at https://github.com/thu-coai/VPO.

[485] Compositional Generative Model of Unbounded 4D Cities

Haozhe Xie, Zhaoxi Chen, Fangzhou Hong, Ziwei Liu

Main category: cs.CV

TL;DR: CityDreamer4D is a compositional generative model for unbounded 4D city generation that separates dynamic objects from static scenes using specialized neural fields and BEV representations.

DetailsMotivation: 4D city generation is more challenging than 3D scenes due to structurally complex objects like buildings/vehicles and human sensitivity to urban distortions. Existing methods struggle with the compositional nature and dynamic elements of urban environments.

Method: Proposes Traffic Scenario Generator and Unbounded Layout Generator using BEV representation. Combines stuff-oriented and instance-oriented neural fields with customized generative hash grids and periodic positional embeddings for different object types.

Result: Delivers state-of-the-art performance in generating realistic 4D cities. Supports downstream applications including instance editing, city stylization, and urban simulation.

Conclusion: The compositional approach effectively addresses the challenges of 4D city generation by separating dynamic and static elements and using specialized neural field representations for different urban components.

Abstract: 3D scene generation has garnered growing attention in recent years and has made significant progress. Generating 4D cities is more challenging than 3D scenes due to the presence of structurally complex, visually diverse objects like buildings and vehicles, and heightened human sensitivity to distortions in urban environments. To tackle these issues, we propose CityDreamer4D, a compositional generative model specifically tailored for generating unbounded 4D cities. Our main insights are 1) 4D city generation should separate dynamic objects (e.g., vehicles) from static scenes (e.g., buildings and roads), and 2) all objects in the 4D scene should be composed of different types of neural fields for buildings, vehicles, and background stuff. Specifically, we propose Traffic Scenario Generator and Unbounded Layout Generator to produce dynamic traffic scenarios and static city layouts using a highly compact BEV representation. Objects in 4D cities are generated by combining stuff-oriented and instance-oriented neural fields for background stuff, buildings, and vehicles. To suit the distinct characteristics of background stuff and instances, the neural fields employ customized generative hash grids and periodic positional embeddings as scene parameterizations. Furthermore, we offer a comprehensive suite of datasets for city generation, including OSM, GoogleEarth, and CityTopia. The OSM dataset provides a variety of real-world city layouts, while the Google Earth and CityTopia datasets deliver large-scale, high-quality city imagery complete with 3D instance annotations. Leveraging its compositional design, CityDreamer4D supports a range of downstream applications, such as instance editing, city stylization, and urban simulation, while delivering state-of-the-art performance in generating realistic 4D cities.

[486] BloomScene: Lightweight Structured 3D Gaussian Splatting for Crossmodal Scene Generation

Xiaolu Hou, Mingcheng Li, Dingkang Yang, Jiawei Chen, Ziyun Qian, Xiao Zhao, Yue Jiang, Jinjie Wei, Qingyao Xu, Lihua Zhang

Main category: cs.CV

TL;DR: BloomScene is a lightweight 3D Gaussian splatting method for crossmodal scene generation from text or images, addressing storage efficiency and geometric distortion issues in current methods.

DetailsMotivation: Current 3D scene generation methods rely on pre-trained text-to-image diffusion models and monocular depth estimators, but they occupy large storage space and suffer from geometric distortions due to lack of effective regularization.

Method: Proposes a crossmodal progressive scene generation framework with incremental point cloud reconstruction and 3D Gaussian splatting, hierarchical depth prior-based regularization for depth accuracy and smoothness, and structured context-guided compression using hash grids to reduce redundancy.

Result: Comprehensive experiments across multiple scenes demonstrate significant potential and advantages over baseline methods, generating diverse and high-quality 3D scenes.

Conclusion: BloomScene provides an effective solution for lightweight, structured 3D scene generation with reduced storage overhead and improved geometric quality compared to existing approaches.

Abstract: With the widespread use of virtual reality applications, 3D scene generation has become a new challenging research frontier. 3D scenes have highly complex structures and need to ensure that the output is dense, coherent, and contains all necessary structures. Many current 3D scene generation methods rely on pre-trained text-to-image diffusion models and monocular depth estimators. However, the generated scenes occupy large amounts of storage space and often lack effective regularisation methods, leading to geometric distortions. To this end, we propose BloomScene, a lightweight structured 3D Gaussian splatting for crossmodal scene generation, which creates diverse and high-quality 3D scenes from text or image inputs. Specifically, a crossmodal progressive scene generation framework is proposed to generate coherent scenes utilizing incremental point cloud reconstruction and 3D Gaussian splatting. Additionally, we propose a hierarchical depth prior-based regularization mechanism that utilizes multi-level constraints on depth accuracy and smoothness to enhance the realism and continuity of the generated scenes. Ultimately, we propose a structured context-guided compression mechanism that exploits structured hash grids to model the context of unorganized anchor attributes, which significantly eliminates structural redundancy and reduces storage overhead. Comprehensive experiments across multiple scenes demonstrate the significant potential and advantages of our framework compared with several baselines.

[487] Learning Visual Proxy for Compositional Zero-Shot Learning

Shiyu Zhang, Cheng Yan, Yang Liu, Chenchen Jing, Lei Zhou, Wenjun Wang

Main category: cs.CV

TL;DR: Visual Proxy Learning method reduces modality gaps in compositional zero-shot learning by creating visual proxies from text representations and using cross-modal constraints to improve discrimination of similar compositions.

DetailsMotivation: Current CZSL methods using VLMs suffer from modality gaps that hinder discrimination of semantically similar pairs and lack fine-grained visual cues in textual prototypes.

Method: Introduces Visual Proxy Learning that initializes visual proxies for attributes, objects, and compositions from text representations, and Cross-Modal Joint Learning (CMJL) that imposes cross-modal constraints between text-image and fine-grained visual spaces.

Result: Achieves state-of-the-art performance in closed-world scenarios and competitive results in open-world settings across four CZSL benchmarks.

Conclusion: The approach effectively reduces modality gaps and enhances compositional generalization, demonstrating improved discrimination of similar composition pairs.

Abstract: Compositional Zero-Shot Learning (CZSL) aims to recognize novel attribute-object compositions by leveraging knowledge from seen compositions. Current methods align textual prototypes with visual features via Vision-Language Models (VLMs), but suffer from two limitations: (1) modality gaps hinder the discrimination of semantically similar pairs, and (2) single-modal textual prototypes lack fine-grained visual cues. In this paper, we introduce Visual Proxy Learning, a method that reduces modality gaps and enhances compositional generalization. We initialize visual proxies for attributes, objects, and their compositions using text representations and optimize the visual space to capture fine-grained cues, improving visual representations. Additionally, we propose Cross-Modal Joint Learning (CMJL), which imposes cross-modal constraints between the text-image and fine-grained visual spaces, improving generalization for unseen compositions and discriminating similar pairs. Experiments show state-of-the-art performance in closed-world scenarios and competitive results in open-world settings across four CZSL benchmarks, demonstrating the effectiveness of our approach in compositional generalization.

[488] StoryReasoning Dataset: Using Chain-of-Thought for Scene Understanding and Grounded Story Generation

Daniel A. P. Oliveira, David Martins de Matos

Main category: cs.CV

TL;DR: StoryReasoning dataset with 4,178 stories from movie images addresses visual storytelling hallucinations through character grounding, cross-frame re-identification, and chain-of-thought reasoning, reducing hallucinations by 12.3% and improving creativity by 31%.

DetailsMotivation: Visual storytelling systems struggle with maintaining character identity across frames and linking actions to appropriate subjects, leading to referential hallucinations that need addressing through visual grounding.

Method: Proposed StoryReasoning dataset with structured scene analyses and grounded stories, featuring cross-frame object re-identification using visual similarity and face recognition, chain-of-thought reasoning for narrative modeling, and grounding scheme linking text to visual entities across frames.

Result: Fine-tuned Qwen2.5-VL 7B model (Qwen Storyteller) reduced hallucinations from 4.06 to 3.56 (-12.3%) per story and improved creativity from 2.58 to 3.38 (+31.0%) compared to non-fine-tuned model.

Conclusion: The proposed approach effectively addresses referential hallucinations in visual storytelling through structured grounding and cross-frame consistency, demonstrating significant improvements in both accuracy and creativity.

Abstract: Visual storytelling systems struggle to maintain character identity across frames and link actions to appropriate subjects, frequently leading to referential hallucinations. These issues can be addressed through grounding of characters, objects, and other entities on the visual elements. We propose StoryReasoning, a dataset containing 4,178 stories derived from 52,016 movie images, with both structured scene analyses and grounded stories. Each story maintains character and object consistency across frames while explicitly modeling multi-frame relationships through structured tabular representations. Our approach features cross-frame object re-identification using visual similarity and face recognition, chain-of-thought reasoning for explicit narrative modeling, and a grounding scheme that links textual elements to visual entities across multiple frames. We establish baseline performance by fine-tuning Qwen2.5-VL 7B, creating Qwen Storyteller, which performs end-to-end object detection, re-identification, and landmark detection while maintaining consistent object references throughout the story. Evaluation demonstrates a reduction from 4.06 to 3.56 (-12.3%) hallucinations on average per story and an improvement in creativity from 2.58 to 3.38 (+31.0%) when compared to a non-fine-tuned model.

[489] Vision without Images: End-to-End Computer Vision from Single Compressive Measurements

Fengpu Pan, Heting Gao, Jiangtao Wen, Yuxing Han

Main category: cs.CV

TL;DR: Novel SCI framework using 8x8 binary masks with CompDAE autoencoder that performs computer vision tasks directly from noisy compressive measurements without image reconstruction, achieving SOTA performance with low complexity especially in low-light conditions.

DetailsMotivation: Address challenges in Snapshot Compressed Imaging (SCI) under low-light/low-SNR conditions and hardware constraints that limit large mask sizes, requiring smaller hardware-friendly designs.

Method: CompDAE (Compressive Denoising Autoencoder) based on STFormer architecture with rate-constrained training strategy, using 8x8 pseudo-random binary masks and shared encoder with lightweight task-specific decoders.

Result: Achieves state-of-the-art performance with significantly lower complexity across multiple datasets, especially effective under ultra-low-light conditions where traditional CMOS and SCI pipelines fail.

Conclusion: The framework enables physically feasible implementations with small masks and performs downstream tasks directly from compressive measurements without reconstruction, making it suitable for practical hardware-constrained applications.

Abstract: Snapshot Compressed Imaging (SCI) offers high-speed, low-bandwidth, and energy-efficient image acquisition, but remains challenged by low-light and low signal-to-noise ratio (SNR) conditions. Moreover, practical hardware constraints in high-resolution sensors limit the use of large frame-sized masks, necessitating smaller, hardware-friendly designs. In this work, we present a novel SCI-based computer vision framework using pseudo-random binary masks of only 8$\times$8 in size for physically feasible implementations. At its core is CompDAE, a Compressive Denoising Autoencoder built on the STFormer architecture, designed to perform downstream tasks–such as edge detection and depth estimation–directly from noisy compressive raw pixel measurements without image reconstruction. CompDAE incorporates a rate-constrained training strategy inspired by BackSlash to promote compact, compressible models. A shared encoder paired with lightweight task-specific decoders enables a unified multi-task platform. Extensive experiments across multiple datasets demonstrate that CompDAE achieves state-of-the-art performance with significantly lower complexity, especially under ultra-low-light conditions where traditional CMOS and SCI pipelines fail.

[490] Exploring Primitive Visual Measurement Understanding and the Role of Output Format in Learning in Vision-Language Models

Ankit Yadav, Lingqiao Liu, Yuankai Qi

Main category: cs.CV

TL;DR: Vision-language models show improved performance on spatial and measurement tasks when using coherent sentence outputs and scaled numeric tokens, with better out-of-domain generalization.

DetailsMotivation: To investigate current VLMs' capabilities in visual understanding and attribute measurement of primitive shapes, particularly focusing on spatial positioning, occlusion, rotation, and other shape attributes.

Method: Fine-tuned state-of-the-art VLMs (2B-8B parameters) using Low-Rank Adaptation (LoRA) and validated on multiple out-of-domain scenarios from a proposed benchmark with controlled 2D shape configurations.

Result: Coherent sentence-based outputs outperformed tuple formats, especially in out-of-domain scenarios. Scaling numeric tokens during loss computation enhanced numerical approximation capabilities for spatial and measurement tasks.

Conclusion: Output format design, loss scaling strategies, and robust generalization techniques are crucial for enhancing VLM training and fine-tuning, particularly for tasks requiring precise spatial approximations and strong out-of-domain generalization.

Abstract: This work investigates the capabilities of current vision-language models (VLMs) in visual understanding and attribute measurement of primitive shapes using a benchmark focused on controlled 2D shape configurations with variations in spatial positioning, occlusion, rotation, size, and shape attributes such as type, quadrant, center-coordinates, rotation, occlusion status, and color as shown in Figure 1 and supplementary Figures S3-S81. We fine-tune state-of-the-art VLMs (2B-8B parameters) using Low-Rank Adaptation (LoRA) and validate them on multiple out-of-domain (OD) scenarios from our proposed benchmark. Our findings reveal that coherent sentence-based outputs outperform tuple formats, particularly in OD scenarios with large domain gaps. Additionally, we demonstrate that scaling numeric tokens during loss computation enhances numerical approximation capabilities, further improving performance on spatial and measurement tasks. These results highlight the importance of output format design, loss scaling strategies, and robust generalization techniques in enhancing the training and fine-tuning of VLMs, particularly for tasks requiring precise spatial approximations and strong OD generalization.

[491] Event Camera Tuning for Detection Applications

David El-Chai Ben-Ezra, Daniel Brisk, Adar Tal

Main category: cs.CV

TL;DR: A heuristic method for tuning event camera bias parameters to optimize small object detection in staring scenarios, reducing the multi-variable problem to a two-parameter optimization.

DetailsMotivation: Current lack of systematic methods for adjusting neuromorphic camera biases to accommodate specific tasks, particularly for small object detection applications.

Method: Translate experimental properties and systemic constraints into mathematical terms, use functional analysis tools to collapse multi-variable bias optimization into a two-parameter problem solvable experimentally.

Result: Demonstrates that optimal camera parameter values for certain signals (like incandescent lamps powered by electrical grid) are significantly different from manufacturer defaults.

Conclusion: The proposed heuristic successfully squeezes event camera potential and expands detection capabilities for specific applications, showing that default manufacturer settings are suboptimal for certain use cases.

Abstract: One of the main challenges in unlocking the potential of neuromorphic cameras, also called ‘’event camera’’, is the development of novel methods that solve the multi-variable problem of adjusting their biases parameters to accommodate a desired task. Actually, it is very difficult to find in the literature a systematic heuristic that solves the problem for any desired application. In this paper we present a tuning parameters heuristic for the biases of event cameras, for tasks that require small objects detection in staring scenarios. The main purpose of the heuristic is to squeeze the camera’s potential, optimize its performance, and expand its detection capabilities as much as possible. In the presentation, we translate the experimental properties of event camera and systemic constrains into mathematical terms, and show, under certain assumptions and classical tools from functional analysis, how the multi-variable problem collapses into a two-parameter problem that can be solved experimentally. A main conclusion that will be demonstrated is that for certain desired signals, such as the one provided by an incandescent lamp powered by the periodic electrical grid, the optimal values of the camera are very far from the default values recommended by the manufacturer.

[492] CytoDiff: AI-Driven Cytomorphology Image Synthesis for Medical Diagnostics

Jan Carreras Boada, Rao Muhammad Umer, Carsten Marr

Main category: cs.CV

TL;DR: CytoDiff uses stable diffusion with LoRA fine-tuning to generate synthetic white blood cell images, improving classifier accuracy from 27% to 78% when training data is limited and imbalanced.

DetailsMotivation: Biomedical datasets face privacy constraints and severe class imbalance, hindering accurate ML model development for critical tasks like white blood cell classification in leukemia diagnosis.

Method: CytoDiff - a stable diffusion model fine-tuned with LoRA weights and guided by few-shot samples to generate high-fidelity synthetic white blood cell images.

Result: Addition of 5,000 synthetic images per class improved ResNet classifier accuracy from 27% to 78% (+51%) and CLIP-based classification from 62% to 77% (+15%) using small, imbalanced real datasets.

Conclusion: Synthetic image generation is a valuable tool for biomedical ML, enhancing data coverage and enabling secure data sharing while preserving patient privacy.

Abstract: Biomedical datasets are often constrained by stringent privacy requirements and frequently suffer from severe class imbalance. These two aspects hinder the development of accurate machine learning models. While generative AI offers a promising solution, producing synthetic images of sufficient quality for training robust classifiers remains challenging. This work addresses the classification of individual white blood cells, a critical task in diagnosing hematological malignancies such as acute myeloid leukemia (AML). We introduce CytoDiff, a stable diffusion model fine-tuned with LoRA weights and guided by few-shot samples that generates high-fidelity synthetic white blood cell images. Our approach demonstrates substantial improvements in classifier performance when training data is limited. Using a small, highly imbalanced real dataset, the addition of 5,000 synthetic images per class improved ResNet classifier accuracy from 27% to 78% (+51%). Similarly, CLIP-based classification accuracy increased from 62% to 77% (+15%). These results establish synthetic image generation as a valuable tool for biomedical machine learning, enhancing data coverage and facilitating secure data sharing while preserving patient privacy. Paper code is publicly available at https://github.com/JanCarreras24/CytoDiff.

[493] DesCLIP: Robust Continual Learning via General Attribute Descriptions for VLM-Based Visual Recognition

Chiyuan He, Zihuan Qiu, Fanman Meng, Linfeng Xu, Qingbo Wu, Hongliang Li

Main category: cs.CV

TL;DR: DesCLIP improves continual learning for vision-language models by using general attribute descriptions to create trilateral vision-GA-class associations instead of direct vision-class connections, reducing knowledge forgetting.

DetailsMotivation: Existing continual learning methods for VLMs focus on connecting visual features with specific class text, which overlooks relationships between general and specialized knowledge and exacerbates forgetting of VLM's recognition ability.

Method: Proposes DesCLIP with a language assistant to generate general attribute descriptions via prompts, an anchor-based embedding filter to obtain relevant GA description embeddings, and uses these as paired text embeddings for visual-textual matching to tune the visual encoder while calibrating class text embeddings.

Result: Extensive experiments show superior performance in VLM-based recognition compared to existing continual learning methods, demonstrating advancements and efficacy.

Conclusion: Leveraging general attribute descriptions to establish robust vision-GA-class trilateral associations effectively addresses knowledge forgetting in continual learning of vision-language models.

Abstract: Continual learning of vision-language models (VLMs) focuses on leveraging cross-modal pretrained knowledge to incrementally adapt to expanding downstream tasks and datasets, while tackling the challenge of knowledge forgetting. Existing research often focuses on connecting visual features with specific class text in downstream tasks, overlooking the latent relationships between general and specialized knowledge. Our findings reveal that forcing models to optimize inappropriate visual-text matches exacerbates forgetting of VLM’s recognition ability. To tackle this issue, we propose DesCLIP, which leverages general attribute (GA) descriptions to guide the understanding of specific class objects, enabling VLMs to establish robust vision-GA-class trilateral associations rather than relying solely on vision-class connections. Specifically, we introduce a language assistant to generate concrete GA description candidates via proper request prompts. Then, an anchor-based embedding filter is designed to obtain highly relevant GA description embeddings, which are leveraged as the paired text embeddings for visual-textual instance matching, thereby tuning the visual encoder. Correspondingly, the class text embeddings are gradually calibrated to align with these shared GA description embeddings. Extensive experiments demonstrate the advancements and efficacy of our proposed method, with comprehensive empirical evaluations highlighting its superior performance in VLM-based recognition compared to existing continual learning methods.

[494] VIKSER: Visual Knowledge-Driven Self-Reinforcing Reasoning Framework

Chao Wang, Chunbai Zhang, Yongxiao Tian, Yang Zhou, Yan Peng

Main category: cs.CV

TL;DR: VIKSER is a visual reasoning framework that uses fine-grained visual knowledge and Chain-of-Evidence prompting to address underspecification issues and improve interpretability, achieving SOTA results comparable to proprietary models.

DetailsMotivation: Current visual reasoning methods suffer from limited interpretability, underspecification in question texts, and absence of fine-grained visual knowledge for precise subject behavior understanding.

Method: Uses knowledge distillation from large language models, visual relationship detection for fine-grained knowledge extraction, question paraphrasing for underspecification, Chain-of-Evidence prompting for interpretable reasoning, and self-reflection technology for learning from mistakes.

Result: Achieves new state-of-the-art results on widely used datasets and performs on par with leading proprietary models like ChatGPT-5.

Conclusion: VIKSER effectively addresses key limitations in visual reasoning through fine-grained knowledge extraction and interpretable reasoning frameworks, demonstrating superior performance compared to existing methods.

Abstract: Visual reasoning refers to the task of solving questions about visual information. Current visual reasoning methods typically employ pre-trained vision-language model (VLM) strategies or deep neural network approaches. However, existing efforts are constrained by limited reasoning interpretability, while hindering by the phenomenon of underspecification in the question text. Additionally, the absence of fine-grained visual knowledge limits the precise understanding of subject behavior in visual reasoning tasks. To address these issues, we propose VIKSER (Visual Knowledge-Driven Self-Reinforcing Reasoning Framework). Specifically, VIKSER, trained using knowledge distilled from large language models, extracts fine-grained visual knowledge with the assistance of visual relationship detection techniques. Subsequently, VIKSER utilizes fine-grained visual knowledge to paraphrase the question with underspecification. Additionally, we design a novel prompting method called Chain-of-Evidence (CoE), which leverages the power of “evidence for reasoning” to endow VIKSER with interpretable reasoning capabilities. Meanwhile, the integration of self-reflection technology empowers VIKSER with the ability to learn and improve from its mistakes. Experiments conducted on widely used datasets demonstrate that VIKSER achieves new state-of-the-art (SOTA) results in relevant tasks. Moreover, VIKSER achieves performance on par with leading proprietary models, such as the latest ChatGPT-5.

[495] SpatialViz-Bench: An MLLM Benchmark for Spatial Visualization

Siting Wang, Minnan Pei, Luoyang Sun, Cheng Deng, Kun Shao, Zheng Tian, Haifeng Zhang, Jun Wang

Main category: cs.CV

TL;DR: SpatialViz-Bench is a new benchmark for evaluating spatial visualization in multi-modal LLMs, revealing significant performance gaps and counter-intuitive behaviors in current models.

DetailsMotivation: Existing evaluations for spatial visualization in MLLMs are inadequate, often embedded in broader assessments and potentially contaminated by training data overlap, requiring a dedicated benchmark.

Method: Created SpatialViz-Bench with 1,180 automatically generated problems across 12 tasks and 4 sub-abilities, then evaluated 33 state-of-the-art MLLMs on this benchmark.

Result: Found wide performance variations, difficulty perception misaligned with human intuition, 2D-to-3D performance cliffs, formulaic derivation preferences, and performance degradation from Chain-of-Thought prompting in open-source models.

Conclusion: State-of-the-art MLLMs still exhibit significant deficiencies in spatial visualization tasks, and SpatialViz-Bench effectively addresses this evaluation gap with publicly available data and code.

Abstract: Humans can directly imagine and manipulate visual images in their minds, a capability known as spatial visualization. While multi-modal Large Language Models (MLLMs) support imagination-based reasoning, spatial visualization remains insufficiently evaluated, typically embedded within broader mathematical and logical assessments. Existing evaluations often rely on IQ tests or math competitions that may overlap with training data, compromising assessment reliability. To this end, we introduce SpatialViz-Bench, a comprehensive multi-modal benchmark for spatial visualization with 12 tasks across 4 sub-abilities, comprising 1,180 automatically generated problems. Our evaluation of 33 state-of-the-art MLLMs not only reveals wide performance variations and demonstrates the benchmark’s strong discriminative power, but also uncovers counter-intuitive findings: models show difficulty perception misaligned with human intuition, exhibit dramatic 2Dto-3D performance cliffs, default to formulaic derivation over visualization, and paradoxically suffer performance degradation from Chain-of-Thought prompting in open-source models. Through statistical and qualitative analysis of error types, SpatialViz-Bench demonstrates that state-of-the-art MLLMs continue to exhibit deficiencies in spatial visualization tasks, thereby addressing a significant lacuna in the field. The benchmark data and evaluation code are publicly available.

[496] RAMer: Reconstruction-based Adversarial Model for Multi-party Multi-modal Multi-label Emotion Recognition

Xudong Yang, Yizhang Zhu, Hanfeng Liu, Zeyi Wen, Nan Tang, Yuyu Luo

Main category: cs.CV

TL;DR: RAMer is a novel multi-modal emotion recognition model that addresses data incompleteness in real-world scenarios by using reconstruction-based adversarial learning and contrastive learning to handle missing modalities, while incorporating personality auxiliary tasks and stack shuffle strategy for improved performance.

DetailsMotivation: Real-world multi-party settings often have incomplete modalities (non-speakers lack acoustic/textual inputs), causing performance degradation in conventional MMER models. Existing approaches also fail to properly handle modality heterogeneity.

Method: Proposes RAMer with: 1) Reconstruction-based adversarial learning with contrastive learning to handle missing modalities and enrich features, 2) Personality auxiliary task with modality-level attention to complement missing data, 3) Stack shuffle strategy to capture label-modality interdependencies.

Result: Achieves state-of-the-art performance on three benchmarks: MEmoR, CMU-MOSEI, and M³ED in both dyadic and multi-party MMER scenarios.

Conclusion: RAMer effectively addresses the challenges of incomplete modalities in real-world settings while capturing modality-specific characteristics and label-modality correlations, demonstrating superior performance across multiple benchmarks.

Abstract: Conventional Multi-modal multi-label emotion recognition (MMER) assumes complete access to visual, textual, and acoustic modalities. However, real-world multi-party settings often violate this assumption, as non-speakers frequently lack acoustic and textual inputs, leading to a significant degradation in model performance. Existing approaches also tend to unify heterogeneous modalities into a single representation, overlooking each modality’s unique characteristics. To address these challenges, we propose RAMer (Reconstruction-based Adversarial Model for Emotion Recognition), which refines multi-modal representations by not only exploring modality commonality and specificity but crucially by leveraging reconstructed features, enhanced by contrastive learning, to overcome data incompleteness and enrich feature quality. RAMer also introduces a personality auxiliary task to complement missing modalities using modality-level attention, improving emotion reasoning. To further strengthen the model’s ability to capture label and modality interdependency, we propose a stack shuffle strategy to enrich correlations between labels and modality-specific features. Experiments on three benchmarks, i.e., MEmoR, CMU-MOSEI, and $M^3ED$, demonstrate that RAMer achieves state-of-the-art performance in dyadic and multi-party MMER scenarios.

[497] Investigating and Enhancing Vision-Audio Capability in Omnimodal Large Language Models

Rui Hu, Delai Qiu, Shuyu Wei, Jiaming Zhang, Yining Wang, Shengping Liu, Jitao Sang

Main category: cs.CV

TL;DR: Self-Knowledge Distillation method improves vision-audio integration in Omnimodal LLMs by using vision-text components as teachers for vision-audio components.

DetailsMotivation: OLLMs struggle with integrating vision and audio modalities, showing suboptimal performance on audio queries compared to text queries due to insufficient alignment between vision and audio during training.

Method: Proposed Self-Knowledge Distillation training where vision-text component serves as teacher and vision-audio component as student, enabling audio processing analogous to text processing.

Result: Experimental results show Self-KD effectively enhances vision-audio capabilities by learning from vision-text components, improving audio-image interaction and multimodal task performance.

Conclusion: Self-Knowledge Distillation is an effective approach for bridging the performance gap between text and audio queries in Omnimodal LLMs by leveraging knowledge from well-aligned vision-text components.

Abstract: Omnimodal Large Language Models (OLLMs) have shown significant progress in integrating vision and text, but still struggle with integrating vision and audio, often exhibiting suboptimal performance when processing audio queries compared to text queries. This disparity is primarily due to insufficient alignment between vision and audio modalities during training, leading to inadequate attention to visual information when using audio queries. To mitigate this issue, we propose a Self-Knowledge Distillation (Self-KD) training method where the vision-text component of the OLLM serves as the teacher and the vision-audio component as the student. This enables the model to process audio in a manner analogous to its text processing. Our experimental results demonstrate that Self-KD is an effective method for enhancing the vision-audio capabilities of OLLMs by learning from the vision-text components, which subsequently improves the interaction between audio and images and results in improved performance on multimodal tasks.

[498] Removing Averaging: Personalized Lip-Sync Driven Characters Based on Identity Adapter

Yanyu Zhu, Lichen Bai, Jintao Xu, Hai-tao Zheng

Main category: cs.CV

TL;DR: UnAvgLip addresses the ’lip averaging’ problem in diffusion-based lip-syncing models by preserving facial identity details while maintaining accurate lip synchronization through identity embeddings and specialized attention mechanisms.

DetailsMotivation: Existing diffusion-based lip-syncing models struggle to maintain fine-grained facial details when dubbing unseen in-the-wild videos, leading to a 'lip averaging' phenomenon where subtle facial characteristics are lost despite good lip synchronization.

Method: Proposes UnAvgLip with two main components: (1) Identity Perceiver module that encodes facial embeddings to align with audio features, and (2) ID-CrossAttn module that injects facial embeddings into the generation process to enhance identity retention.

Result: Achieves significant improvements of 5% on identity consistency metric and 2% on SSIM metric across HDTF and LRW benchmark datasets, effectively mitigating the averaging phenomenon at modest training and inference cost.

Conclusion: UnAvgLip successfully addresses the lip averaging problem in visual dubbing by preserving unique facial characteristics while maintaining precise lip synchronization, demonstrating superior performance over existing approaches.

Abstract: Recent advances in diffusion-based lip-syncing generative models have demonstrated their ability to produce highly synchronized talking face videos for visual dubbing. Although these models excel at lip synchronization, they often struggle to maintain fine-grained control over facial details in generated images. In this work, we identify “lip averaging” phenomenon where the model fails to preserve subtle facial details when dubbing unseen in-the-wild videos. This issue arises because the commonly used UNet backbone primarily integrates audio features into visual representations in the latent space via cross-attention mechanisms and multi-scale fusion, but it struggles to retain fine-grained lip details in the generated faces. To address this issue, we propose UnAvgLip, which extracts identity embeddings from reference videos to generate highly faithful facial sequences while maintaining accurate lip synchronization. Specifically, our method comprises two primary components: (1) an Identity Perceiver module that encodes facial embeddings to align with conditioned audio features; and (2) an ID-CrossAttn module that injects facial embeddings into the generation process, enhancing model’s capability of identity retention. Extensive experiments demonstrate that, at a modest training and inference cost, UnAvgLip effectively mitigates the “averaging” phenomenon in lip inpainting, significantly preserving unique facial characteristics while maintaining precise lip synchronization. Compared with the original approach, our method demonstrates significant improvements of 5% on the identity consistency metric and 2% on the SSIM metric across two benchmark datasets (HDTF and LRW).

[499] Certifiably Optimal Anisotropic Rotation Averaging

Carl Olsson, Yaroslava Lochman, Johan Malmport, Christopher Zach

Main category: cs.CV

TL;DR: Certifiably optimal rotation averaging with anisotropic costs using stronger relaxation for improved accuracy

DetailsMotivation: Most rotation averaging methods focus on isotropic settings, but recent empirical results show that incorporating anisotropic uncertainties (measurement uncertainties) can improve solution quality, though global optimization remains challenging in anisotropic scenarios

Method: Proposed a stronger relaxation that incorporates anisotropic costs into certifiably optimal rotation averaging, addressing the limitations of existing isotropic solvers

Result: The proposed method recovers global optima in all tested datasets and leads to more accurate reconstructions in almost all scenes compared to existing isotropic solvers

Conclusion: Anisotropic rotation averaging with the proposed stronger relaxation provides certifiably optimal solutions and significantly improves reconstruction accuracy over traditional isotropic approaches

Abstract: Rotation averaging is a key subproblem in applications of computer vision and robotics. Many methods for solving this problem exist, and there are also several theoretical results analyzing difficulty and optimality. However, one aspect that most of these have in common is a focus on the isotropic setting, where the intrinsic uncertainties in the measurements are not fully incorporated into the resulting optimization task. Recent empirical results suggest that moving to an anisotropic framework, where these uncertainties are explicitly included, can result in an improvement of solution quality. However, global optimization for rotation averaging has remained a challenge in this scenario. In this work we show how anisotropic costs can be incorporated in certifiably optimal rotation averaging. We also demonstrate how existing solvers, designed for isotropic situations, fail in the anisotropic setting. Finally, we propose a stronger relaxation and empirically show that it recovers global optima in all tested datasets and leads to more accurate reconstructions in almost all scenes.

[500] CoViPAL: Layer-wise Contextualized Visual Token Pruning for Large Vision-Language Models

Zicong Tang, Ziyang Ma, Suqing Wang, Zuchao Li, Lefei Zhang, Hai Zhao, Yun Li, Qianren Wang

Main category: cs.CV

TL;DR: CoViPAL is a layer-wise contextualized visual token pruning method that uses a lightweight Plug-and-Play Pruning Module to remove redundant vision tokens in LVLMs, improving inference efficiency without accuracy loss.

DetailsMotivation: Large Vision-Language Models process thousands of vision tokens per image, causing high computational costs and memory overhead during prefilling and decoding stages. Existing pruning methods struggle in shallow layers due to insufficient contextual information.

Method: Proposes CoViPAL with a Plug-and-Play Pruning Module (PPM) that predicts and removes redundant vision tokens before LVLM processing. The PPM is lightweight, model-agnostic, and operates independently of LVLM architecture.

Result: Extensive experiments show CoViPAL outperforms training-free pruning methods under equal token budgets and surpasses training-based methods with comparable supervision while maintaining accuracy.

Conclusion: CoViPAL provides a scalable and efficient solution to improve inference efficiency in LVLMs without compromising performance, offering seamless integration with various models.

Abstract: Large Vision-Language Models (LVLMs) process multimodal inputs consisting of text tokens and vision tokens extracted from images or videos. Due to the rich visual information, a single image can generate thousands of vision tokens, leading to high computational costs during the prefilling stage and significant memory overhead during decoding. Existing methods attempt to prune redundant vision tokens, revealing substantial redundancy in visual representations. However, these methods often struggle in shallow layers due to the lack of sufficient contextual information. We argue that many visual tokens are inherently redundant even in shallow layers and can be safely and effectively pruned with appropriate contextual signals. In this work, we propose CoViPAL, a layer-wise contextualized visual token pruning method that employs a Plug-and-Play Pruning Module (PPM) to predict and remove redundant vision tokens before they are processed by the LVLM. The PPM is lightweight, model-agnostic, and operates independently of the LVLM architecture, ensuring seamless integration with various models. Extensive experiments on multiple benchmarks demonstrate that CoViPAL outperforms training-free pruning methods under equal token budgets and surpasses training-based methods with comparable supervision. CoViPAL offers a scalable and efficient solution to improve inference efficiency in LVLMs without compromising accuracy.

[501] Pathology-Aware Adaptive Watermarking for Text-Driven Medical Image Synthesis

Chanyoung Kim, Dayun Ju, Jinyeong Kim, Woojung Han, Roberto Alcover-Couso, Seong Jae Hwang

Main category: cs.CV

TL;DR: MedSign is a specialized watermarking framework for text-to-medical image synthesis that preserves diagnostic integrity by adaptively adjusting watermark strength based on pathological regions identified through cross-attention mechanisms.

DetailsMotivation: Address the urgent need for safeguards against unethical use of AI-generated medical images (e.g., insurance fraud, falsified records) while overcoming the challenge that standard watermarking techniques may distort fine-grained disease manifestations and compromise diagnostic integrity.

Method: Generate pathology localization maps using cross-attention between medical text tokens and diffusion denoising network, aggregate attention across layers/heads/time steps, then optimize LDM decoder to incorporate adaptive watermarking that minimizes interference in critical diagnostic regions during image synthesis.

Result: Achieves state-of-the-art performance in both image quality and detection accuracy on MIMIC-CXR and OIA-ODIR datasets, preserving diagnostic integrity while ensuring watermark robustness.

Conclusion: MedSign provides an effective solution for ethical watermarking in medical image generation by adaptively preserving pathological regions, enabling reliable detection of AI-generated medical images without compromising clinical diagnostic value.

Abstract: As recent text-conditioned diffusion models have enabled the generation of high-quality images, concerns over their potential misuse have also grown. This issue is critical in the medical domain, where text-conditioned generated medical images could enable insurance fraud or falsified records, highlighting the urgent need for reliable safeguards against unethical use. While watermarking techniques have emerged as a promising solution in general image domains, their direct application to medical imaging presents significant challenges. A key challenge is preserving fine-grained disease manifestations, as even minor distortions from a watermark may lead to clinical misinterpretation, which compromises diagnostic integrity. To overcome this gap, we present MedSign, a deep learning-based watermarking framework specifically designed for text-to-medical image synthesis, which preserves pathologically significant regions by adaptively adjusting watermark strength. Specifically, we generate a pathology localization map using cross-attention between medical text tokens and the diffusion denoising network, aggregating token-wise attention across layers, heads, and time steps. Leveraging this map, we optimize the LDM decoder to incorporate watermarking during image synthesis, ensuring cohesive integration while minimizing interference in diagnostically critical regions. Experimental results show that our MedSign preserves diagnostic integrity while ensuring watermark robustness, achieving state-of-the-art performance in image quality and detection accuracy on MIMIC-CXR and OIA-ODIR datasets.

[502] Accelerating Diffusion Sampling via Exploiting Local Transition Coherence

Shangwen Zhu, Han Zhang, Zhantao Yang, Qianyu Peng, Zhao Pu, Huangji Wang, Fan Cheng

Main category: cs.CV

TL;DR: LTC-Accel is a training-free acceleration method that speeds up text-to-image and text-to-video diffusion models by discovering statistical relationships in transition operators between adjacent denoising steps, achieving 1.67x speedup in Stable Diffusion and 1.55x in video generation.

DetailsMotivation: Text-based diffusion models suffer from lengthy sampling times, and existing acceleration methods either ignore statistical relationships between steps or rely on specific network structures, limiting their applicability.

Method: The method discovers a new statistical relationship in transition operators between adjacent denoising steps, focusing on network outputs rather than internal structures. This relationship is used to estimate current transition operators based on adjacent steps without requiring specific network assumptions.

Result: LTC-Accel achieves 1.67x speedup in Stable Diffusion v2 and 1.55x speedup in video generation models. When combined with distillation models, it achieves 10x speedup in video generation (over 16FPS real-time generation) while maintaining competitive sample quality.

Conclusion: LTC-Accel provides a universal, training-free acceleration method that works with almost all diffusion-based methods and is orthogonal to existing techniques, enabling significant speed improvements without compromising quality.

Abstract: Text-based diffusion models have made significant breakthroughs in generating high-quality images and videos from textual descriptions. However, the lengthy sampling time of the denoising process remains a significant bottleneck in practical applications. Previous methods either ignore the statistical relationships between adjacent steps or rely on attention or feature similarity between them, which often only works with specific network structures. To address this issue, we discover a new statistical relationship in the transition operator between adjacent steps, focusing on the relationship of the outputs from the network. This relationship does not impose any requirements on the network structure. Based on this observation, we propose a novel training-free acceleration method called LTC-Accel, which uses the identified relationship to estimate the current transition operator based on adjacent steps. Due to no specific assumptions regarding the network structure, LTC-Accel is applicable to almost all diffusion-based methods and orthogonal to almost all existing acceleration techniques, making it easy to combine with them. Experimental results demonstrate that LTC-Accel significantly speeds up sampling in text-to-image and text-to-video synthesis while maintaining competitive sample quality. Specifically, LTC-Accel achieves a speedup of 1.67-fold in Stable Diffusion v2 and a speedup of 1.55-fold in video generation models. When combined with distillation models, LTC-Accel achieves a remarkable 10-fold speedup in video generation, allowing real-time generation of more than 16FPS.

[503] KRETA: A Benchmark for Korean Reading and Reasoning in Text-Rich VQA Attuned to Diverse Visual Contexts

Taebaek Hwang, Minseo Kim, Gisang Lee, Seonuk Kim, Hyunjun Eun

Main category: cs.CV

TL;DR: KRETA is a new benchmark for Korean text-rich visual question answering that addresses the lack of evaluation resources for low-resource languages, featuring diverse visual contexts and a semi-automated data generation pipeline.

DetailsMotivation: Existing text-rich VQA benchmarks focus on high-resource languages like English, creating a critical gap for low-resource languages such as Korean, which hinders robust model evaluation and comparison.

Method: Developed a semi-automated VQA generation pipeline optimized for text-rich settings using refined stepwise image decomposition and a rigorous seven-metric evaluation protocol to ensure data quality.

Result: KRETA benchmark supports comprehensive evaluation across 15 domains and 26 image types, facilitating in-depth assessment of both visual text understanding and reasoning capabilities for Korean.

Conclusion: KRETA bridges the evaluation gap for Korean VLM research and provides an adaptable pipeline that can facilitate similar benchmark development for other languages, accelerating multilingual VLM research.

Abstract: Understanding and reasoning over text within visual contexts poses a significant challenge for Vision-Language Models (VLMs), given the complexity and diversity of real-world scenarios. To address this challenge, text-rich Visual Question Answering (VQA) datasets and benchmarks have emerged for high-resource languages like English. However, a critical gap persists for low-resource languages such as Korean, where the lack of comprehensive benchmarks hinders robust model evaluation and comparison. To bridge this gap, we introduce KRETA, a benchmark for Korean Reading and rEasoning in Text-rich VQA Attuned to diverse visual contexts. KRETA facilitates an in-depth evaluation of both visual text understanding and reasoning capabilities, while also supporting a multifaceted assessment across 15 domains and 26 image types. Additionally, we introduce a semi-automated VQA generation pipeline specifically optimized for text-rich settings, leveraging refined stepwise image decomposition and a rigorous seven-metric evaluation protocol to ensure data quality. While KRETA is tailored for Korean, we hope our adaptable and extensible pipeline will facilitate the development of similar benchmarks in other languages, thereby accelerating multilingual VLM research. The code and dataset for KRETA are available at https://github.com/tabtoyou/KRETA.

[504] Open-World Skill Discovery from Unsegmented Demonstrations

Jingwen Deng, Zihao Wang, Shaofei Cai, Anji Liu, Yitao Liang

Main category: cs.CV

TL;DR: Self-supervised Skill Boundary Detection (SBD) algorithm for temporal video segmentation without annotations, using prediction errors from action-prediction models to detect skill transitions in long gameplay videos.

DetailsMotivation: Open-world environments require agents to learn diverse skills from long, unsegmented demonstration videos. Existing methods rely on sequence sampling or human labeling, which is inefficient and doesn't scale well with the abundance of online video content.

Method: Developed SBD algorithm inspired by human cognitive event segmentation theory. Uses prediction errors from a pretrained unconditional action-prediction model to detect skill boundaries - significant prediction error increases indicate skill shifts. Evaluated in Minecraft environment.

Result: SBD-generated segments improved conditioned policies by 63.7% and 52.1% on short-term atomic skill tasks, and hierarchical agents by 11.3% and 20.8% on long-horizon tasks. Method enables leveraging diverse YouTube videos for training instruction-following agents.

Conclusion: The proposed self-supervised SBD approach effectively segments long demonstration videos into skill-consistent segments without human annotation, significantly improving agent performance in both atomic and hierarchical tasks while enabling scalable learning from online video content.

Abstract: Learning skills in open-world environments is essential for developing agents capable of handling a variety of tasks by combining basic skills. Online demonstration videos are typically long but unsegmented, making them difficult to segment and label with skill identifiers. Unlike existing methods that rely on sequence sampling or human labeling, we have developed a self-supervised learning-based approach to segment these long videos into a series of semantic-aware and skill-consistent segments. Drawing inspiration from human cognitive event segmentation theory, we introduce Skill Boundary Detection (SBD), an annotation-free temporal video segmentation algorithm. SBD detects skill boundaries in a video by leveraging prediction errors from a pretrained unconditional action-prediction model. This approach is based on the assumption that a significant increase in prediction error indicates a shift in the skill being executed. We evaluated our method in Minecraft, a rich open-world simulator with extensive gameplay videos available online. Our SBD-generated segments improved the average performance of conditioned policies by 63.7% and 52.1% on short-term atomic skill tasks, and their corresponding hierarchical agents by 11.3% and 20.8% on long-horizon tasks. Our method can leverage the diverse YouTube videos to train instruction-following agents. The project page can be found in https://craftjarvis.github.io/SkillDiscovery.

[505] NuiScene: Exploring Efficient Generation of Unbounded Outdoor Scenes

Han-Hung Lee, Qinghong Han, Angel X. Chang

Main category: cs.CV

TL;DR: Efficient outdoor scene generation method using vector encoding and explicit outpainting, with a curated dataset enabling style blending.

DetailsMotivation: Outdoor scene generation presents unique challenges compared to indoor scenes, including wide height variations and the need for rapid large landscape production, which prior methods focused on indoor generation haven't adequately addressed.

Method: Proposed approach encodes scene chunks as uniform vector sets for better compression and performance than spatially structured latents. Trained an explicit outpainting model for unbounded generation to improve coherence and speed compared to resampling-based inpainting. Created NuiScene43 dataset for joint training.

Result: Method offers better compression and performance than prior approaches. The explicit outpainting model improves coherence and speeds up generation by eliminating extra diffusion steps. Model can blend different environments (e.g., rural houses and city skyscrapers) when trained on scenes of varying styles.

Conclusion: The proposed approach effectively addresses outdoor scene generation challenges, demonstrating the potential of heterogeneous scene curation for joint training and enabling coherent blending of diverse environmental styles within single scenes.

Abstract: In this paper, we explore the task of generating expansive outdoor scenes, ranging from castles to high-rises. Unlike indoor scene generation, which has been a primary focus of prior work, outdoor scene generation presents unique challenges, including wide variations in scene heights and the need for a method capable of rapidly producing large landscapes. To address this, we propose an efficient approach that encodes scene chunks as uniform vector sets, offering better compression and performance than the spatially structured latents used in prior methods. Furthermore, we train an explicit outpainting model for unbounded generation, which improves coherence compared to prior resampling-based inpainting schemes while also speeding up generation by eliminating extra diffusion steps. To facilitate this task, we curate NuiScene43, a small but high-quality set of scenes, preprocessed for joint training. Notably, when trained on scenes of varying styles, our model can blend different environments, such as rural houses and city skyscrapers, within the same scene, highlighting the potential of our curation process to leverage heterogeneous scenes for joint training.

[506] Flip Learning: Weakly Supervised Erase to Segment Nodules in Breast Ultrasound

Yuhao Huang, Ao Chang, Haoran Dou, Xing Tao, Xinrui Zhou, Yan Cao, Ruobing Huang, Alejandro F Frangi, Lingyun Bao, Xin Yang, Dong Ni

Main category: cs.CV

TL;DR: Flip Learning: A multi-agent reinforcement learning framework for weakly-supervised nodule segmentation in breast ultrasound using only 2D/3D boxes, achieving performance comparable to fully-supervised methods.

DetailsMotivation: To develop an automated nodule segmentation system that reduces reliance on labor-intensive pixel-level annotations while maintaining high accuracy for clinical diagnosis and treatment planning in breast ultrasound imaging.

Method: Multi-agent reinforcement learning framework where agents erase target regions from boxes to flip classification tags, using superpixel/supervoxel encoding, three specialized rewards (classification score + two intensity distribution rewards), and progressive curriculum learning.

Result: Outperforms state-of-the-art weakly-supervised methods and foundation models, achieving comparable performance to fully-supervised learning algorithms on large in-house BUS and ABUS datasets.

Conclusion: Flip Learning provides an effective weakly-supervised segmentation solution that eliminates the need for pixel-level annotations while delivering precise nodule segmentation results suitable for clinical applications.

Abstract: Accurate segmentation of nodules in both 2D breast ultrasound (BUS) and 3D automated breast ultrasound (ABUS) is crucial for clinical diagnosis and treatment planning. Therefore, developing an automated system for nodule segmentation can enhance user independence and expedite clinical analysis. Unlike fully-supervised learning, weakly-supervised segmentation (WSS) can streamline the laborious and intricate annotation process. However, current WSS methods face challenges in achieving precise nodule segmentation, as many of them depend on inaccurate activation maps or inefficient pseudo-mask generation algorithms. In this study, we introduce a novel multi-agent reinforcement learning-based WSS framework called Flip Learning, which relies solely on 2D/3D boxes for accurate segmentation. Specifically, multiple agents are employed to erase the target from the box to facilitate classification tag flipping, with the erased region serving as the predicted segmentation mask. The key contributions of this research are as follows: (1) Adoption of a superpixel/supervoxel-based approach to encode the standardized environment, capturing boundary priors and expediting the learning process. (2) Introduction of three meticulously designed rewards, comprising a classification score reward and two intensity distribution rewards, to steer the agents’ erasing process precisely, thereby avoiding both under- and over-segmentation. (3) Implementation of a progressive curriculum learning strategy to enable agents to interact with the environment in a progressively challenging manner, thereby enhancing learning efficiency. Extensively validated on the large in-house BUS and ABUS datasets, our Flip Learning method outperforms state-of-the-art WSS methods and foundation models, and achieves comparable performance as fully-supervised learning algorithms.

[507] LATTE-MV: Learning to Anticipate Table Tennis Hits from Monocular Videos

Daniel Etaat, Dvij Kalaria, Nima Rahmanian, Shankar Sastry

Main category: cs.CV

TL;DR: This paper presents a system for 3D reconstruction of table tennis matches from monocular video and an uncertainty-aware controller that anticipates opponent actions, improving ball return rates by nearly 10% compared to non-anticipatory baselines.

DetailsMotivation: Physical agility alone is insufficient for competitive table tennis - champions excel by anticipating opponent intent to gain reaction time. Previous systems either lack anticipation capabilities or are limited by small datasets.

Method: Developed (1) a scalable system for 3D reconstruction of table tennis matches from monocular video, and (2) an uncertainty-aware controller that anticipates opponent actions.

Result: In simulation, the policy improved ball return rate against high-speed hits from 49.9% to 59.0% compared to a baseline non-anticipatory policy.

Conclusion: The proposed anticipatory agent with uncertainty-aware control and scalable 3D reconstruction significantly enhances table tennis gameplay performance by effectively predicting opponent actions.

Abstract: Physical agility is a necessary skill in competitive table tennis, but by no means sufficient. Champions excel in this fast-paced and highly dynamic environment by anticipating their opponent’s intent - buying themselves the necessary time to react. In this work, we take one step towards designing such an anticipatory agent. Previous works have developed systems capable of real-time table tennis gameplay, though they often do not leverage anticipation. Among the works that forecast opponent actions, their approaches are limited by dataset size and variety. Our paper contributes (1) a scalable system for reconstructing monocular video of table tennis matches in 3D and (2) an uncertainty-aware controller that anticipates opponent actions. We demonstrate in simulation that our policy improves the ball return rate against high-speed hits from 49.9% to 59.0% as compared to a baseline non-anticipatory policy.

[508] PS-ReID: Advancing Person Re-Identification and Precise Segmentation with Multimodal Retrieval

Jincheng Yan, Yun Wang, Xiaoyan Luo, Yu-Wing Tai

Main category: cs.CV

TL;DR: PS-ReID is a multimodal person re-identification model that combines image and text inputs to overcome limitations of traditional image-based methods, using dual-path encoding and token-level supervision for improved performance in challenging scenarios.

DetailsMotivation: Traditional image-based ReID methods struggle with occlusions and lighting changes, while text provides complementary information. The integration of both modalities remains underexplored, especially in full-scene settings.

Method: Proposes PS-ReID with dual-path asymmetric encoding: query branch captures identity-discriminative cues, target branch performs holistic scene reasoning. Uses token-level ReID loss to supervise identity-aware tokens, coupling retrieval and segmentation. Built M2ReID dataset with 200K+ images and 4,894 identities.

Result: PS-ReID significantly outperforms unimodal query-based models in both ReID and segmentation tasks. Excels in challenging scenarios like occlusion, low lighting, and background clutter.

Conclusion: The model offers a robust and flexible solution for person retrieval and segmentation, demonstrating the effectiveness of multimodal integration in ReID tasks. All resources will be publicly available.

Abstract: Person re-identification (ReID) plays a critical role in applications such as security surveillance and criminal investigations. Most traditional image-based ReID methods face challenges including occlusions and lighting changes, while text provides complementary information to mitigate these issues. However, the integration of both image and text modalities remains underexplored. To address this gap, we propose {\bf PS-ReID}, a multimodal model that combines image and text inputs to enhance ReID performance. In contrast to existing ReID methods limited by cropped pedestrian images, our PS-ReID focuses on full-scene settings and introduces a multimodal ReID task that incorporates segmentation, enabling precise feature extraction of the queried individual, even under challenging conditions such as occlusion. To this end, our model adopts a dual-path asymmetric encoding scheme that explicitly separates query and target roles: the query branch captures identity-discriminative cues, while the target branch performs holistic scene reasoning. Additionally, a token-level ReID loss supervises identity-aware tokens, coupling retrieval and segmentation to yield masks that are both spatially precise and identity-consistent. To facilitate systematic evaluation, we construct M2ReID, currently the largest full-scene multimodal ReID dataset, with over 200K images and 4,894 identities, featuring multimodal queries and high-quality segmentation masks. Experimental results demonstrate that PS-ReID significantly outperforms unimodal query-based models in both ReID and segmentation tasks. The model excels in challenging real-world scenarios such as occlusion, low lighting, and background clutter, offering a robust and flexible solution for person retrieval and segmentation. All code, models, and datasets will be publicly available.

[509] A Hybrid Fully Convolutional CNN-Transformer Model for Inherently Interpretable Disease Detection from Retinal Fundus Images

Kerol Djoumessi, Samuel Ofosu Mensah, Philipp Berens

Main category: cs.CV

TL;DR: Interpretable hybrid CNN-Transformer architecture for retinal disease detection that generates faithful evidence maps directly reflecting model decisions, achieving state-of-the-art performance.

DetailsMotivation: Hybrid CNN-Transformer models combine local feature extraction and global dependencies but lack interpretability, which is crucial for medical imaging applications where understanding model decisions is essential.

Method: Developed an interpretable-by-design hybrid fully convolutional CNN-Transformer architecture that generates class-specific sparse evidence maps in a single forward pass, unlike post-hoc saliency methods.

Result: Achieved state-of-the-art predictive performance on retinal disease detection tasks using color fundus images, outperforming both black-box and interpretable models while providing faithful localized evidence.

Conclusion: The proposed architecture successfully combines the strengths of CNNs and Transformers while maintaining interpretability, making it suitable for medical imaging applications where both performance and transparency are critical.

Abstract: In many medical imaging tasks, convolutional neural networks (CNNs) efficiently extract local features hierarchically. More recently, vision transformers (ViTs) have gained popularity, using self-attention mechanisms to capture global dependencies, but lacking the inherent spatial localization of convolutions. Therefore, hybrid models combining CNNs and ViTs have been developed to combine the strengths of both architectures. However, such hybrid models are difficult to interpret, which hinders their application in medical imaging. In this work, we introduce an interpretable-by-design hybrid fully convolutional CNN-Transformer architecture for retinal disease detection. Unlike widely used post-hoc saliency methods for ViTs, our approach generates faithful and localized evidence maps that directly reflect the mode’s decision process. We evaluated our method on two medical tasks focused on disease detection using color fundus images. Our model achieves state-of-the-art predictive performance compared to black-box and interpretable models and provides class-specific sparse evidence maps in a single forward pass. The code is available at: https://github.com/kdjoumessi/Self-Explainable-CNN-Transformer.

[510] A Multi-Stage Auto-Context Deep Learning Framework for Tissue and Nuclei Segmentation and Classification in H&E-Stained Histological Images of Advanced Melanoma

Nima Torbati, Anastasia Meshcheryakova, Ramona Woitek, Sepideh Hatamikia, Diana Mechtcheriakova, Amirreza Mahbod

Main category: cs.CV

TL;DR: A multi-stage deep learning approach that combines tissue and nuclei information in a unified framework for melanoma histological image analysis, achieving top rankings in the PUMA challenge with scores of 73.40% and 63.48% for different tracks.

DetailsMotivation: Melanoma diagnosis relies on histological image analysis, but existing computerized approaches treat tissue-based and nuclei-based analysis as separate tasks, which may be suboptimal for accurate diagnosis.

Method: Proposed a novel multi-stage deep learning approach using auto-context concept to unify tissue and nuclei information. Includes pre-training and post-processing stages to perform segmentation and classification in melanoma histological images.

Result: Achieved second and first place rankings in PUMA challenge with average micro Dice tissue score of 73.40% for Track 1 and summed nuclei F1-score of 63.48% for Track 2. Demonstrated effectiveness through ablation study and generalization on external dataset.

Conclusion: The unified framework combining tissue and nuclei information proves effective for melanoma histological image analysis, showing strong performance in segmentation and classification tasks with good generalization capabilities.

Abstract: Melanoma is the most lethal form of skin cancer, with an increasing incidence rate worldwide. Analyzing histological images of melanoma by localizing and classifying tissues and cell nuclei is considered the gold standard method for diagnosis and treatment options for patients. While many computerized approaches have been proposed for automatic analysis, most perform tissue-based analysis and nuclei (cell)-based analysis as separate tasks, which might be suboptimal. In this work, using the PUMA challenge dataset, we propose a novel multi-stage deep learning approach by combining tissue and nuclei information in a unified framework based on the auto-context concept to perform segmentation and classification in histological images of melanoma. Through pre-training and further post-processing, our approach achieved second and first place rankings in the PUMA challenge, with average micro Dice tissue score and summed nuclei F1-score of 73.40% for Track 1 and 63.48% for Track 2, respectively. Furthermore, through a comprehensive ablation study and additional evaluation on an external dataset, we demonstrated the effectiveness of the framework components as well as the generalization capabilities of the proposed approach. Our implementation for training and testing is available at: https://github.com/NimaTorbati/PumaSubmit

[511] Leveraging Modality Tags for Enhanced Cross-Modal Video Retrieval

Adriano Fragomeni, Dima Damen, Michael Wray

Main category: cs.CV

TL;DR: MAC-VR uses modality-specific tags from foundation models to enhance video retrieval by aligning visual and textual concepts in latent space, outperforming state-of-the-art methods on multiple datasets.

DetailsMotivation: Video retrieval requires better alignment between visual content and natural language descriptions. Current methods need improvement in cross-modal alignment to distinguish concepts more effectively.

Method: Leverages automatically extracted modality-specific tags from foundation models, aligns modalities in latent space, and learns auxiliary latent concepts from video and caption features to improve concept distinction.

Result: Outperforms current state-of-the-art methods across three datasets (MSR-VTT splits, DiDeMo, TGIF, Charades, YouCook2) and performs comparably or better on others, demonstrating improved cross-modal alignment.

Conclusion: Modality-specific tags significantly enhance video retrieval by improving the alignment of visual and textual latent concepts, allowing better concept distinction and superior performance across diverse datasets.

Abstract: Video retrieval requires aligning visual content with corresponding natural language descriptions. In this paper, we introduce Modality Auxiliary Concepts for Video Retrieval (MAC-VR), a novel approach that leverages modality-specific tags – automatically extracted from foundation models – to enhance video retrieval. We propose to align modalities in a latent space, along with learning and aligning auxiliary latent concepts derived from the features of a video and its corresponding caption. We introduce these auxiliary concepts to improve the alignment of visual and textual latent concepts, allowing concepts to be distinguished from one another. We conduct extensive experiments on six diverse datasets: two different splits of MSR-VTT, DiDeMo, TGIF, Charades and YouCook2. The experimental results consistently demonstrate that modality-specific tags improve cross-modal alignment, outperforming current state-of-the-art methods across three datasets and performing comparably or better across others. Project Webpage: https://adrianofragomeni.github.io/MAC-VR/

[512] Shadow Erosion and Nighttime Adaptability for Camera-Based Automated Driving Applications

Mohamed Sabry, Gregory Schroeder, Joshua Varughese, Cristina Olaverri-Monreal

Main category: cs.CV

TL;DR: Proposes a pipeline for Shadow Erosion and Nighttime Adaptability to enhance RGB images for autonomous driving, outperforming CLAHE technique in illumination uniformity and visual quality metrics.

DetailsMotivation: RGB image enhancement is crucial for autonomous driving applications to handle challenging lighting conditions like poor nighttime visibility and shadow effects in bright daylight.

Method: Developed a Shadow Erosion and Nighttime Adaptability pipeline that preserves color and texture details, compared against the widely used CLAHE technique.

Result: Significant improvement over CLAHE in both illumination uniformity and visual perception quality metrics, and enhanced performance of YOLO-based drivable area segmentation algorithm.

Conclusion: The proposed pipeline effectively addresses shadow and nighttime challenges in autonomous driving imagery, providing superior image enhancement compared to traditional methods.

Abstract: Enhancement of images from RGB cameras is of particular interest due to its wide range of ever-increasing applications such as medical imaging, satellite imaging, automated driving, etc. In autonomous driving, various techniques are used to enhance image quality under challenging lighting conditions. These include artificial augmentation to improve visibility in poor nighttime conditions, illumination-invariant imaging to reduce the impact of lighting variations, and shadow mitigation to ensure consistent image clarity in bright daylight. This paper proposes a pipeline for Shadow Erosion and Nighttime Adaptability in images for automated driving applications while preserving color and texture details. The Shadow Erosion and Nighttime Adaptability pipeline is compared to the widely used CLAHE technique and evaluated based on illumination uniformity and visual perception quality metrics. The results also demonstrate a significant improvement over CLAHE, enhancing a YOLO-based drivable area segmentation algorithm.

[513] Cognitive-Inspired Hierarchical Attention Fusion With Visual and Textual for Cross-Domain Sequential Recommendation

Wangyu Wu, Zhenhong Chen, Siqi Song, Xianglin Qiu, Xiaowei Huang, Fei Ma, Jimin Xiao

Main category: cs.CV

TL;DR: HAF-VT integrates visual and textual data using CLIP embeddings and hierarchical attention to model cross-domain user preferences in sequential recommendation, outperforming existing methods on e-commerce datasets.

DetailsMotivation: To enhance cross-domain sequential recommendation by modeling human cognitive processes through multimodal data integration, addressing the limitations of existing methods in capturing cross-domain user interests.

Method: Uses frozen CLIP model to generate image and text embeddings, employs hierarchical attention mechanism to jointly learn single-domain and cross-domain preferences, mimicking human information integration processes.

Result: Outperforms existing methods on four e-commerce datasets, demonstrating superior capability in capturing cross-domain user interests and sequential decision-making patterns.

Conclusion: Successfully bridges cognitive principles with computational models, highlighting the significant role of multimodal data in enhancing cross-domain sequential recommendation systems.

Abstract: Cross-Domain Sequential Recommendation (CDSR) predicts user behavior by leveraging historical interactions across multiple domains, focusing on modeling cross-domain preferences through intra- and inter-sequence item relationships. Inspired by human cognitive processes, we propose Hierarchical Attention Fusion of Visual and Textual Representations (HAF-VT), a novel approach integrating visual and textual data to enhance cognitive modeling. Using the frozen CLIP model, we generate image and text embeddings, enriching item representations with multimodal data. A hierarchical attention mechanism jointly learns single-domain and cross-domain preferences, mimicking human information integration. Evaluated on four e-commerce datasets, HAF-VT outperforms existing methods in capturing cross-domain user interests, bridging cognitive principles with computational models and highlighting the role of multimodal data in sequential decision-making.

[514] Boosting Generative Image Modeling via Joint Image-Feature Synthesis

Theodoros Kouzelis, Efstathios Karypidis, Ioannis Kakogeorgiou, Spyros Gidaris, Nikos Komodakis

Main category: cs.CV

TL;DR: A novel diffusion framework that jointly models low-level image latents and high-level semantic features, improving image generation quality and training efficiency while enabling semantic guidance.

DetailsMotivation: To bridge the gap between representation learning and generative modeling in latent diffusion models, addressing the challenge of integrating semantic understanding with high-quality image generation.

Method: Leverages a diffusion model to jointly model low-level image latents from a VAE and high-level semantic features from a pretrained self-supervised encoder (like DINO), learning to generate coherent image-feature pairs from noise with minimal modifications to standard Diffusion Transformer architectures.

Result: Significant improvements in both image quality and training convergence speed in conditional and unconditional settings, while eliminating the need for complex distillation objectives.

Conclusion: Establishes a new direction for representation-aware generative modeling with a unified design that simplifies training and enables powerful semantic guidance during inference.

Abstract: Latent diffusion models (LDMs) dominate high-quality image generation, yet integrating representation learning with generative modeling remains a challenge. We introduce a novel generative image modeling framework that seamlessly bridges this gap by leveraging a diffusion model to jointly model low-level image latents (from a variational autoencoder) and high-level semantic features (from a pretrained self-supervised encoder like DINO). Our latent-semantic diffusion approach learns to generate coherent image-feature pairs from pure noise, significantly enhancing both generative quality and training efficiency, all while requiring only minimal modifications to standard Diffusion Transformer architectures. By eliminating the need for complex distillation objectives, our unified design simplifies training and unlocks a powerful new inference strategy: Representation Guidance, which leverages learned semantics to steer and refine image generation. Evaluated in both conditional and unconditional settings, our method delivers substantial improvements in image quality and training convergence speed, establishing a new direction for representation-aware generative modeling. Project page and code: https://representationdiffusion.github.io

[515] Depth3DLane: Monocular 3D Lane Detection via Depth Prior Distillation

Dongxin Lyu, Han Huang, Cheng Tan, Zimu Li

Main category: cs.CV

TL;DR: A novel BEV-based framework for monocular 3D lane detection that addresses IPM limitations through hierarchical depth features, depth prior distillation, and spatial coherence enforcement.

DetailsMotivation: Overcome limitations of inverse perspective mapping (IPM) in monocular 3D lane detection, particularly the flat-ground assumption and loss of contextual information that cause inaccuracies in 3D reconstruction, especially height estimation.

Method: Proposes three key components: 1) Hierarchical Depth-Aware Head for multi-scale depth features to mitigate flat-ground assumption, 2) Depth Prior Distillation to transfer semantic depth knowledge from teacher model, and 3) Conditional Random Field module to enforce spatial coherence and smooth lane reconstruction.

Result: Achieves state-of-the-art performance in terms of z-axis error and outperforms other methods in overall performance, as validated through extensive experiments.

Conclusion: The proposed framework effectively addresses the challenges of monocular 3D lane detection by enhancing depth awareness, leveraging semantic depth knowledge, and ensuring spatial coherence, leading to improved accuracy in 3D lane reconstruction.

Abstract: Monocular 3D lane detection is challenging due to the difficulty in capturing depth information from single-camera images. A common strategy involves transforming front-view (FV) images into bird’s-eye-view (BEV) space through inverse perspective mapping (IPM), facilitating lane detection using BEV features. However, IPM’s flat-ground assumption and loss of contextual information lead to inaccuracies in reconstructing 3D information, especially height. In this paper, we introduce a BEV-based framework to address these limitations and improve 3D lane detection accuracy. Our approach incorporates a Hierarchical Depth-Aware Head that provides multi-scale depth features, mitigating the flat-ground assumption by enhancing spatial awareness across varying depths. Additionally, we leverage Depth Prior Distillation to transfer semantic depth knowledge from a teacher model, capturing richer structural and contextual information for complex lane structures. To further refine lane continuity and ensure smooth lane reconstruction, we introduce a Conditional Random Field module that enforces spatial coherence in lane predictions. Extensive experiments validate that our method achieves state-of-the-art performance in terms of z-axis error and outperforms other methods in the field in overall performance. The code is released at: https://anonymous.4open.science/r/Depth3DLane-DCDD.

[516] Automated Parsing of Engineering Drawings for Structured Information Extraction Using a Fine-tuned Document Understanding Transformer

Muhammad Tayyab Khan, Zane Yong, Lequn Chen, Jun Ming Tan, Wenhe Feng, Seung Ki Moon

Main category: cs.CV

TL;DR: A hybrid deep learning framework combining YOLOv11 OBB detection with Donut transformer achieves high-precision structured information extraction from 2D engineering drawings, outperforming category-specific models with 97.3% F1 score.

DetailsMotivation: Manual extraction of key information from 2D engineering drawings is time-consuming and error-prone, while traditional OCR struggles with complex layouts and overlapping symbols, producing unstructured outputs that hinder precision manufacturing.

Method: Proposes a hybrid framework integrating oriented bounding box (OBB) detection using YOLOv11 for 9 key categories (GD&T, tolerances, measures, materials, etc.) with transformer-based Donut model for structured JSON output. Uses in-house annotated dataset and compares single model vs category-specific fine-tuning strategies.

Result: Single model consistently outperforms category-specific models across all metrics: achieves 94.77% precision for GD&T, 100% recall for most categories, 97.3% F1 score, while reducing hallucination to 5.23%.

Conclusion: The proposed hybrid framework significantly improves accuracy, reduces manual effort, and supports scalable deployment in precision-driven industries by providing structured information extraction from complex engineering drawings.

Abstract: Accurate extraction of key information from 2D engineering drawings is crucial for high-precision manufacturing. Manual extraction is time-consuming and error-prone, while traditional Optical Character Recognition (OCR) techniques often struggle with complex layouts and overlapping symbols, resulting in unstructured outputs. To address these challenges, this paper proposes a novel hybrid deep learning framework for structured information extraction by integrating an oriented bounding box (OBB) detection model with a transformer-based document parsing model (Donut). An in-house annotated dataset is used to train YOLOv11 for detecting nine key categories: Geometric Dimensioning and Tolerancing (GD&T), General Tolerances, Measures, Materials, Notes, Radii, Surface Roughness, Threads, and Title Blocks. Detected OBBs are cropped into images and labeled to fine-tune Donut for structured JSON output. Fine-tuning strategies include a single model trained across all categories and category-specific models. Results show that the single model consistently outperforms category-specific ones across all evaluation metrics, achieving higher precision (94.77% for GD&T), recall (100% for most), and F1 score (97.3%), while reducing hallucination (5.23%). The proposed framework improves accuracy, reduces manual effort, and supports scalable deployment in precision-driven industries.

[517] FedMVP: Federated Multimodal Visual Prompt Tuning for Vision-Language Models

Mainak Singha, Subhankar Roy, Sarthak Mehrotra, Ankit Jha, Moloud Abdar, Biplab Banerjee, Elisa Ricci

Main category: cs.CV

TL;DR: FedMVP proposes multimodal visual prompt tuning for federated learning, using both image and textual features to generate dynamic prompts that improve generalization to unseen classes and domains.

DetailsMotivation: Textual prompt tuning in federated learning suffers from overfitting to known concepts, limiting generalizability to unseen concepts. The paper aims to address this limitation by incorporating multimodal contextual information.

Method: FedMVP uses a PromptFormer module with cross-attention to align textual and visual features, generating multimodal visual prompts. These prompts are input to CLIP’s frozen vision encoder and trained with CLIP similarity loss plus consistency loss.

Result: Extensive evaluation on 20 datasets across three generalization settings shows FedMVP preserves in-distribution performance while achieving +1.57%-2.26% better generalization to unseen classes/domains compared to state-of-the-art methods.

Conclusion: Multimodal visual prompt tuning with cross-attention alignment effectively addresses overfitting issues in federated learning, significantly improving generalization capabilities while maintaining performance on known concepts.

Abstract: In federated learning, textual prompt tuning adapts Vision-Language Models (e.g., CLIP) by tuning lightweight input tokens (or prompts) on local client data, while keeping network weights frozen. After training, only the prompts are shared by the clients with the central server for aggregation. However, textual prompt tuning suffers from overfitting to known concepts, limiting its generalizability to unseen concepts. To address this limitation, we propose Multimodal Visual Prompt Tuning (FedMVP) that conditions the prompts on multimodal contextual information - derived from the input image and textual attribute features of a class. At the core of FedMVP is a PromptFormer module that synergistically aligns textual and visual features through a cross-attention mechanism. The dynamically generated multimodal visual prompts are then input to the frozen vision encoder of CLIP, and trained with a combination of CLIP similarity loss and a consistency loss. Extensive evaluation on 20 datasets, spanning three generalization settings, demonstrates that FedMVP not only preserves performance on in-distribution classes and domains, but also displays higher generalizability to unseen classes and domains, surpassing state-of-the-art methods by a notable margin of +1.57% - 2.26%. Code is available at https://github.com/mainaksingha01/FedMVP.

[518] A Rate-Quality Model for Learned Video Coding

Sang NguyenQuang, Cheng-Wei Chen, Xiem HoangVan, Wen-Hsiao Peng

Main category: cs.CV

TL;DR: Proposes RQNet, a neural network that models rate-quality relationship for learned video coding, enabling online parameter adaptation with improved accuracy and minimal complexity.

DetailsMotivation: Learned video coding has shown superior performance but needs better modeling of the rate-quality relationship to enhance flexibility and precision in video compression.

Method: Develop RQNet neural network to characterize bitrate-quality relationship based on video content and coding context, integrate with least-squares method using previous frames for online parameter adaptation.

Result: Achieves significantly smaller bitrate deviations than baseline methods on common datasets with minimal additional computational complexity.

Conclusion: The proposed R-Q model provides accurate rate-quality estimation and enables effective online parameter adaptation, improving learned video coding performance.

Abstract: Learned video coding (LVC) has recently achieved superior coding performance. In this paper, we model the rate-quality (R-Q) relationship for learned video coding by a parametric function. We learn a neural network, termed RQNet, to characterize the relationship between the bitrate and quality level according to video content and coding context. The predicted (R,Q) results are further integrated with those from previously coded frames using the least-squares method to determine the parameters of our R-Q model on-the-fly. Compared to the conventional approaches, our method accurately estimates the R-Q relationship, enabling the online adaptation of model parameters to enhance both flexibility and precision. Experimental results show that our R-Q model achieves significantly smaller bitrate deviations than the baseline method on commonly used datasets with minimal additional complexity.

[519] Uncertainty-Aware Prototype Semantic Decoupling for Text-Based Person Search in Full Images

Zengli Luo, Canlong Zhang, Zhixin Li, Zhiwen Wang, Chunrong Wei

Main category: cs.CV

TL;DR: UPD-TBPS is a novel framework for text-based pedestrian search that addresses uncertainties in detection and matching through multi-granularity uncertainty estimation, prototype-based decoupling, and cross-modal re-identification.

DetailsMotivation: Existing text-based pedestrian search methods struggle with uncertainties in detection and matching in complex scenes with multiple pedestrians, leading to degraded performance.

Method: Three-module framework: 1) Multi-granularity Uncertainty Estimation (MUE) for identifying potential targets with confidence scores, 2) Prototype-based Uncertainty Decoupling (PUD) for visual context decoupling and prototype mining at cluster and individual levels, 3) Cross-modal Re-identification for evaluating candidates with varying confidence levels.

Result: Experiments on CUHK-SYSU-TBPS and PRW-TBPS datasets validate the effectiveness of the proposed framework.

Conclusion: UPD-TBPS successfully addresses uncertainties in text-based pedestrian search through its three-module approach, improving detection and retrieval accuracy in complex scenes.

Abstract: Text-based pedestrian search (TBPS) in full images aims to locate a target pedestrian in untrimmed images using natural language descriptions. However, in complex scenes with multiple pedestrians, existing methods are limited by uncertainties in detection and matching, leading to degraded performance. To address this, we propose UPD-TBPS, a novel framework comprising three modules: Multi-granularity Uncertainty Estimation (MUE), Prototype-based Uncertainty Decoupling (PUD), and Cross-modal Re-identification (ReID). MUE conducts multi-granularity queries to identify potential targets and assigns confidence scores to reduce early-stage uncertainty. PUD leverages visual context decoupling and prototype mining to extract features of the target pedestrian described in the query. It separates and learns pedestrian prototype representations at both the coarse-grained cluster level and the fine-grained individual level, thereby reducing matching uncertainty. ReID evaluates candidates with varying confidence levels, improving detection and retrieval accuracy. Experiments on CUHK-SYSU-TBPS and PRW-TBPS datasets validate the effectiveness of our framework.

[520] Multi-Agent System for Comprehensive Soccer Understanding

Jiayuan Rao, Zifeng Li, Haoning Wu, Ya Zhang, Yanfeng Wang, Weidi Xie

Main category: cs.CV

TL;DR: A comprehensive framework for holistic soccer understanding featuring a multimodal knowledge base (SoccerWiki), a large benchmark (SoccerBench), and a multi-agent reasoning system (SoccerAgent) that outperforms existing MLLMs.

DetailsMotivation: Existing soccer understanding research focuses on isolated or narrow tasks, creating a gap that needs a holistic approach to enable comprehensive soccer analysis and reasoning.

Method: Constructed SoccerWiki (multimodal knowledge base), created SoccerBench (10K multimodal QA benchmark), and developed SoccerAgent (multi-agent system for collaborative reasoning using domain expertise).

Result: The proposed framework achieves superior performance compared to representative multimodal large language models (MLLMs) on the comprehensive SoccerBench benchmark.

Conclusion: The holistic soccer understanding framework successfully bridges the gap in existing research by providing integrated knowledge, comprehensive evaluation, and robust agentic reasoning capabilities for complex soccer analysis.

Abstract: Recent advances in soccer understanding have demonstrated rapid progress, yet existing research predominantly focuses on isolated or narrow tasks. To bridge this gap, we propose a comprehensive framework for holistic soccer understanding. Concretely, we make the following contributions in this paper: (i) we construct SoccerWiki, the first large-scale multimodal soccer knowledge base, integrating rich domain knowledge about players, teams, referees, and venues to enable knowledge-driven reasoning; (ii) we present SoccerBench, the largest and most comprehensive soccer-specific benchmark, featuring around 10K multimodal (text, image, video) multi-choice QA pairs across 13 distinct tasks; (iii) we introduce SoccerAgent, a novel multi-agent system that decomposes complex soccer questions via collaborative reasoning, leveraging domain expertise from SoccerWiki and achieving robust performance; (iv) extensive evaluations and comparisons with representative MLLMs on SoccerBench highlight the superiority of our agentic system.

[521] A Large-scale Benchmark on Geological Fault Delineation Models: Domain Shift, Training Dynamics, Generalizability, Evaluation and Inferential Behavior

Jorge Quesada, Chen Zhou, Prithwijit Chowdhury, Mohammad Alotaibi, Ahmad Mustafa, Yusufjon Kumakov, Mohit Prabhushankar, Ghassan AlRegib

Main category: cs.CV

TL;DR: First large-scale benchmarking study on domain shift strategies for seismic fault delineation models, analyzing 200+ model-dataset combinations across synthetic and real data to provide guidelines for reliable deployment.

DetailsMotivation: Lack of systematic understanding of model generalizability limits across diverse seismic data settings, with distributional shifts, fine-tuning limitations, and inconsistent evaluation protocols hindering real-world deployment.

Method: Benchmark spanning 200+ combinations of model architectures, datasets and training strategies across three datasets (FaultSeg3D, CRACKS, Thebe), systematically assessing pretraining, fine-tuning, and joint training under varying domain shifts.

Result: Fine-tuning practices can cause catastrophic forgetting with disjoint datasets; larger models like Segformer are more robust; domain adaptation outperforms fine-tuning for large shifts but underperforms for similar domains; novel fault characteristic analysis reveals structural biases.

Conclusion: Establishes robust experimental baseline providing insights into tradeoffs in fault delineation workflows and highlights directions for building more generalizable and interpretable seismic interpretation models.

Abstract: Machine learning has taken a critical role in seismic interpretation workflows, especially in fault delineation tasks. However, despite the recent proliferation of pretrained models and synthetic datasets, the field still lacks a systematic understanding of the generalizability limits of these models across seismic data representing diverse geologic, acquisition and processing settings. Distributional shifts between data sources, limitations in fine-tuning strategies and labeled data accessibility, and inconsistent evaluation protocols all remain major roadblocks to deploying reliable models in real-world exploration. In this paper, we present the first large-scale benchmarking study explicitly designed to provide guidelines for domain shift strategies in seismic interpretation. Our benchmark spans over 200 combinations of model architectures, datasets and training strategies, across three datasets (synthetic and real) including FaultSeg3D, CRACKS, and Thebe. We systematically assess pretraining, fine-tuning, and joint training under varying domain shifts. Our analysis shows that common fine-tuning practices can lead to catastrophic forgetting, especially when source and target datasets are disjoint, and that larger models such as Segformer are more robust than smaller architectures. We also find that domain adaptation methods outperform fine-tuning when shifts are large, yet underperform when domains are similar. Finally, we complement segmentation metrics with a novel analysis based on fault characteristic descriptors, revealing how models absorb structural biases from training datasets. Overall, we establish a robust experimental baseline that provides insights into tradeoffs in current fault delineation workflows and highlights directions for building more generalizable and interpretable models.

[522] ViEEG: Hierarchical Visual Neural Representation for EEG Brain Decoding

Minxu Liu, Donghai Guan, Chuhang Zheng, Chunwei Tian, Jie Wen, Qi Zhu

Main category: cs.CV

TL;DR: ViEEG is a hierarchical EEG visual decoding framework that addresses the neglect of hierarchical neural encoding in brain activity decoding, achieving superior performance in both subject-dependent and subject-independent settings.

DetailsMotivation: Existing EEG visual decoding methods suffer from Hierarchical Neural Encoding Neglect (HNEN), failing to model the brain's hierarchical visual processing structure, which limits their effectiveness in decoding brain activity into visual representations.

Method: ViEEG decomposes visual stimuli into three biologically aligned components (contour, foreground object, contextual scene) and uses a three-stream EEG encoder with cross-attention routing to progressively integrate features, simulating cortical information flow from low-level to high-level vision.

Result: Extensive experiments on THINGS-EEG dataset show ViEEG significantly outperforms previous methods by a large margin. Results on THINGS-MEG dataset confirm generalization to different neural modalities.

Conclusion: ViEEG not only advances performance but also sets a new paradigm for EEG brain decoding by properly modeling the brain’s hierarchical visual processing structure.

Abstract: Understanding and decoding brain activity into visual representations is a fundamental challenge at the intersection of neuroscience and artificial intelligence. While EEG visual decoding has shown promise due to its non-invasive, and low-cost nature, existing methods suffer from Hierarchical Neural Encoding Neglect (HNEN)-a critical limitation where flat neural representations fail to model the brain’s hierarchical visual processing hierarchy. Inspired by the hierarchical organization of visual cortex, we propose ViEEG, a neuro-We further adopt hierarchical contrastive learning for EEG-CLIP representation alignment, enabling zero-shot object recognition. Extensive experiments on the THINGS-EEG dataset demonstrate that ViEEG significantly outperforms previous methods by a large margin in both subject-dependent and subject-independent settings. Results on the THINGS-MEG dataset further confirm ViEEG’s generalization to different neural modalities. Our framework not only advances the performance frontier but also sets a new paradigm for EEG brain decoding. inspired framework that addresses HNEN. ViEEG decomposes each visual stimulus into three biologically aligned components-contour, foreground object, and contextual scene-serving as anchors for a three-stream EEG encoder. These EEG features are progressively integrated via cross-attention routing, simulating cortical information flow from low-level to high-level vision.

[523] BusterX: MLLM-Powered AI-Generated Video Forgery Detection and Explanation

Haiquan Wen, Yiwei He, Zhenglin Huang, Tianxiao Li, Zihan Yu, Xingru Huang, Lu Qi, Baoyuan Wu, Xiangtai Li, Guangliang Cheng

Main category: cs.CV

TL;DR: Proposes GenBuster-200K dataset (200K AI-generated videos) and BusterX framework for explainable deepfake detection using MLLM and reinforcement learning.

DetailsMotivation: Address the lack of large-scale AI-generated video datasets and the need for explainable detection methods beyond binary classification to combat misinformation risks from advanced video generation models.

Method: Created GenBuster-200K dataset with diverse generative techniques and real-world scenes. Developed BusterX framework combining multimodal large language model (MLLM) with reinforcement learning for detection and explainable rationale.

Result: Extensive comparisons show BusterX’s effectiveness and generalizability. The framework provides both authenticity determination and explainable decision-making.

Conclusion: First large-scale AI-generated video dataset with latest techniques and first explainable detection framework using MLLM+RL, addressing critical gaps in deepfake video detection and explanation.

Abstract: Advances in AI generative models facilitate super-realistic video synthesis, amplifying misinformation risks via social media and eroding trust in digital content. Several research works have explored new deepfake detection methods on AI-generated images to alleviate these risks. However, with the fast development of video generation models, such as Sora and WanX, there is currently a lack of large-scale, high-quality AI-generated video datasets for forgery detection. In addition, existing detection approaches predominantly treat the task as binary classification, lacking explainability in model decision-making and failing to provide actionable insights or guidance for the public. To address these challenges, we propose \textbf{GenBuster-200K}, a large-scale AI-generated video dataset featuring 200K high-resolution video clips, diverse latest generative techniques, and real-world scenes. We further introduce \textbf{BusterX}, a novel AI-generated video detection and explanation framework leveraging multimodal large language model (MLLM) and reinforcement learning for authenticity determination and explainable rationale. To our knowledge, GenBuster-200K is the {\it \textbf{first}} large-scale, high-quality AI-generated video dataset that incorporates the latest generative techniques for real-world scenarios. BusterX is the {\it \textbf{first}} framework to integrate MLLM with reinforcement learning for explainable AI-generated video detection. Extensive comparisons with state-of-the-art methods and ablation studies validate the effectiveness and generalizability of BusterX. The code, models, and datasets will be released.

[524] DiffDecompose: Layer-Wise Decomposition of Alpha-Composited Images via Diffusion Transformers

Zitong Wang, Hang Zhao, Qianyu Zhou, Xuequan Lu, Xiangtai Li, Yiren Song

Main category: cs.CV

TL;DR: DiffDecompose: A diffusion Transformer framework for layer-wise decomposition of alpha-composited images, addressing semi-transparent/transparent layer separation with a new dataset AlphaBlend.

DetailsMotivation: Existing image decomposition methods struggle with semi-transparent/transparent layer occlusions due to mask prior dependencies, static object assumptions, and lack of appropriate datasets.

Method: Proposes DiffDecompose - a diffusion Transformer-based framework that learns posterior over possible layer decompositions using input image, semantic prompts, and blending type. Uses In-Context Decomposition and Layer Position Encoding Cloning to maintain pixel correspondence.

Result: Extensive experiments on AlphaBlend dataset (6 real-world subtasks) and public LOGO dataset verify effectiveness. Outperforms existing methods in transparent/semi-transparent layer decomposition.

Conclusion: The approach successfully addresses layer ambiguity, generalization, and data scarcity challenges in alpha-composited image decomposition, providing a robust solution for various real-world applications.

Abstract: Diffusion models have recently motivated great success in many generation tasks like object removal. Nevertheless, existing image decomposition methods struggle to disentangle semi-transparent or transparent layer occlusions due to mask prior dependencies, static object assumptions, and the lack of datasets. In this paper, we delve into a novel task: Layer-Wise Decomposition of Alpha-Composited Images, aiming to recover constituent layers from single overlapped images under the condition of semi-transparent/transparent alpha layer non-linear occlusion. To address challenges in layer ambiguity, generalization, and data scarcity, we first introduce AlphaBlend, the first large-scale and high-quality dataset for transparent and semi-transparent layer decomposition, supporting six real-world subtasks (e.g., translucent flare removal, semi-transparent cell decomposition, glassware decomposition). Building on this dataset, we present DiffDecompose, a diffusion Transformer-based framework that learns the posterior over possible layer decompositions conditioned on the input image, semantic prompts, and blending type. Rather than regressing alpha mattes directly, DiffDecompose performs In-Context Decomposition, enabling the model to predict one or multiple layers without per-layer supervision, and introduces Layer Position Encoding Cloning to maintain pixel-level correspondence across layers. Extensive experiments on the proposed AlphaBlend dataset and public LOGO dataset verify the effectiveness of DiffDecompose. The code and dataset will be available upon paper acceptance. Our code will be available at: https://github.com/Wangzt1121/DiffDecompose.

[525] RUSplatting: Robust 3D Gaussian Splatting for Sparse-View Underwater Scene Reconstruction

Zhuodong Jiang, Haoran Wang, Guoxi Huang, Brett Seymour, Nantheera Anantrasirichai

Main category: cs.CV

TL;DR: Enhanced Gaussian Splatting framework for high-fidelity underwater scene reconstruction with improved color restoration, view consistency, and noise reduction.

DetailsMotivation: Underwater scene reconstruction is challenging due to light absorption, scattering, and limited visibility in aquatic environments, requiring specialized approaches for accurate rendering.

Method: Proposes decoupled RGB channel learning guided by underwater physics, frame interpolation with adaptive weighting, and a novel loss function for noise reduction and edge preservation.

Result: Outperforms state-of-the-art methods with PSNR gains up to 1.90dB, delivering superior perceptual quality and robustness in deep-sea environments.

Conclusion: The framework offers promising directions for marine robotics and underwater visual analytics, with code and new Submerged3D dataset publicly available.

Abstract: Reconstructing high-fidelity underwater scenes remains a challenging task due to light absorption, scattering, and limited visibility inherent in aquatic environments. This paper presents an enhanced Gaussian Splatting-based framework that improves both the visual quality and geometric accuracy of deep underwater rendering. We propose decoupled learning for RGB channels, guided by the physics of underwater attenuation, to enable more accurate colour restoration. To address sparse-view limitations and improve view consistency, we introduce a frame interpolation strategy with a novel adaptive weighting scheme. Additionally, we introduce a new loss function aimed at reducing noise while preserving edges, which is essential for deep-sea content. We also release a newly collected dataset, Submerged3D, captured specifically in deep-sea environments. Experimental results demonstrate that our framework consistently outperforms state-of-the-art methods with PSNR gains up to 1.90dB, delivering superior perceptual quality and robustness, and offering promising directions for marine robotics and underwater visual analytics. The code of RUSplatting is available at https://github.com/theflash987/RUSplatting and the dataset Submerged3D can be downloaded at https://zenodo.org/records/15482420.

[526] Can NeRFs See without Cameras?

Chaitanya Amballa, Sattwik Basu, Yu-Lin Wei, Zhijian Yang, Mehmet Ergezer, Romit Roy Choudhury

Main category: cs.CV

TL;DR: NeRFs can be adapted to learn from multipath signals like WiFi to infer indoor environments, enabling floorplan reconstruction from sparse measurements.

DetailsMotivation: Traditional NeRFs work with optical rays but RF/audio signals contain multipath reflections. The paper explores whether environment inference is possible using such complex signal mixtures.

Method: Redesigned Neural Radiance Fields (NeRFs) that can learn from multipath signals, specifically using sparse WiFi measurements at multiple indoor locations to model the environment.

Result: Successfully inferred indoor floorplans from WiFi measurements, with promising results that enable forward applications like indoor signal prediction and basic ray tracing.

Conclusion: NeRFs can be effectively adapted to work with multipath signals, demonstrating the ability to “see” environments through RF measurements and opening new possibilities for indoor mapping and signal analysis.

Abstract: Neural Radiance Fields (NeRFs) have been remarkably successful at synthesizing novel views of 3D scenes by optimizing a volumetric scene function. This scene function models how optical rays bring color information from a 3D object to the camera pixels. Radio frequency (RF) or audio signals can also be viewed as a vehicle for delivering information about the environment to a sensor. However, unlike camera pixels, an RF/audio sensor receives a mixture of signals that contain many environmental reflections (also called “multipath”). Is it still possible to infer the environment using such multipath signals? We show that with redesign, NeRFs can be taught to learn from multipath signals, and thereby “see” the environment. As a grounding application, we aim to infer the indoor floorplan of a home from sparse WiFi measurements made at multiple locations inside the home. Although a difficult inverse problem, our implicitly learnt floorplans look promising, and enables forward applications, such as indoor signal prediction and basic ray tracing.

[527] So-Fake: Benchmarking and Explaining Social Media Image Forgery Detection

Zhenglin Huang, Tianxiao Li, Xiangtai Li, Haiquan Wen, Yiwei He, Jiangning Zhang, Hao Fei, Xi Yang, Xiaowei Huang, Bei Peng, Guangliang Cheng

Main category: cs.CV

TL;DR: So-Fake-Set: A comprehensive 2M+ image dataset and So-Fake-R1 detection framework for social media synthetic image detection, with 1.3% accuracy improvement and 4.5% IoU gain.

DetailsMotivation: Address limitations of existing synthetic image detection methods that lack diversity, scale, and realism for social media contexts, and struggle with generalization to unseen generative technologies.

Method: Created So-Fake-Set dataset with 2M+ images from 35 state-of-the-art generative models, established So-Fake-OOD benchmark (100K images) for cross-domain testing, and developed So-Fake-R1 vision-language framework using reinforcement learning for detection, localization, and explainable inference.

Result: So-Fake-R1 outperforms second-best method with 1.3% gain in detection accuracy and 4.5% increase in localization IoU, demonstrating superior performance on the challenging benchmark.

Conclusion: This work establishes a new foundation for social media-centric forgery detection research by integrating scalable dataset, challenging OOD benchmark, and advanced detection framework, with public release of code, models, and datasets.

Abstract: Recent advances in AI-powered generative models have enabled the creation of increasingly realistic synthetic images, posing significant risks to information integrity and public trust on social media platforms. While robust detection frameworks and diverse, large-scale datasets are essential to mitigate these risks, existing academic efforts remain limited in scope: current datasets lack the diversity, scale, and realism required for social media contexts, while detection methods struggle with generalization to unseen generative technologies. To bridge this gap, we introduce So-Fake-Set, a comprehensive social media-oriented dataset with over 2 million high-quality images, diverse generative sources, and photorealistic imagery synthesized using 35 state-of-the-art generative models. To rigorously evaluate cross-domain robustness, we establish a novel and large-scale (100K) out-of-domain benchmark (So-Fake-OOD) featuring synthetic imagery from commercial models explicitly excluded from the training distribution, creating a realistic testbed for evaluating real-world performance. Leveraging these resources, we present So-Fake-R1, an advanced vision-language framework that employs reinforcement learning for highly accurate forgery detection, precise localization, and explainable inference through interpretable visual rationales. Extensive experiments show that So-Fake-R1 outperforms the second-best method, with a 1.3% gain in detection accuracy and a 4.5% increase in localization IoU. By integrating a scalable dataset, a challenging OOD benchmark, and an advanced detection framework, this work establishes a new foundation for social media-centric forgery detection research. The code, models, and datasets will be released publicly.

[528] Diagnosing Reliability in Text-Guided Medical Image Editing

Minghao Liu, Zhitao He, Zhiyuan Fan, Qingyun Wang, Yi R. Fung

Main category: cs.CV

TL;DR: MedEBench is a comprehensive benchmark for evaluating text-guided medical image editing, featuring 1,182 clinically sourced image-prompt triplets across 70 tasks and 13 anatomical regions, with standardized evaluation metrics and failure analysis.

DetailsMotivation: Text-guided image editing has advanced in natural images but lacks standardized evaluation in medical imaging, despite its clinical potential for surgical simulation, teaching materials, and patient communication.

Method: Created MedEBench benchmark with clinically sourced data, three evaluation metrics (Editing Accuracy, Contextual Preservation, Visual Quality), ROI masks, and systematic comparison of 7 state-of-the-art models using attention grounding and IoU analysis.

Result: The benchmark reveals common failure patterns in existing models and provides a failure analysis protocol using attention map-ROI IoU to identify mislocalization issues.

Conclusion: MedEBench establishes a solid foundation for developing and evaluating reliable, clinically meaningful medical image editing systems with standardized evaluation framework.

Abstract: Text-guided image editing has seen rapid progress in natural image domains, but its adaptation to medical imaging remains limited and lacks standardized evaluation. Clinically, such editing holds promise for simulating surgical outcomes, creating personalized teaching materials, and enhancing patient communication. To bridge this gap, we introduce MedEBench, a comprehensive benchmark for evaluating text-guided medical image editing. It consists of 1,182 clinically sourced image-prompt triplets spanning 70 tasks across 13 anatomical regions. MedEBench offers three key contributions: (1) a clinically relevant evaluation framework covering Editing Accuracy, Contextual Preservation, and Visual Quality, supported by detailed descriptions of expected change and ROI (Region of Interest) masks; (2) a systematic comparison of seven state-of-the-art models, revealing common failure patterns; and (3) a failure analysis protocol based on attention grounding, using IoU between attention maps and ROIs to identify mislocalization. MedEBench provides a solid foundation for developing and evaluating reliable, clinically meaningful medical image editing systems. Project website: https://mliuby.github.io/MedEBench_Website/

[529] CarboFormer: A Lightweight Semantic Segmentation Architecture for Efficient Carbon Dioxide Detection Using Optical Gas Imaging

Taminul Islam, Toqi Tahamid Sarker, Mohamed G Embaby, Khaled R Ahmed, Amer AbuGhazaleh

Main category: cs.CV

TL;DR: CarboFormer is a lightweight semantic segmentation framework for Optical Gas Imaging that detects and quantifies CO2 emissions with high accuracy and computational efficiency, using optimized encoder-decoder architecture and multi-scale feature fusion.

DetailsMotivation: Carbon dioxide emissions are critical environmental indicators and important for industrial processes like livestock management. There's a need for efficient real-time monitoring tools that can operate on resource-constrained platforms.

Method: Integrates optimized encoder-decoder architecture with specialized multi-scale feature fusion and auxiliary supervision strategies. Introduces two novel datasets: Controlled Carbon Dioxide Release (CCR) dataset and Real Time Ankom (RTA) dataset for dairy cow rumen emissions.

Result: Achieves 84.88% mIoU on CCR dataset and 92.98% mIoU on RTA dataset with only 5.07M parameters and 84.68 FPS. Outperforms other lightweight methods like SegFormer-B0 and SegNeXt, especially in low-flow scenarios.

Conclusion: CarboFormer provides robust and efficient tools for CO2 emission analysis, advancing environmental sensing and precision livestock management, making it suitable for real-time monitoring on resource-constrained platforms like drones.

Abstract: Carbon dioxide (CO$_2$) emissions are critical indicators of both environmental impact and various industrial processes, including livestock management. We introduce CarboFormer, a lightweight semantic segmentation framework for Optical Gas Imaging (OGI), designed to detect and quantify CO$_2$ emissions across diverse applications. Our approach integrates an optimized encoder-decoder architecture with specialized multi-scale feature fusion and auxiliary supervision strategies to effectively model both local details and global relationships in gas plume imagery while achieving competitive accuracy with minimal computational overhead for resource-constrained environments. We contribute two novel datasets: (1) the Controlled Carbon Dioxide Release (CCR) dataset, which simulates gas leaks with systematically varied flow rates (10-100 SCCM), and (2) the Real Time Ankom (RTA) dataset, focusing on emissions from dairy cow rumen fluid in vitro experiments. Extensive evaluations demonstrate that CarboFormer achieves competitive performance with 84.88% mIoU on CCR and 92.98% mIoU on RTA, while maintaining computational efficiency with only 5.07M parameters and operating at 84.68 FPS. The model shows particular effectiveness in challenging low-flow scenarios and significantly outperforms other lightweight methods like SegFormer-B0 (83.36% mIoU on CCR) and SegNeXt (82.55% mIoU on CCR), making it suitable for real-time monitoring on resource-constrained platforms such as programmable drones. Our work advances both environmental sensing and precision livestock management by providing robust and efficient tools for CO$_2$ emission analysis.

[530] Domain Adaptation for Big Data in Agricultural Image Analysis: A Comprehensive Review

Xing Hu, Siyuan Chen, Qianqian Duan, Choon Ki Ahn, Huiliang Shang, Dawei Zhang

Main category: cs.CV

TL;DR: This paper reviews domain adaptation techniques for agricultural image analysis to address domain shift challenges caused by environmental changes, crop variations, and data acquisition differences.

DetailsMotivation: Computer vision applications in agriculture face significant domain shifts that hinder model generalization across regions, seasons, and scenarios. Domain adaptation offers solutions for limited labeled data, model adaptability issues, and dynamic field environment changes.

Method: Systematic review of domain adaptation methods categorized into shallow and deep learning approaches, including supervised, semi-supervised, and unsupervised strategies. Special focus on adversarial learning techniques for complex scenarios. Also reviews public agricultural image datasets.

Result: Domain adaptation methods have significantly improved cross-domain performance in agricultural applications including crop health monitoring, pest detection, and fruit identification. The study provides a complete framework for DA research.

Conclusion: This review provides key insights and a comprehensive framework that can serve as a reference for future research and development of domain adaptation methods in agricultural vision tasks.

Abstract: With the wide application of computer vision in agriculture, image analysis has become the key to tasks such as crop health monitoring and pest detection. However, the significant domain shifts caused by environmental changes, different crop types, and diverse data acquisition methods seriously hinder the generalization ability of the model in cross-region, cross-season, and complex agricultural scenarios. This paper explores how domain adaptation (DA) techniques can address these challenges to improve cross-domain transferability in agricultural image analysis. DA is considered a promising solution in the case of limited labeled data, insufficient model adaptability, and dynamic changes in the field environment. This paper systematically reviews the latest advances in DA in agricultural images in recent years, focusing on application scenarios such as crop health monitoring, pest and disease detection, and fruit identification, in which DA methods have significantly improved cross-domain performance. We categorize DA methods into shallow learning and deep learning methods, including supervised, semi-supervised and unsupervised strategies, and pay special attention to the adversarial learning-based techniques that perform well in complex scenarios. In addition, this paper also reviews the main public datasets of agricultural images, and evaluates their advantages and limitations in DA research. Overall, this study provides a complete framework and some key insights that can be used as a reference for the research and development of domain adaptation methods in future agricultural vision tasks.

[531] MCA-Bench: A Multimodal Benchmark for Evaluating CAPTCHA Robustness Against VLM-based Attacks

Zonglin Wu, Yule Xue, Yaoyao Feng, Xiaolong Wang, Yiren Song

Main category: cs.CV

TL;DR: MCA-Bench is a unified multimodal benchmarking suite for evaluating CAPTCHA security across diverse modalities using fine-tuned vision-language models, revealing vulnerabilities and providing design principles.

DetailsMotivation: The lack of a comprehensive, large-scale multimodal benchmark for evaluating CAPTCHA security robustness against advanced automated attacks.

Method: Developed MCA-Bench with heterogeneous CAPTCHA types integrated into a single evaluation protocol, using fine-tuned vision-language model agents for each CAPTCHA category.

Result: Effectively mapped vulnerability spectrum of modern CAPTCHAs, provided first quantitative analysis of challenge complexity, interaction depth, and model solvability relationships.

Conclusion: Proposed three actionable design principles and identified key open challenges for systematic CAPTCHA hardening and fair benchmarking.

Abstract: As automated attack techniques rapidly advance, CAPTCHAs remain a critical defense mechanism against malicious bots. However, existing CAPTCHA schemes encompass a diverse range of modalities – from static distorted text and obfuscated images to interactive clicks, sliding puzzles, and logic-based questions – yet the community still lacks a unified, large-scale, multimodal benchmark to rigorously evaluate their security robustness. To address this gap, we introduce MCA-Bench, a comprehensive and reproducible benchmarking suite that integrates heterogeneous CAPTCHA types into a single evaluation protocol. Leveraging a shared vision-language model backbone, we fine-tune specialized cracking agents for each CAPTCHA category, enabling consistent, cross-modal assessments. Extensive experiments reveal that MCA-Bench effectively maps the vulnerability spectrum of modern CAPTCHA designs under varied attack settings, and crucially offers the first quantitative analysis of how challenge complexity, interaction depth, and model solvability interrelate. Based on these findings, we propose three actionable design principles and identify key open challenges, laying the groundwork for systematic CAPTCHA hardening, fair benchmarking, and broader community collaboration. Datasets and code are available online.

[532] Compressed Feature Quality Assessment: Dataset and Baselines

Changsheng Gao, Wei Zhou, Guosheng Lin, Weisi Lin

Main category: cs.CV

TL;DR: First benchmark for Compressed Feature Quality Assessment (CFQA) with 300 original and 12,000 compressed features across vision tasks, showing traditional metrics like MSE and cosine similarity are inadequate for measuring semantic fidelity.

DetailsMotivation: Large models deployed in resource-constrained environments require efficient transmission of intermediate features, but compression causes semantic degradation that traditional metrics fail to quantify.

Method: Created benchmark dataset with features from three vision tasks compressed by four codecs, providing task-specific performance degradation as ground truth semantic distortion for evaluating metrics.

Result: Systematic evaluation shows MSE, cosine similarity, and CKA metrics are insufficient for capturing semantic degradation in compressed features, demonstrating the need for better CFQA metrics.

Conclusion: Establishes foundational CFQA benchmark and releases dataset/code to advance research on semantic fidelity assessment for compressed features in resource-constrained deployments.

Abstract: The widespread deployment of large models in resource-constrained environments has underscored the need for efficient transmission of intermediate feature representations. In this context, feature coding, which compresses features into compact bitstreams, becomes a critical component for scenarios involving feature transmission, storage, and reuse. However, this compression process inevitably introduces semantic degradation that is difficult to quantify with traditional metrics. To address this, we formalize the research problem of Compressed Feature Quality Assessment (CFQA), aiming to evaluate the semantic fidelity of compressed features. To advance CFQA research, we propose the first benchmark dataset, comprising 300 original features and 12000 compressed features derived from three vision tasks and four feature codecs. Task-specific performance degradation is provided as true semantic distortion for evaluating CFQA metrics. We systematically assess three widely used metrics – MSE, cosine similarity, and Centered Kernel Alignment (CKA) – in terms of their ability to capture semantic degradation. Our findings demonstrate the representativeness of the proposed dataset while underscoring the need for more sophisticated metrics capable of measuring semantic distortion in compressed features. This work advances the field by establishing a foundational benchmark and providing a critical resource for the community to explore CFQA. To foster further research, we release the dataset and all associated source code at https://github.com/chansongoal/Compressed-Feature-Quality-Assessment.

[533] A Vision-Language Agent System for Compositional Reasoning with VLM-assisted Script and Executable Generation

Yichang Xu, Gaowen Liu, Ramana Rao Kompella, Sihao Hu, Tiansheng Huang, Fatih Ilhan, Selim Furkan Tekin, Zachary Yahn, Ling Liu

Main category: cs.CV

TL;DR: VLAgent is a vision-language agent system that improves compositional reasoning by generating executable planning scripts from LLMs, correcting logic errors, and validating outputs through complementary reasoning techniques.

DetailsMotivation: Existing vision-language models perform poorly on compositional reasoning tasks, creating a need for systems that can better handle complex vision-text reasoning.

Method: Uses pre-trained LLM with few-shot learning to generate planning scripts, SS-parser to correct logic errors, and output verifier with ensemble learning and caption analysis to validate results.

Result: Outperforms state-of-the-art visual reasoning models on six visual benchmarks, demonstrating superior compositional text-visual reasoning capabilities.

Conclusion: VLAgent provides an effective framework for compositional vision-text reasoning through executable script generation, error correction, and output validation.

Abstract: The advancement in large language models (LLMs) and large vision models has fueled the rapid progress in multi-modal vision-text reasoning capabilities. However, existing vision-language models (VLMs) to date offer poor performance for compositional reasoning. This paper presents VLAgent, a vision-language agent system for vision-text compositional reasoning with three novel features. First, VLAgent leverages a pre-trained LLM with few-shot context learning to generate the planning script for each compositional reasoning task and provides a backend engine to generate and perform executable runtime, which maps the planning script into executable code using the VLAgent library for VLAgent executor. Second, VLAgent introduces the SS-parser, which identifies and corrects logic errors embedded in the LLM-generated planning script, to further enhance the quality of script-executable mapping. Third, VLAgent introduces the compositional reasoning output verifier, which validates and refines the output of complex compositional reasoning steps, by leveraging complementary reasoning techniques, e.g., ensemble learning and caption analysis. Extensive experiments are conducted on six visual benchmarks and compared to a dozen of the SoTA visual reasoning models. The results show that VLAgent outperforms existing representative approaches for compositional text-visual reasoning. Our code and datasets with outputs will be made available upon acceptance.

[534] DART: Differentiable Dynamic Adaptive Region Tokenizer for Vision Transformer and Mamba

Shicheng Yin, Kaixuan Yin, Yang Liu, Weixing Chen, Liang Lin

Main category: cs.CV

TL;DR: DART is a dynamic adaptive region tokenizer that creates content-dependent patches of varying sizes to focus tokens on information-rich areas, improving vision transformer performance while reducing computational costs.

DetailsMotivation: Fixed-size patches in vision transformers like ViT and Vim often waste tokens on background regions and miss critical local details, especially when objects are sparsely distributed.

Method: DART uses learnable region scores with piecewise differentiable quantile operations to adaptively partition images into varying-sized patches, allocating denser tokens to information-rich areas.

Result: DART improves DeiT accuracy by 2.1% on ImageNet-1K with only ~1M additional parameters, achieves 45% FLOPs reduction while maintaining superior performance, and consistently enhances accuracy across DeiT, Vim, and VideoMamba with minimal computational overhead.

Conclusion: DART provides an efficient alternative to uniform token density methods, demonstrating that adaptive token allocation significantly improves vision transformer performance while reducing computational costs.

Abstract: Recently, non-convolutional models such as the Vision Transformer (ViT) and Vision Mamba (Vim) have achieved remarkable performance in computer vision tasks. However, their reliance on fixed-size patches often results in excessive encoding of background regions and omission of critical local details, especially when informative objects are sparsely distributed. To address this, we introduce a fully differentiable Dynamic Adaptive Region Tokenizer (DART), which adaptively partitions images into content-dependent patches of varying sizes. DART combines learnable region scores with piecewise differentiable quantile operations to allocate denser tokens to information-rich areas. Despite introducing only approximately 1 million (1M) additional parameters, DART improves accuracy by 2.1% on DeiT (ImageNet-1K). Unlike methods that uniformly increase token density to capture fine-grained details, DART offers a more efficient alternative, achieving 45% FLOPs reduction with superior performance. Extensive experiments on DeiT, Vim, and VideoMamba confirm that DART consistently enhances accuracy while incurring minimal or even reduced computational overhead. Code is available at https://github.com/HCPLab-SYSU/DART.

[535] NOCTIS: Novel Object Cyclic Threshold based Instance Segmentation

Max Gandyra, Alessandro Santonicola, Michael Beetz

Main category: cs.CV

TL;DR: NOCTIS is a training-free framework for novel object instance segmentation that combines Grounded-SAM 2 for object proposals and DINOv2 for embeddings, using cyclic thresholding to achieve state-of-the-art results without retraining.

DetailsMotivation: Instance segmentation of novel objects without retraining is challenging. Existing methods struggle with generalization across diverse object types, requiring a solution that can handle various novel objects without additional training.

Method: Integrates pre-trained Grounded-SAM 2 for object proposals and DINOv2 for embeddings. Uses cyclic thresholding for stable matching, appearance scoring, and confidence-based scoring in an RGB-only pipeline.

Result: Achieves state-of-the-art mean AP scores on BOP 2023 challenge datasets, outperforming both RGB and RGB-D methods for unseen object segmentation.

Conclusion: NOCTIS demonstrates that effective novel object instance segmentation can be achieved without retraining by combining existing pre-trained models with intelligent matching mechanisms and scoring strategies.

Abstract: Instance segmentation of novel objects instances in RGB images, given some example images for each object, is a well known problem in computer vision. Designing a model general enough to be employed for all kinds of novel objects without (re-) training has proven to be a difficult task. To handle this, we present a new training-free framework, called: Novel Object Cyclic Threshold based Instance Segmentation (NOCTIS). NOCTIS integrates two pre-trained models: Grounded-SAM 2 for object proposals with precise bounding boxes and corresponding segmentation masks; and DINOv2 for robust class and patch embeddings, due to its zero-shot capabilities. Internally, the proposal-object matching is realized by determining an object matching score based on the similarity of the class embeddings and the average maximum similarity of the patch embeddings with a new cyclic thresholding (CT) mechanism that mitigates unstable matches caused by repetitive textures or visually similar patterns. Beyond CT, NOCTIS introduces: (i) an appearance score that is unaffected by object selection bias; (ii) the usage of the average confidence of the proposals bounding box and mask as a scoring component; and (iii) an RGB-only pipeline that performs even better than RGB-D ones. We empirically show that NOCTIS, without further training/fine tuning, attains state-of-the-art results regarding the mean AP score, w.r.t. the best RGB and RGB-D methods on the seven core datasets of the BOP 2023 challenge for the “Model-based 2D segmentation of unseen objects” task.

[536] STAGE: A Stream-Centric Generative World Model for Long-Horizon Driving-Scene Simulation

Jiamin Wang, Yichen Yao, Xiang Feng, Hang Wu, Yaming Wang, Qingqiu Huang, Yuexin Ma, Xinge Zhu

Main category: cs.CV

TL;DR: STAGE introduces hierarchical feature coordination and multi-phase optimization for generating high-quality, temporally consistent long-horizon driving videos, significantly outperforming existing methods.

DetailsMotivation: Existing approaches suffer from error accumulation and feature misalignment due to inadequate spatio-temporal decoupling and limited cross-frame feature propagation in autonomous driving video generation.

Method: Proposes Hierarchical Temporal Feature Transfer (HTFT) to model temporal and denoising processes separately with feature transfer between frames, plus a three-stage training strategy with model decoupling and auto-regressive inference simulation.

Result: Significantly surpasses existing methods on Nuscenes dataset and demonstrates ability to generate 600 frames of high-quality driving videos, far exceeding previous maximum lengths.

Conclusion: STAGE provides an effective solution for sustainable, high-fidelity long-horizon driving video generation with superior temporal consistency and reduced error accumulation.

Abstract: The generation of temporally consistent, high-fidelity driving videos over extended horizons presents a fundamental challenge in autonomous driving world modeling. Existing approaches often suffer from error accumulation and feature misalignment due to inadequate decoupling of spatio-temporal dynamics and limited cross-frame feature propagation mechanisms. To address these limitations, we present STAGE (Streaming Temporal Attention Generative Engine), a novel auto-regressive framework that pioneers hierarchical feature coordination and multi-phase optimization for sustainable video synthesis. To achieve high-quality long-horizon driving video generation, we introduce Hierarchical Temporal Feature Transfer (HTFT) and a novel multi-stage training strategy. HTFT enhances temporal consistency between video frames throughout the video generation process by modeling the temporal and denoising process separately and transferring denoising features between frames. The multi-stage training strategy is to divide the training into three stages, through model decoupling and auto-regressive inference process simulation, thereby accelerating model convergence and reducing error accumulation. Experiments on the Nuscenes dataset show that STAGE has significantly surpassed existing methods in the long-horizon driving video generation task. In addition, we also explored STAGE’s ability to generate unlimited-length driving videos. We generated 600 frames of high-quality driving videos on the Nuscenes dataset, which far exceeds the maximum length achievable by existing methods.

[537] Driving as a Diagnostic Tool: Scenario-based Cognitive Assessment in Older Drivers from Driving Video

Md Zahid Hasan, Guillermo Basulto-Elias, Jun Ha Chang, Sahuna Hallmark, Matthew Rizzo, Anuj Sharma, Soumik Sarkar

Main category: cs.CV

TL;DR: Using large vision models to analyze naturalistic driving videos for early detection of cognitive decline in older drivers by identifying behavioral patterns that correlate with dementia and MCI.

DetailsMotivation: Current cognitive decline diagnostic methods are time-consuming and costly, leading to underdiagnosis of Dementia and Mild Cognitive Impairment. There's a need for scalable, non-invasive monitoring systems to address the growing burden of cognitive decline in aging populations.

Method: A framework that leverages large vision models to analyze naturalistic driving videos captured through in-vehicle sensors. The method extracts “digital fingerprints” from real-world driving behavior across different roadway scenarios to identify cognitive status and predict disease progression.

Result: The approach can identify early warning signs of functional impairment by correlating driving patterns with clinical features of dementia, enabling the vehicle to serve as a “diagnostic tool” for cognitive status assessment.

Conclusion: This work enhances early detection capabilities and supports the development of proactive intervention strategies, contributing to scalable monitoring systems that can mitigate the societal and economic burden of cognitive decline in older adults.

Abstract: We introduce scenario-based cognitive status identification in older drivers from naturalistic driving videos, leveraging large vision models. In recent times, cognitive decline including Dementia and Mild Cognitive Impairment (MCI), is often underdiagnosed due to the time-consuming and costly nature of current diagnostic methods. By analyzing real-world driving behavior captured through in-vehicle sensors, this study aims to extract “digital fingerprints” that correlate with functional decline and clinical features of dementia. Moreover, modern large vision models can draw meaningful insights from everyday driving patterns across different roadway scenarios to early detect cognitive decline. We propose a framework that uses large vision models and naturalistic driving videos to analyze driver behavior, identify cognitive status and predict disease progression. We leverage the strong relationship between real-world driving behavior as an observation of the current cognitive status of the drivers where the vehicle can be utilized as a “diagnostic tool”. Our method identifies early warning signs of functional impairment, contributing to proactive intervention strategies. This work enhances early detection and supports the development of scalable, non-invasive monitoring systems to mitigate the growing societal and economic burden of cognitive decline in the aging population.

[538] DExNet: Combining Observations of Domain Adapted Critics for Leaf Disease Classification with Limited Data

Sabbir Ahmed, Md. Bakhtiar Hasan, Tasnim Ahmed, Md. Hasanul Kabir

Main category: cs.CV

TL;DR: DExNet is a few-shot learning framework that combines multiple pre-trained CNN critics with domain adaptation and Bi-LSTM fusion for plant disease classification with limited data, achieving near state-of-the-art performance with 94.5% less training data.

DetailsMotivation: Deep learning models require large datasets for plant disease classification, but obtaining sufficient training data for leaf diseases is challenging. This work addresses the need for effective classification with limited samples.

Method: Uses 9 pre-trained CNN architectures as ‘critics’ that are domain-adapted on a non-overlapping leaf disease dataset. Extracts feature embeddings, fuses them through a Feature Fusion Block, and classifies using Bi-LSTM layers in a few-shot learning framework.

Result: Achieved 89.06% (5-shot), 92.46% (10-shot), 94.07% (15-shot), and 98.09% (80-shot) accuracy on tomato leaf disease classification, requiring only 5.5% of the data compared to state-of-the-art while maintaining similar performance.

Conclusion: DExNet provides an effective solution for plant disease classification with limited data, outperforming existing methods across single-domain, mixed-domain, and cross-domain scenarios while significantly reducing data requirements.

Abstract: While deep learning-based architectures have been widely used for correctly detecting and classifying plant diseases, they require large-scale datasets to learn generalized features and achieve state-of-the-art performance. This poses a challenge for such models to obtain satisfactory performance in classifying leaf diseases with limited samples. This work proposes a few-shot learning framework, Domain-adapted Expert Network (DExNet), for plant disease classification that compensates for the lack of sufficient training data by combining observations of a number of expert critics. It starts with extracting the feature embeddings as ‘observations’ from nine ‘critics’ that are state-of-the-art pre-trained CNN-based architectures. These critics are ‘domain adapted’ using a publicly available leaf disease dataset having no overlapping classes with the specific downstream task of interest. The observations are then passed to the ‘Feature Fusion Block’ and finally to a classifier network consisting of Bi-LSTM layers. The proposed pipeline is evaluated on the 10 classes of tomato leaf images from the PlantVillage dataset, achieving promising accuracies of 89.06%, 92.46%, and 94.07%, respectively, for 5-shot, 10-shot, and 15-shot classification. Furthermore, an accuracy of 98.09+-0.7% has been achieved in 80-shot classification, which is only 1.2% less than state-of-the-art, allowing a 94.5% reduction in the training data requirement. The proposed pipeline also outperforms existing works on leaf disease classification with limited data in both laboratory and real-life conditions in single-domain, mixed-domain, and cross-domain scenarios.

[539] Demographic-aware fine-grained classification of pediatric wrist fractures

Ammar Ahmed, Ali Shariq Imran, Zenun Kastrati, Sher Muhammad Daudpota

Main category: cs.CV

TL;DR: This paper presents a novel approach for wrist pathology recognition using fine-grained transformers, metadata fusion, and specialized pre-training, achieving significant accuracy improvements over traditional methods.

DetailsMotivation: Wrist pathologies are common, especially in children, but medical imaging datasets are limited. Relying solely on image data is insufficient, and there's a need to leverage multiple data types for better diagnosis.

Method: Framed as fine-grained recognition task, fused patient metadata with X-rays, used fine-grained pre-training weights instead of ImageNet, and applied transformer architecture for the first time with metadata integration in wrist pathology.

Result: Combination of fine-grained transformer approach, fine-grained pre-training, and metadata integration improved diagnostic accuracy by 2% on small custom dataset and over 10% on larger fracture dataset.

Conclusion: The multifaceted approach integrating metadata with medical images and using specialized fine-grained pre-training significantly enhances wrist pathology recognition accuracy, demonstrating the value of combining multiple data modalities in medical imaging analysis.

Abstract: Wrist pathologies are frequently observed, particularly among children who constitute the majority of fracture cases. Computer vision presents a promising avenue, contingent upon the availability of extensive datasets, a notable challenge in medical imaging. Therefore, reliance solely on one modality, such as images, proves inadequate, especially in an era of diverse and plentiful data types. This study addresses the problem using a multifaceted approach: framing it as a fine-grained recognition task, fusing patient metadata with X-rays, and leveraging weights from a separate fine-grained dataset rather than from a coarse-grained dataset like ImageNet. Unlike prior work, this is the first application of metadata integration for wrist pathology recognition. Our results show that combining fine-grained transformer approach, fine-grained pre-training, and metadata integration improves diagnostic accuracy by 2% on small custom curated dataset and over 10% on a larger fracture dataset.

[540] Reconstructing Tornadoes in 3D with Gaussian Splatting

Adam Yang, Nadula Kadawedduwa, Tianfu Wang, Sunny Sharma, Emily F. Wisinski, Jhayron S. Pérez-Carrasquilla, Kyle J. C. Hall, Dean Calhoun, Jonathan Starfeldt, Timothy P. Canty, Maria Molina, Christopher Metzler

Main category: cs.CV

TL;DR: Researchers created a lab-based tornado dataset and used 3D Gaussian splatting to successfully reconstruct its 3D structure.

DetailsMotivation: Accurate 3D reconstruction of tornadoes is crucial for understanding and preparing for these destructive weather phenomena, but there's a lack of controlled tornado datasets to develop and validate 3D reconstruction tools.

Method: Captured and released a novel multiview dataset of a small lab-based tornado, then used 3D Gaussian splatting (3DGS) technique for reconstruction.

Result: Successfully demonstrated effective 3D reconstruction and visualization of the tornado structure using 3DGS.

Conclusion: The created dataset enables development of 3D reconstruction tools for tornadoes, and 3DGS proves effective for visualizing tornado structures from controlled lab data.

Abstract: Accurately reconstructing the 3D structure of tornadoes is critically important for understanding and preparing for this highly destructive weather phenomenon. While modern 3D scene reconstruction techniques, such as 3D Gaussian splatting (3DGS), could provide a valuable tool for reconstructing the 3D structure of tornados, at present we are critically lacking a controlled tornado dataset with which to develop and validate these tools. In this work we capture and release a novel multiview dataset of a small lab-based tornado. We demonstrate one can effectively reconstruct and visualize the 3D structure of this tornado using 3DGS.

[541] GPI-Net: Gestalt-Guided Parallel Interaction Network via Orthogonal Geometric Consistency for Robust Point Cloud Registration

Weikang Gu, Mingyue Han, Li Xue, Heng Dong, Changcai Yang, Riqing Chen, Lifang Wei

Main category: cs.CV

TL;DR: Proposes GPI-Net, a Gestalt-guided parallel interaction network for point cloud registration that uses orthogonal geometric consistency to improve correspondence quality by integrating local and global features.

DetailsMotivation: Accurate identification of high-quality correspondences is crucial for point cloud registration, but challenging due to feature redundancy and complex spatial relationships between local and global features.

Method: Uses Gestalt principles with orthogonal integration strategy to reduce redundancy. Implements Gestalt Feature Attention block with self/cross-attention mechanisms and Dual-path Multi-Granularity parallel interaction block for information exchange across granularities.

Result: Extensive experiments demonstrate superior performance compared to existing methods on various challenging tasks.

Conclusion: GPI-Net effectively handles the fusion of local and global features for high-quality correspondences in point cloud registration through Gestalt-guided parallel interaction.

Abstract: The accurate identification of high-quality correspondences is a prerequisite task in feature-based point cloud registration. However, it is extremely challenging to handle the fusion of local and global features due to feature redundancy and complex spatial relationships. Given that Gestalt principles provide key advantages in analyzing local and global relationships, we propose a novel Gestalt-guided Parallel Interaction Network via orthogonal geometric consistency (GPI-Net) in this paper. It utilizes Gestalt principles to facilitate complementary communication between local and global information. Specifically, we introduce an orthogonal integration strategy to optimally reduce redundant information and generate a more compact global structure for high-quality correspondences. To capture geometric features in correspondences, we leverage a Gestalt Feature Attention (GFA) block through a hybrid utilization of self-attention and cross-attention mechanisms. Furthermore, to facilitate the integration of local detail information into the global structure, we design an innovative Dual-path Multi-Granularity parallel interaction aggregation (DMG) block to promote information exchange across different granularities. Extensive experiments on various challenging tasks demonstrate the superior performance of our proposed GPI-Net in comparison to existing methods. The code will be released at https://github.com/gwk429/GPI-Net.

[542] VSRM: A Robust Mamba-Based Framework for Video Super-Resolution

Dinh Phu Tran, Dao Duy Hung, Daeyoung Kim

Main category: cs.CV

TL;DR: VSRM is a novel video super-resolution framework that uses Mamba architecture for efficient long-range spatio-temporal feature extraction with linear complexity, achieving state-of-the-art results.

DetailsMotivation: Current CNN-based methods have limited receptive fields while Transformers suffer from quadratic complexity, making them inefficient for long video sequences in super-resolution tasks.

Method: Proposes VSRM framework with Spatial-to-Temporal and Temporal-to-Spatial Mamba blocks for feature extraction, Deformable Cross-Mamba Alignment for dynamic frame alignment, and Frequency Charbonnier-like loss for frequency domain optimization.

Result: Achieves state-of-the-art performance on diverse benchmarks, demonstrating superior video super-resolution quality with efficient computation.

Conclusion: VSRM establishes a solid foundation for future video super-resolution research by effectively addressing limitations of existing methods through Mamba architecture and novel alignment techniques.

Abstract: Video super-resolution remains a major challenge in low-level vision tasks. To date, CNN- and Transformer-based methods have delivered impressive results. However, CNNs are limited by local receptive fields, while Transformers struggle with quadratic complexity, posing challenges for processing long sequences in VSR. Recently, Mamba has drawn attention for its long-sequence modeling, linear complexity, and large receptive fields. In this work, we propose VSRM, a novel \textbf{V}ideo \textbf{S}uper-\textbf{R}esolution framework that leverages the power of \textbf{M}amba. VSRM introduces Spatial-to-Temporal Mamba and Temporal-to-Spatial Mamba blocks to extract long-range spatio-temporal features and enhance receptive fields efficiently. To better align adjacent frames, we propose Deformable Cross-Mamba Alignment module. This module utilizes a deformable cross-mamba mechanism to make the compensation stage more dynamic and flexible, preventing feature distortions. Finally, we minimize the frequency domain gaps between reconstructed and ground-truth frames by proposing a simple yet effective Frequency Charbonnier-like loss that better preserves high-frequency content and enhances visual quality. Through extensive experiments, VSRM achieves state-of-the-art results on diverse benchmarks, establishing itself as a solid foundation for future research.

[543] LOD-GS: Level-of-Detail-Sensitive 3D Gaussian Splatting for Detail Conserved Anti-Aliasing

Zhenya Yang, Bingchen Gong, Kai Chen

Main category: cs.CV

TL;DR: LOD-GS is a Level-of-Detail-sensitive filtering framework for 3D Gaussian Splatting that dynamically predicts optimal filtering strength for each Gaussian primitive based on sampling rate, eliminating aliasing artifacts while maintaining rendering quality.

DetailsMotivation: Existing anti-aliasing methods in 3D Gaussian Splatting rely on low-pass filtering but are insensitive to sampling rate, leading to under-filtering and over-smoothing. Current evaluation methods also overlook camera distance impact.

Method: Introduces basis functions to each Gaussian that take sampling rate as input to model appearance variations. Parameters are jointly optimized with 3D Gaussians end-to-end. Also creates a new synthetic dataset with varying camera distances for comprehensive evaluation.

Result: Achieves state-of-the-art rendering quality while effectively eliminating aliasing artifacts. Extensive experiments on public datasets and new synthetic dataset demonstrate superior performance.

Conclusion: LOD-GS provides a sampling-rate-sensitive filtering framework that dynamically adapts to different viewing conditions, solving aliasing issues in 3D Gaussian Splatting while maintaining high rendering efficiency and quality.

Abstract: Despite the advancements in quality and efficiency achieved by 3D Gaussian Splatting (3DGS) in 3D scene rendering, aliasing artifacts remain a persistent challenge. Existing approaches primarily rely on low-pass filtering to mitigate aliasing. However, these methods are not sensitive to the sampling rate, often resulting in under-filtering and over-smoothing renderings. To address this limitation, we propose LOD-GS, a Level-of-Detail-sensitive filtering framework for Gaussian Splatting, which dynamically predicts the optimal filtering strength for each 3D Gaussian primitive. Specifically, we introduce a set of basis functions to each Gaussian, which take the sampling rate as input to model appearance variations, enabling sampling-rate-sensitive filtering. These basis function parameters are jointly optimized with the 3D Gaussian in an end-to-end manner. The sampling rate is influenced by both focal length and camera distance. However, existing methods and datasets rely solely on down-sampling to simulate focal length changes for anti-aliasing evaluation, overlooking the impact of camera distance. To enable a more comprehensive assessment, we introduce a new synthetic dataset featuring objects rendered at varying camera distances. Extensive experiments on both public datasets and our newly collected dataset demonstrate that our method achieves SOTA rendering quality while effectively eliminating aliasing. The code and dataset have been open-sourced.

[544] IC-Custom: Diverse Image Customization via In-Context Learning

Yaowei Li, Xiaoyu Li, Zhaoyang Zhang, Yuxuan Bian, Gan Liu, Xinyuan Li, Jiale Xu, Wenbo Hu, Yating Liu, Lingen Li, Jing Cai, Yuexian Zou, Yancheng He, Ying Shan

Main category: cs.CV

TL;DR: IC-Custom is a unified framework that integrates position-aware and position-free image customization through in-context learning, achieving 73% higher human preference while training only 0.4% of parameters.

DetailsMotivation: Current approaches separate image customization into position-aware and position-free paradigms, lacking a universal framework for diverse customization scenarios in industrial media production.

Method: Proposes IC-Custom with In-context Multi-Modal Attention (ICMA) mechanism using learnable task-oriented register tokens and boundary-aware positional embeddings. Uses polyptych configuration of reference and target images with DiT’s attention for token-level interactions. Curates 12k high-quality dataset with real-world and synthetic data.

Result: Significantly outperforms community workflows, closed-source models, and state-of-the-art open-source approaches on ProductBench and DreamBench. Achieves ~73% higher human preference across identity consistency, harmonicity, and text alignment metrics.

Conclusion: IC-Custom provides a unified framework for diverse industrial applications including try-on, accessory placement, furniture arrangement, and creative IP customization, demonstrating superior performance with minimal parameter training.

Abstract: Image customization, a crucial technique for industrial media production, aims to generate content that is consistent with reference images. However, current approaches conventionally separate image customization into position-aware and position-free customization paradigms and lack a universal framework for diverse customization, limiting their applications across various scenarios. To overcome these limitations, we propose IC-Custom, a unified framework that seamlessly integrates position-aware and position-free image customization through in-context learning. IC-Custom concatenates reference images with target images to a polyptych, leveraging DiT’s multi-modal attention mechanism for fine-grained token-level interactions. We introduce the In-context Multi-Modal Attention (ICMA) mechanism with learnable task-oriented register tokens and boundary-aware positional embeddings to enable the model to correctly handle different task types and distinguish various inputs in polyptych configurations. To bridge the data gap, we carefully curated a high-quality dataset of 12k identity-consistent samples with 8k from real-world sources and 4k from high-quality synthetic data, avoiding the overly glossy and over-saturated synthetic appearance. IC-Custom supports various industrial applications, including try-on, accessory placement, furniture arrangement, and creative IP customization. Extensive evaluations on our proposed ProductBench and the publicly available DreamBench demonstrate that IC-Custom significantly outperforms community workflows, closed-source models, and state-of-the-art open-source approaches. IC-Custom achieves approximately 73% higher human preference across identity consistency, harmonicity, and text alignment metrics, while training only 0.4% of the original model parameters. Project page: https://liyaowei-stu.github.io/project/IC_Custom

[545] Discontinuity-aware Normal Integration for Generic Central Camera Models

Francesco Milano, Manuel López-Antequera, Naina Dhingra, Roland Siegwart, Robert Thiel

Main category: cs.CV

TL;DR: Novel normal integration method that explicitly handles depth discontinuities and works with generic central cameras using local planarity constraints between surface normals and ray directions.

DetailsMotivation: Existing normal integration approaches fail to properly handle depth discontinuities and are limited to orthographic or ideal pinhole cameras, limiting their practical application.

Method: Proposes a formulation based on local planarity assumption, modeling constraints between surface normals and ray directions to accurately approximate depth-normal relations.

Result: Achieves state-of-the-art results on standard normal integration benchmark and is the first method to directly handle generic central camera models.

Conclusion: The approach provides more accurate normal integration by explicitly modeling discontinuities and supporting various camera models, advancing photometric shape reconstruction techniques.

Abstract: Recovering a 3D surface from its surface normal map, a problem known as normal integration, is a key component for photometric shape reconstruction techniques such as shape-from-shading and photometric stereo. The vast majority of existing approaches for normal integration handle only implicitly the presence of depth discontinuities and are limited to orthographic or ideal pinhole cameras. In this paper, we propose a novel formulation that allows modeling discontinuities explicitly and handling generic central cameras. Our key idea is based on a local planarity assumption, that we model through constraints between surface normals and ray directions. Compared to existing methods, our approach more accurately approximates the relation between depth and surface normals, achieves state-of-the-art results on the standard normal integration benchmark, and is the first to directly handle generic central camera models.

[546] Exploring the Application of Visual Question Answering (VQA) for Classroom Activity Monitoring

Sinh Trong Vu, Hieu Trung Pham, Dung Manh Nguyen, Hieu Minh Hoang, Nhu Hoang Le, Thu Ha Pham, Tai Tan Mai

Main category: cs.CV

TL;DR: Evaluation of four state-of-the-art VQA models (LLaMA2, LLaMA3, QWEN3, NVILA) for classroom behavior analysis using a new dataset from Vietnamese classroom recordings, showing promising performance for automated classroom monitoring.

DetailsMotivation: Classroom behavior monitoring is crucial for educational research and student outcomes, and recent VQA advancements offer automated tools for analyzing complex classroom interactions from video.

Method: Introduces BAV-Classroom-VQA dataset from real Vietnamese classroom recordings, with methodology for data collection and annotation. Benchmarks four open-source VQA models on this dataset.

Result: Initial experimental results show all four models achieve promising performance levels in answering behavior-related visual questions from classroom videos.

Conclusion: The VQA models demonstrate potential for future classroom analytics and intervention systems, enabling automated behavior monitoring and analysis.

Abstract: Classroom behavior monitoring is a critical aspect of educational research, with significant implications for student engagement and learning outcomes. Recent advancements in Visual Question Answering (VQA) models offer promising tools for automatically analyzing complex classroom interactions from video recordings. In this paper, we investigate the applicability of several state-of-the-art open-source VQA models, including LLaMA2, LLaMA3, QWEN3, and NVILA, in the context of classroom behavior analysis. To facilitate rigorous evaluation, we introduce our BAV-Classroom-VQA dataset derived from real-world classroom video recordings at the Banking Academy of Vietnam. We present the methodology for data collection, annotation, and benchmark the performance of the selected VQA models on this dataset. Our initial experimental results demonstrate that all four models achieve promising performance levels in answering behavior-related visual questions, showcasing their potential in future classroom analytics and intervention systems.

[547] HiM2SAM: Enhancing SAM2 with Hierarchical Motion Estimation and Memory Optimization towards Long-term Tracking

Ruixiang Chen, Guolei Sun, Yawei Li, Jie Qin, Luca Benini

Main category: cs.CV

TL;DR: Enhanced SAM2 framework for video object tracking with hierarchical motion estimation and optimized memory bank, achieving state-of-the-art performance without additional training.

DetailsMotivation: Address challenges in video object tracking including occlusions, background clutter, and target reappearance to improve long-term tracking performance.

Method: Hierarchical motion estimation combining lightweight linear prediction with selective non-linear refinement, plus optimized memory bank with long-term and short-term memory frame distinction.

Result: Achieved 9.6% and 7.2% relative improvements in AUC on LaSOT and LaSOText with large model, with even larger gains on smaller models.

Conclusion: Trainless, low-overhead improvements effectively boost long-term tracking performance across different model scales.

Abstract: This paper presents enhancements to the SAM2 framework for video object tracking task, addressing challenges such as occlusions, background clutter, and target reappearance. We introduce a hierarchical motion estimation strategy, combining lightweight linear prediction with selective non-linear refinement to improve tracking accuracy without requiring additional training. In addition, we optimize the memory bank by distinguishing long-term and short-term memory frames, enabling more reliable tracking under long-term occlusions and appearance changes. Experimental results show consistent improvements across different model scales. Our method achieves state-of-the-art performance on LaSOT and LaSOText with the large model, achieving 9.6% and 7.2% relative improvements in AUC over the original SAM2, and demonstrates even larger relative gains on smaller models, highlighting the effectiveness of our trainless, low-overhead improvements for boosting long-term tracking performance. The code is available at https://github.com/LouisFinner/HiM2SAM.

[548] Simplifying Traffic Anomaly Detection with Video Foundation Models

Svetlana Orlova, Tommie Kerssies, Brunó B. Englert, Gijs Dubbelman

Main category: cs.CV

TL;DR: Simple Video Vision Transformers with advanced pre-training (especially self-supervised Masked Video Modeling) match or exceed complex specialized methods for Traffic Anomaly Detection while being more efficient.

DetailsMotivation: To investigate whether complex multi-stage architectures are necessary for traffic anomaly detection, and explore if simple encoder-only models with proper pre-training can achieve comparable or better performance.

Method: Uses plain Video Vision Transformers (Video ViTs) with different pre-training strategies including self-supervised Masked Video Modeling (MVM), weakly-supervised, fully-supervised pre-training, and Domain-Adaptive Pre-Training (DAPT) on unlabeled driving videos.

Result: Simple encoder-only models match or surpass specialized state-of-the-art TAD methods while being significantly more efficient. Self-supervised MVM provides the strongest signal, and DAPT further improves performance without requiring anomalous examples.

Conclusion: Effective, efficient, and scalable traffic anomaly detection models can be built with minimal architectural complexity through proper pre-training strategies, particularly self-supervised approaches and domain adaptation.

Abstract: Recent methods for ego-centric Traffic Anomaly Detection (TAD) often rely on complex multi-stage or multi-representation fusion architectures, yet it remains unclear whether such complexity is necessary. Recent findings in visual perception suggest that foundation models, enabled by advanced pre-training, allow simple yet flexible architectures to outperform specialized designs. Therefore, in this work, we investigate an architecturally simple encoder-only approach using plain Video Vision Transformers (Video ViTs) and study how pre-training enables strong TAD performance. We find that: (i) advanced pre-training enables simple encoder-only models to match or even surpass the performance of specialized state-of-the-art TAD methods, while also being significantly more efficient; (ii) although weakly- and fully-supervised pre-training are advantageous on standard benchmarks, we find them less effective for TAD. Instead, self-supervised Masked Video Modeling (MVM) provides the strongest signal; and (iii) Domain-Adaptive Pre-Training (DAPT) on unlabeled driving videos further improves downstream performance, without requiring anomalous examples. Our findings highlight the importance of pre-training and show that effective, efficient, and scalable TAD models can be built with minimal architectural complexity. We release our code, domain-adapted encoders, and fine-tuned models to support future work: https://github.com/tue-mps/simple-tad.

[549] Memory-Efficient Personalization of Text-to-Image Diffusion Models via Selective Optimization Strategies

Seokeon Choi, Sunghyun Park, Hyoungwoo Park, Jeongho Kim, Sungrack Yun

Main category: cs.CV

TL;DR: Selective optimization framework combining BP-low (backpropagation on low-res images) and ZO-high (zeroth-order optimization on high-res images) for memory-efficient diffusion model personalization on edge devices.

DetailsMotivation: Need for memory-efficient personalization of text-to-image diffusion models that preserves user privacy and works within limited computational resources of edge devices.

Method: Timestep-aware probabilistic function that dynamically selects between BP-low for target-specific adaptation and ZO-high for high-resolution detail refinement based on diffusion timesteps.

Result: Achieves competitive performance while significantly reducing memory consumption, enabling scalable high-quality on-device personalization without increasing inference latency.

Conclusion: The framework successfully combines both optimization strategies to leverage their complementary strengths - BP-low for effective personalization and ZO-high for structural consistency, achieving memory-efficient fine-tuning.

Abstract: Memory-efficient personalization is critical for adapting text-to-image diffusion models while preserving user privacy and operating within the limited computational resources of edge devices. To this end, we propose a selective optimization framework that adaptively chooses between backpropagation on low-resolution images (BP-low) and zeroth-order optimization on high-resolution images (ZO-high), guided by the characteristics of the diffusion process. As observed in our experiments, BP-low efficiently adapts the model to target-specific features, but suffers from structural distortions due to resolution mismatch. Conversely, ZO-high refines high-resolution details with minimal memory overhead but faces slow convergence when applied without prior adaptation. By complementing both methods, our framework leverages BP-low for effective personalization while using ZO-high to maintain structural consistency, achieving memory-efficient and high-quality fine-tuning. To maximize the efficacy of both BP-low and ZO-high, we introduce a timestep-aware probabilistic function that dynamically selects the appropriate optimization strategy based on diffusion timesteps. This function mitigates the overfitting from BP-low at high timesteps, where structural information is critical, while ensuring ZO-high is applied more effectively as training progresses. Experimental results demonstrate that our method achieves competitive performance while significantly reducing memory consumption, enabling scalable, high-quality on-device personalization without increasing inference latency.

[550] CF3: Compact and Fast 3D Feature Fields

Hyunjoon Lee, Joonkyu Min, Jaesik Park

Main category: cs.CV

TL;DR: CF3 proposes a top-down pipeline for efficient 3D Gaussian feature fields using weighted multi-view feature fusion, per-Gaussian autoencoders, and adaptive sparsification to achieve compact representation with only 5% of Gaussians.

DetailsMotivation: Current 3DGS approaches rely on bottom-up optimization treating 2D features as ground truth, leading to high computational costs and inefficiency.

Method: Fast weighted fusion of multi-view 2D features with pre-trained Gaussians, training per-Gaussian autoencoders directly on lifted features, and adaptive sparsification with pruning/merging of redundant Gaussians.

Result: Achieves competitive 3D feature field using only 5% of Gaussians compared to Feature-3DGS while preserving geometric details.

Conclusion: CF3 provides an efficient top-down approach for constructing compact 3D Gaussian feature fields with significantly reduced computational requirements.

Abstract: 3D Gaussian Splatting (3DGS) has begun incorporating rich information from 2D foundation models. However, most approaches rely on a bottom-up optimization process that treats raw 2D features as ground truth, incurring increased computational costs. We propose a top-down pipeline for constructing compact and fast 3D Gaussian feature fields, namely, CF3. We first perform a fast weighted fusion of multi-view 2D features with pre-trained Gaussians. This approach enables training a per-Gaussian autoencoder directly on the lifted features, instead of training autoencoders in the 2D domain. As a result, the autoencoder better aligns with the feature distribution. More importantly, we introduce an adaptive sparsification method that optimizes the Gaussian attributes of the feature field while pruning and merging the redundant Gaussians, constructing an efficient representation with preserved geometric details. Our approach achieves a competitive 3D feature field using as little as 5% of the Gaussians compared to Feature-3DGS.

[551] Detecting Visual Information Manipulation Attacks in Augmented Reality: A Multimodal Semantic Reasoning Approach

Yanming Xiu, Maria Gorlatova

Main category: cs.CV

TL;DR: This paper addresses visual information manipulation attacks in AR, categorizes them into a taxonomy, creates a dataset, and proposes a multimodal detection framework that achieves 88.94% accuracy.

DetailsMotivation: AR virtual content can introduce misleading or harmful information that causes semantic misunderstandings and user errors, necessitating detection methods for visual information manipulation attacks.

Method: Proposed a multimodal semantic reasoning framework (VIM-Sense) combining vision-language models with OCR-based textual analysis to detect attacks categorized into character, phrase, and pattern manipulation formats.

Result: Achieved 88.94% attack detection accuracy on the AR-VIM dataset, outperforming vision-only and text-only baselines, with average detection latency of 7.07-7.17 seconds.

Conclusion: The proposed VIM-Sense framework effectively detects visual information manipulation attacks in AR through multimodal semantic reasoning, demonstrating practical applicability in real-world AR scenarios.

Abstract: The virtual content in augmented reality (AR) can introduce misleading or harmful information, leading to semantic misunderstandings or user errors. In this work, we focus on visual information manipulation (VIM) attacks in AR, where virtual content changes the meaning of real-world scenes in subtle but impactful ways. We introduce a taxonomy that categorizes these attacks into three formats: character, phrase, and pattern manipulation, and three purposes: information replacement, information obfuscation, and extra wrong information. Based on the taxonomy, we construct a dataset, AR-VIM, which consists of 452 raw-AR video pairs spanning 202 different scenes, each simulating a real-world AR scenario. To detect the attacks in the dataset, we propose a multimodal semantic reasoning framework, VIM-Sense. It combines the language and visual understanding capabilities of vision-language models (VLMs) with optical character recognition (OCR)-based textual analysis. VIM-Sense achieves an attack detection accuracy of 88.94% on AR-VIM, consistently outperforming vision-only and text-only baselines. The system achieves an average attack detection latency of 7.07 seconds in a simulated video processing framework and 7.17 seconds in a real-world evaluation conducted on a mobile Android AR application.

[552] Exploiting Diffusion Prior for Task-driven Image Restoration

Jaeha Kim, Junghun Oh, Kyoung Mu Lee

Main category: cs.CV

TL;DR: EDTR leverages diffusion prior to restore task-relevant details in degraded images, improving both visual quality and task performance through pixel-error-based pre-restoration and controlled denoising steps.

DetailsMotivation: Existing task-driven image restoration methods struggle with multiple complex degradations that leave minimal restoration clues, and diffusion priors alone fail to preserve task-relevant details.

Method: Proposes EDTR that uses pixel-error-based pre-restored images with mild noise as starting points for diffusion, and employs limited denoising steps to avoid generating redundant details that dilute task-critical information.

Result: The method effectively utilizes diffusion prior for TDIR, significantly enhancing both task performance and visual quality across diverse tasks with multiple complex degradations.

Conclusion: EDTR successfully bridges the gap between diffusion-based generation and task-driven restoration by strategically controlling the diffusion process to preserve and enhance task-relevant details in severely degraded images.

Abstract: Task-driven image restoration (TDIR) has recently emerged to address performance drops in high-level vision tasks caused by low-quality (LQ) inputs. Previous TDIR methods struggle to handle practical scenarios in which images are degraded by multiple complex factors, leaving minimal clues for restoration. This motivates us to leverage the diffusion prior, one of the most powerful natural image priors. However, while the diffusion prior can help generate visually plausible results, using it to restore task-relevant details remains challenging, even when combined with recent TDIR methods. To address this, we propose EDTR, which effectively harnesses the power of diffusion prior to restore task-relevant details. Specifically, we propose directly leveraging useful clues from LQ images in the diffusion process by generating from pixel-error-based pre-restored LQ images with mild noise added. Moreover, we employ a small number of denoising steps to prevent the generation of redundant details that dilute crucial task-related information. We demonstrate that our method effectively utilizes diffusion prior for TDIR, significantly enhancing task performance and visual quality across diverse tasks with multiple complex degradations.

[553] Beyond Gloss: A Hand-Centric Framework for Gloss-Free Sign Language Translation

Sobhan Asasi, Mohamed Ilyas Lakhal, Ozge Mercanoglu Sincan, Richard Bowden

Main category: cs.CV

TL;DR: BeyondGloss is a gloss-free sign language translation framework that uses VideoLLMs with novel hand motion descriptions and contrastive alignment to bridge visual-linguistic modality gaps, achieving state-of-the-art results.

DetailsMotivation: Sign Language Translation faces challenges in bridging visual-linguistic modality gaps and capturing subtle hand shape/movement variations. Existing VideoLLMs struggle with long video details, requiring a gloss-free approach.

Method: Proposes fine-grained temporally-aware textual descriptions of hand motion, contrastive alignment module, distillation of features from HaMeR, and contrastive loss between video representations and target language embeddings.

Result: Achieves state-of-the-art performance on Phoenix14T and CSL-Daily benchmarks, demonstrating effective modality gap reduction and improved sign distinction.

Conclusion: BeyondGloss framework effectively addresses SLT challenges through VideoLLM enhancement with hand-centric temporal modeling and contrastive alignment, providing a robust gloss-free solution.

Abstract: Sign Language Translation (SLT) is a challenging task that requires bridging the modality gap between visual and linguistic information while capturing subtle variations in hand shapes and movements. To address these challenges, we introduce \textbf{BeyondGloss}, a novel gloss-free SLT framework that leverages the spatio-temporal reasoning capabilities of Video Large Language Models (VideoLLMs). Since existing VideoLLMs struggle to model long videos in detail, we propose a novel approach to generate fine-grained, temporally-aware textual descriptions of hand motion. A contrastive alignment module aligns these descriptions with video features during pre-training, encouraging the model to focus on hand-centric temporal dynamics and distinguish signs more effectively. To further enrich hand-specific representations, we distill fine-grained features from HaMeR. Additionally, we apply a contrastive loss between sign video representations and target language embeddings to reduce the modality gap in pre-training. \textbf{BeyondGloss} achieves state-of-the-art performance on the Phoenix14T and CSL-Daily benchmarks, demonstrating the effectiveness of the proposed framework. We will release the code upon acceptance of the paper.

[554] ACD-CLIP: Decoupling Representation and Dynamic Fusion for Zero-Shot Anomaly Detection

Ke Ma, Jun Long, Hongxiao Fei, Liujie Hua, Yiran Qian, Zhen Dai, Yueyi Luo

Main category: cs.CV

TL;DR: Architectural co-design framework with Conv-LoRA adapter and Dynamic Fusion Gateway improves zero-shot anomaly detection in vision-language models by addressing local inductive bias limitations and enabling adaptive cross-modal fusion.

DetailsMotivation: Pre-trained VLMs struggle with zero-shot anomaly detection due to lack of local inductive biases for dense prediction and inflexible feature fusion paradigms, creating an adaptation gap.

Method: Proposes parameter-efficient Convolutional Low-Rank Adaptation (Conv-LoRA) adapter to inject local inductive biases, and Dynamic Fusion Gateway (DFG) that uses visual context to adaptively modulate text prompts for bidirectional fusion.

Result: Extensive experiments on diverse industrial and medical benchmarks demonstrate superior accuracy and robustness compared to existing methods.

Conclusion: The synergistic architectural co-design framework is critical for robustly adapting foundation models to dense perception tasks like zero-shot anomaly detection.

Abstract: Pre-trained Vision-Language Models (VLMs) struggle with Zero-Shot Anomaly Detection (ZSAD) due to a critical adaptation gap: they lack the local inductive biases required for dense prediction and employ inflexible feature fusion paradigms. We address these limitations through an Architectural Co-Design framework that jointly refines feature representation and cross-modal fusion. Our method proposes a parameter-efficient Convolutional Low-Rank Adaptation (Conv-LoRA) adapter to inject local inductive biases for fine-grained representation, and introduces a Dynamic Fusion Gateway (DFG) that leverages visual context to adaptively modulate text prompts, enabling a powerful bidirectional fusion. Extensive experiments on diverse industrial and medical benchmarks demonstrate superior accuracy and robustness, validating that this synergistic co-design is critical for robustly adapting foundation models to dense perception tasks.

[555] TNet: Terrace Convolutional Decoder Network for Remote Sensing Image Semantic Segmentation

Chengqian Dai, Yonghong Guo, Hongzhao Xiang, Yigui Luo

Main category: cs.CV

TL;DR: TNet is a convolutional decoder network that progressively integrates low-resolution global context features into higher-resolution local detail features using only convolution and addition operations, achieving competitive performance on remote sensing segmentation benchmarks.

DetailsMotivation: Current segmentation networks focus on intra-scale relationships but neglect global contextual dependencies across multiple resolutions in remote sensing imagery.

Method: TNet uses a Terrace Convolutional Decoder that progressively fuses low-resolution (global context) features into higher-resolution (local details) features across decoding stages using convolution and addition operations only.

Result: TNet-R (with ResNet-18 encoder) achieves mIoU of 85.35% on ISPRS Vaihingen, 87.05% on ISPRS Potsdam, and 52.19% on LoveDA datasets while maintaining computational efficiency.

Conclusion: TNet demonstrates that simple convolution and addition operations can effectively integrate multi-resolution global-local features for competitive remote sensing segmentation performance without complex modules like Transformers or Mamba.

Abstract: In remote sensing, most segmentation networks adopt the UNet architecture, often incorporating modules such as Transformers or Mamba to enhance global-local feature interactions within decoder stages. However, these enhancements typically focus on intra-scale relationships and neglect the global contextual dependencies across multiple resolutions. To address this limitation, we introduce the Terrace Convolutional Decoder Network (TNet), a simple yet effective architecture that leverages only convolution and addition operations to progressively integrate low-resolution features (rich in global context) into higher-resolution features (rich in local details) across decoding stages. This progressive fusion enables the model to learn spatially-aware convolutional kernels that naturally blend global and local information in a stage-wise manner. We implement TNet with a ResNet-18 encoder (TNet-R) and evaluate it on three benchmark datasets. TNet-R achieves competitive performance with a mean Intersection-over-Union (mIoU) of 85.35% on ISPRS Vaihingen, 87.05% on ISPRS Potsdam, and 52.19% on LoveDA, while maintaining high computational efficiency. Code is publicly available.

[556] Preacher: Paper-to-Video Agentic System

Jingwei Liu, Ling Yang, Hao Luo, Fan Wang, Hongyan Li, Mengdi Wang

Main category: cs.CV

TL;DR: Preacher is the first paper-to-video agentic system that converts research papers into structured video abstracts using a top-down decomposition and bottom-up generation approach with Progressive Chain of Thought planning.

DetailsMotivation: Current video generation models have limitations including limited context windows, rigid duration constraints, limited stylistic diversity, and inability to represent domain-specific knowledge, which hinders effective paper-to-video conversion.

Method: Uses a top-down approach to decompose, summarize and reformulate papers, followed by bottom-up video generation. Introduces Progressive Chain of Thought (P-CoT) for granular iterative planning and defines key scenes to align cross-modal representations.

Result: Successfully generates high-quality video abstracts across five research fields, demonstrating expertise beyond current video generation models.

Conclusion: Preacher addresses the limitations of existing video generation models and provides an effective solution for converting research papers into accessible, well-organized video abstracts.

Abstract: The paper-to-video task converts a research paper into a structured video abstract, distilling key concepts, methods, and conclusions into an accessible, well-organized format. While state-of-the-art video generation models demonstrate potential, they are constrained by limited context windows, rigid video duration constraints, limited stylistic diversity, and an inability to represent domain-specific knowledge. To address these limitations, we introduce Preacher, the first paper-to-video agentic system. Preacher employs a topdown approach to decompose, summarize, and reformulate the paper, followed by bottom-up video generation, synthesizing diverse video segments into a coherent abstract. To align cross-modal representations, we define key scenes and introduce a Progressive Chain of Thought (P-CoT) for granular, iterative planning. Preacher successfully generates high-quality video abstracts across five research fields, demonstrating expertise beyond current video generation models. Code will be released at: https://github.com/GenVerse/Paper2Video

[557] Towards Comprehensive Cellular Characterisation of H&E slides

Benjamin Adjadj, Pierre-Antoine Bannier, Guillaume Horent, Sebastien Mandela, Aurore Lyon, Kathryn Schutte, Ulysse Marteau, Valentin Gaury, Laura Dumont, Thomas Mathieu, MOSAIC consortium, Reda Belbahri, Benoît Schmauch, Eric Durand, Katharina Von Loga, Lucie Gillet

Main category: cs.CV

TL;DR: HistoPLUS is a state-of-the-art cell analysis model that addresses poor performance on rare cell types and limited cross-domain generalization in tumor microenvironment analysis, achieving significant improvements in detection and classification with fewer parameters.

DetailsMotivation: Existing methods for cell detection, segmentation and classification in H&E slides suffer from poor performance on understudied cell types and limited cross-domain generalization, hindering comprehensive tumor microenvironment analysis.

Method: Developed HistoPLUS model trained on a novel curated pan-cancer dataset of 108,722 nuclei covering 13 cell types, using a more efficient architecture with fewer parameters.

Result: Outperforms state-of-the-art models by 5.2% in detection quality and 23.7% in overall F1 classification score, while using 5x fewer parameters. Enables study of 7 understudied cell types and shows robust transfer to unseen oncology indications.

Conclusion: HistoPLUS significantly advances cell analysis capabilities for tumor microenvironment research, particularly for rare cell types, and demonstrates strong cross-domain generalization with improved efficiency.

Abstract: Cell detection, segmentation and classification are essential for analyzing tumor microenvironments (TME) on hematoxylin and eosin (H&E) slides. Existing methods suffer from poor performance on understudied cell types (rare or not present in public datasets) and limited cross-domain generalization. To address these shortcomings, we introduce HistoPLUS, a state-of-the-art model for cell analysis, trained on a novel curated pan-cancer dataset of 108,722 nuclei covering 13 cell types. In external validation across 4 independent cohorts, HistoPLUS outperforms current state-of-the-art models in detection quality by 5.2% and overall F1 classification score by 23.7%, while using 5x fewer parameters. Notably, HistoPLUS unlocks the study of 7 understudied cell types and brings significant improvements on 8 of 13 cell types. Moreover, we show that HistoPLUS robustly transfers to two oncology indications unseen during training. To support broader TME biomarker research, we release the model weights and inference code at https://github.com/owkin/histoplus/.

[558] SocialTrack: Multi-Object Tracking in Complex Urban Traffic Scenes Inspired by Social Behavior

Wenguang Tao, Xiaotian Wang, Tian Yan, Jie Yan, Guodong Li, Kun Bai

Main category: cs.CV

TL;DR: SocialTrack is a novel UAV-based multi-object tracking framework that addresses challenges like small target variations, occlusions, and motion blur in complex urban traffic environments through specialized detection, adaptive filtering, group motion modeling, and spatio-temporal memory prediction.

DetailsMotivation: UAV-based multi-object tracking faces significant challenges in complex urban environments including small target scale variations, occlusions, nonlinear crossing motions, and motion blur, which hinder tracking stability and accuracy.

Method: Proposes SocialTrack framework with four key components: 1) Multi-scale feature enhancement mechanism for small-target detection, 2) Velocity Adaptive Cubature Kalman Filter (VACKF) for trajectory prediction, 3) Group Motion Compensation Strategy (GMCS) for social group motion modeling, and 4) Spatio-Temporal Memory Prediction (STMP) for leveraging historical trajectory information.

Result: Extensive experiments on UAVDT and MOT17 datasets show SocialTrack outperforms state-of-the-art methods, with significant improvements in MOTA and IDF1 metrics, demonstrating superior robustness and adaptability in complex dynamic environments.

Conclusion: SocialTrack provides an effective solution for UAV-based multi-object tracking in complex urban traffic scenarios, offering enhanced accuracy, robustness against identity switching, and modular compatibility with existing trackers for performance improvement.

Abstract: As a key research direction in the field of multi-object tracking (MOT), UAV-based multi-object tracking has significant application value in the analysis and understanding of urban intelligent transportation systems. However, in complex UAV perspectives, challenges such as small target scale variations, occlusions, nonlinear crossing motions, and motion blur severely hinder the stability of multi-object tracking. To address these challenges, this paper proposes a novel multi-object tracking framework, SocialTrack, aimed at enhancing the tracking accuracy and robustness of small targets in complex urban traffic environments. The specialized small-target detector enhances the detection performance by employing a multi-scale feature enhancement mechanism. The Velocity Adaptive Cubature Kalman Filter (VACKF) improves the accuracy of trajectory prediction by incorporating a velocity dynamic modeling mechanism. The Group Motion Compensation Strategy (GMCS) models social group motion priors to provide stable state update references for low-quality tracks, significantly improving the target association accuracy in complex dynamic environments. Furthermore, the Spatio-Temporal Memory Prediction (STMP) leverages historical trajectory information to predict the future state of low-quality tracks, effectively mitigating identity switching issues. Extensive experiments on the UAVDT and MOT17 datasets demonstrate that SocialTrack outperforms existing state-of-the-art (SOTA) methods across several key metrics. Significant improvements in MOTA and IDF1, among other core performance indicators, highlight its superior robustness and adaptability. Additionally, SocialTrack is highly modular and compatible, allowing for seamless integration with existing trackers to further enhance performance.

[559] FLAIR: Frequency- and Locality-Aware Implicit Neural Representations

Sukhun Ko, Dahyeon Kye, Kyle Min, Chanho Eom, Jihyong Oh

Main category: cs.CV

TL;DR: FLAIR introduces frequency- and locality-aware implicit neural representations with RC-GAUSS activation and WEGE encoding to overcome spectral bias and improve representation quality.

DetailsMotivation: Existing INRs lack frequency selectivity, spatial localization, and sparse representations, leading to spectral bias and difficulty capturing high-frequency details.

Method: Proposes FLAIR with two innovations: RC-GAUSS activation for explicit frequency selection and spatial localization under TFUP constraints, and Wavelet-Energy-Guided Encoding (WEGE) using DWT to guide frequency information.

Result: Consistently outperforms existing INRs in 2D image representation/restoration and 3D reconstruction tasks.

Conclusion: FLAIR successfully addresses spectral bias and improves implicit neural representations through frequency-aware activation and wavelet-guided encoding.

Abstract: Implicit Neural Representations (INRs) leverage neural networks to map coordinates to corresponding signals, enabling continuous and compact representations. This paradigm has driven significant advances in various vision tasks. However, existing INRs lack frequency selectivity, spatial localization, and sparse representations, leading to an over-reliance on redundant signal components. Consequently, they exhibit spectral bias, tending to learn low-frequency components early while struggling to capture fine high-frequency details. To address these issues, we propose FLAIR (Frequency- and Locality-Aware Implicit Neural Representations), which incorporates two key innovations. The first is RC-GAUSS, a novel activation designed for explicit frequency selection and spatial localization under the constraints of the time-frequency uncertainty principle (TFUP). The second is Wavelet-Energy-Guided Encoding (WEGE), which leverages the discrete wavelet transform (DWT) to compute energy scores and explicitly guide frequency information to the network. Our method consistently outperforms existing INRs in 2D image representation and restoration, as well as 3D reconstruction.

[560] PersonaVlog: Personalized Multimodal Vlog Generation with Multi-Agent Collaboration and Iterative Self-Correction

Xiaolu Hou, Bing Ma, Jiaxiang Cheng, Xuhua Ren, Kai Yu, Wenyue Li, Tianxiang Zheng, Qinglin Lu

Main category: cs.CV

TL;DR: PersonaVlog is an automated multimodal Vlog generation framework that uses MLLM-based multi-agent collaboration to create personalized videos with music and speech from theme and image inputs, featuring feedback mechanisms and a new evaluation benchmark.

DetailsMotivation: Growing demand for short videos and personalized content, with existing methods relying on predefined scripts lacking dynamism and personal expression.

Method: Multi-agent collaboration framework based on Multimodal Large Language Models (MLLMs) with feedback and rollback mechanism for iterative self-correction of multimodal content.

Result: Comprehensive experiments demonstrate significant advantages over baselines, showing effectiveness for automated Vlog generation.

Conclusion: The framework shows great potential for generating automated Vlogs with high personalization and multimodal collaboration capabilities.

Abstract: With the growing demand for short videos and personalized content, automated Video Log (Vlog) generation has become a key direction in multimodal content creation. Existing methods mostly rely on predefined scripts, lacking dynamism and personal expression. Therefore, there is an urgent need for an automated Vlog generation approach that enables effective multimodal collaboration and high personalization. To this end, we propose PersonaVlog, an automated multimodal stylized Vlog generation framework that can produce personalized Vlogs featuring videos, background music, and inner monologue speech based on a given theme and reference image. Specifically, we propose a multi-agent collaboration framework based on Multimodal Large Language Models (MLLMs). This framework efficiently generates high-quality prompts for multimodal content creation based on user input, thereby improving the efficiency and creativity of the process. In addition, we incorporate a feedback and rollback mechanism that leverages MLLMs to evaluate and provide feedback on generated results, thereby enabling iterative self-correction of multimodal content. We also propose ThemeVlogEval, a theme-based automated benchmarking framework that provides standardized metrics and datasets for fair evaluation. Comprehensive experiments demonstrate the significant advantages and potential of our framework over several baselines, highlighting its effectiveness and great potential for generating automated Vlogs.

[561] TPA: Temporal Prompt Alignment for Fetal Congenital Heart Defect Classification

Darya Taratynova, Alya Almsouti, Beknur Kalmakhanbet, Numan Saeed, Mohammad Yaqub

Main category: cs.CV

TL;DR: TPA is a novel framework for fetal congenital heart defect classification in ultrasound videos that combines temporal modeling, prompt-aware contrastive learning, and uncertainty quantification to achieve state-of-the-art performance with improved calibration.

DetailsMotivation: Current automated methods for CHD detection in ultrasound videos neglect temporal information, limit to binary classification, and lack prediction calibration, which hinders clinical reliability.

Method: Temporal Prompt Alignment (TPA) extracts frame features with an image encoder, aggregates them with a temporal extractor, aligns video representations with class-specific text prompts using margin-hinge contrastive loss, and incorporates CVAESM module for uncertainty quantification and style modulation.

Result: TPA achieves 85.40% macro F1 score for CHD diagnosis, reduces expected calibration error by 5.38% and adaptive ECE by 6.8%, and boosts macro F1 by 4.73% on EchoNet-Dynamic’s three-class task.

Conclusion: TPA effectively addresses limitations of current methods by integrating temporal modeling, prompt learning, and uncertainty quantification, demonstrating superior performance and improved calibration for clinical ultrasound video analysis.

Abstract: Congenital heart defect (CHD) detection in ultrasound videos is hindered by image noise and probe positioning variability. While automated methods can reduce operator dependence, current machine learning approaches often neglect temporal information, limit themselves to binary classification, and do not account for prediction calibration. We propose Temporal Prompt Alignment (TPA), a method leveraging foundation image-text model and prompt-aware contrastive learning to classify fetal CHD on cardiac ultrasound videos. TPA extracts features from each frame of video subclips using an image encoder, aggregates them with a trainable temporal extractor to capture heart motion, and aligns the video representation with class-specific text prompts via a margin-hinge contrastive loss. To enhance calibration for clinical reliability, we introduce a Conditional Variational Autoencoder Style Modulation (CVAESM) module, which learns a latent style vector to modulate embeddings and quantifies classification uncertainty. Evaluated on a private dataset for CHD detection and on a large public dataset, EchoNet-Dynamic, for systolic dysfunction, TPA achieves state-of-the-art macro F1 scores of 85.40% for CHD diagnosis, while also reducing expected calibration error by 5.38% and adaptive ECE by 6.8%. On EchoNet-Dynamic’s three-class task, it boosts macro F1 by 4.73% (from 53.89% to 58.62%). Temporal Prompt Alignment (TPA) is a framework for fetal congenital heart defect (CHD) classification in ultrasound videos that integrates temporal modeling, prompt-aware contrastive learning, and uncertainty quantification.

[562] DIO: Refining Mutual Information and Causal Chain to Enhance Machine Abstract Reasoning Ability

Ruizhuo Song, Beiming Yuan

Main category: cs.CV

TL;DR: This paper addresses the abstract reasoning bottleneck in deep learning by focusing on Raven’s Progressive Matrices problems, proposing causal chain modeling and three improvement methods to overcome limitations in baseline models.

DetailsMotivation: Current deep learning models struggle with abstract reasoning despite strong performance in other domains. Raven's Progressive Matrices serve as an authoritative benchmark to evaluate and enhance abstract reasoning, pattern recognition, and problem-solving capabilities in machine intelligence.

Method: The paper adopts a causal chain modeling perspective to analyze RPM tasks, designs a baseline model (DIO), and progressively proposes three improvement methods to address limitations in mutual information maximization and causal relationship capture.

Result: Experiments revealed that the baseline model’s optimization objective (maximizing variational lower bound of mutual information) failed to enable genuine acquisition of human reasoning logic due to tightness issues and statistical nature of mutual information.

Conclusion: The paper proposes progressive improvements to overcome the limitations of mutual information-based approaches in capturing causal relationships and enabling true abstract reasoning capabilities in deep learning models.

Abstract: Despite the outstanding performance of current deep learning models across various domains, their fundamental bottleneck in abstract reasoning remains unresolved. To address this challenge, the academic community has introduced Raven’s Progressive Matrices (RPM) problems as an authoritative benchmark for evaluating the abstract reasoning capabilities of deep learning algorithms, with a focus on core intelligence dimensions such as abstract reasoning, pattern recognition, and complex problem-solving. Therefore, this paper centers on solving RPM problems, aiming to contribute to enhancing the abstract reasoning abilities of machine intelligence. Firstly, this paper adopts a ``causal chain modeling’’ perspective to systematically analyze the complete causal chain in RPM tasks: image $\rightarrow$ abstract attributes $\rightarrow$ progressive attribute patterns $\rightarrow$ pattern consistency $\rightarrow$ correct answer. Based on this analysis, the network architecture of the baseline model DIO is designed. However, experiments reveal that the optimization objective formulated for DIO, namely maximizing the variational lower bound of mutual information between the context and the correct option, fails to enable the model to genuinely acquire the predefined human reasoning logic. This is attributed to two main reasons: the tightness of the lower bound significantly impacts the effectiveness of mutual information maximization, and mutual information, as a statistical measure, does not capture the causal relationship between subjects and objects. To overcome these limitations, this paper progressively proposes three improvement methods:

[563] Aligning Moments in Time using Video Queries

Yogesh Kumar, Uday Agarwal, Manish Gupta, Anand Mishra

Main category: cs.CV

TL;DR: MATR is a transformer-based model for video-to-video moment retrieval that uses dual-stage sequence alignment and self-supervised pre-training to achieve significant performance improvements over state-of-the-art methods.

DetailsMotivation: Video-to-video moment retrieval requires semantic frame-level alignment and modeling complex dependencies between query and target videos, which existing methods struggle with.

Method: MATR uses transformer architecture with dual-stage sequence alignment to condition target video representations on query features, plus self-supervised pre-training by localizing random clips within videos.

Result: Achieves 13.1% R@1 and 8.1% mIoU improvement on ActivityNet-VRL, and 14.7% R@1 and 14.4% mIoU gain on SportsMoments dataset compared to state-of-the-art methods.

Conclusion: MATR effectively addresses the challenges of video-to-video moment retrieval through its transformer-based architecture and self-supervised pre-training, demonstrating substantial performance gains across multiple datasets.

Abstract: Video-to-video moment retrieval (Vid2VidMR) is the task of localizing unseen events or moments in a target video using a query video. This task poses several challenges, such as the need for semantic frame-level alignment and modeling complex dependencies between query and target videos. To tackle this challenging problem, we introduce MATR (Moment Alignment TRansformer), a transformer-based model designed to capture semantic context as well as the temporal details necessary for precise moment localization. MATR conditions target video representations on query video features using dual-stage sequence alignment that encodes the required correlations and dependencies. These representations are then used to guide foreground/background classification and boundary prediction heads, enabling the model to accurately identify moments in the target video that semantically match with the query video. Additionally, to provide a strong task-specific initialization for MATR, we propose a self-supervised pre-training technique that involves training the model to localize random clips within videos. Extensive experiments demonstrate that MATR achieves notable performance improvements of 13.1% in R@1 and 8.1% in mIoU on an absolute scale compared to state-of-the-art methods on the popular ActivityNet-VRL dataset. Additionally, on our newly proposed dataset, SportsMoments, MATR shows a 14.7% gain in R@1 and a 14.4% gain in mIoU on an absolute scale over strong baselines.

[564] Structuring GUI Elements through Vision Language Models: Towards Action Space Generation

Yi Xu, Yesheng Zhang, Jiajia Liu, Jingdong Chen

Main category: cs.CV

TL;DR: Proposes IAML training paradigm to improve MLLMs’ precision in generating UI element coordinates by addressing semantic void in numerical coordinate prediction through IoU-based data augmentation.

DetailsMotivation: MLLMs struggle with precise UI coordinate generation due to semantic void around numerical values in language spaces and next-token prediction limitations, requiring better training approaches.

Method: Introduces IoU-Augmented Maximum Likelihood (IAML) with novel IoU-based coordinate sampling pipeline for data augmentation, fine-tuning MLLMs to mitigate exposure bias in traditional training.

Result: Extensive experiments show superior performance of IAML approach over traditional training paradigms for GUI element coordinate generation.

Conclusion: IAML training paradigm effectively addresses MLLMs’ coordinate prediction challenges and improves GUI understanding capabilities through strategic data augmentation and bias mitigation.

Abstract: Multimodal large language models (MLLMs) have emerged as pivotal tools in enhancing human-computer interaction. In this paper we focus on the application of MLLMs in the field of graphical user interface (GUI) elements structuring, where they assist in processing user instructions based on screen contents. Despite the promise of MLLMs, their performance in precisely generating UI element coordinates, a critical aspect of GUI understanding, is hindered by the nature of next-token prediction training. This challenge arises from the semantic void surrounding numerical UI coordinates in language representation spaces, necessitating a substantial and diverse dataset to bolster visual module capabilities. To address these limitations, we introduce an IoU-Augmented Maximum Likelihood (IAML) training paradigm. Specifically, our approach involves a novel pipeline for IoU-based coordinate sampling to augment the training data, which considers the proximity to ground truth coordinates. This data augmentation strategy is then employed to fine-tune MLLMs under the IAML paradigm, which is designed to mitigate the exposure bias problem inherent in traditional maximum likelihood estimation. Through extensive experiments, we demonstrate the superior performance of our IAML training approach over traditional training paradigms.

[565] MSNav: Zero-Shot Vision-and-Language Navigation with Dynamic Memory and LLM Spatial Reasoning

Chenghao Liu, Zhimu Zhou, Jiachen Zhang, Minghao Zhang, Songfang Huang, Huiling Duan

Main category: cs.CV

TL;DR: MSNav is a novel framework that addresses VLN challenges by integrating memory, spatial reasoning, and decision modules to overcome LLM limitations in navigation tasks.

DetailsMotivation: Current VLN approaches using single LLMs suffer from poor spatial reasoning, weak cross-modal grounding, and memory overload in long-horizon navigation tasks.

Method: MSNav integrates three modules: Memory Module for dynamic map memory with selective node pruning, Spatial Module for spatial reasoning and object relationship inference, and Decision Module for LLM-based path planning. Also introduces I-O-S dataset and fine-tunes Qwen3-4B into Qwen-Spatial model.

Result: Achieves state-of-the-art performance on R2R and REVERIE datasets with significant improvements in Success Rate (SR) and Success weighted by Path Length (SPL). Qwen-Spatial outperforms commercial LLMs in object list extraction with higher F1 and NDCG scores.

Conclusion: MSNav’s synergistic architecture transforms fragile inference into robust, integrated intelligence for vision-and-language navigation, effectively addressing critical vulnerabilities of single-LLM approaches.

Abstract: Vision-and-Language Navigation (VLN) requires an agent to interpret natural language instructions and navigate complex environments. Current approaches often adopt a “black-box” paradigm, where a single Large Language Model (LLM) makes end-to-end decisions. However, it is plagued by critical vulnerabilities, including poor spatial reasoning, weak cross-modal grounding, and memory overload in long-horizon tasks. To systematically address these issues, we propose Memory Spatial Navigation(MSNav), a framework that fuses three modules into a synergistic architecture, which transforms fragile inference into a robust, integrated intelligence. MSNav integrates three modules: Memory Module, a dynamic map memory module that tackles memory overload through selective node pruning, enhancing long-range exploration; Spatial Module, a module for spatial reasoning and object relationship inference that improves endpoint recognition; and Decision Module, a module using LLM-based path planning to execute robust actions. Powering Spatial Module, we also introduce an Instruction-Object-Space (I-O-S) dataset and fine-tune the Qwen3-4B model into Qwen-Spatial (Qwen-Sp), which outperforms leading commercial LLMs in object list extraction, achieving higher F1 and NDCG scores on the I-O-S test set. Extensive experiments on the Room-to-Room (R2R) and REVERIE datasets demonstrate MSNav’s state-of-the-art performance with significant improvements in Success Rate (SR) and Success weighted by Path Length (SPL).

[566] Contrastive Prompt Clustering for Weakly Supervised Semantic Segmentation

Wangyu Wu, Zhenhong Chen, Xiaowen Ma, Wenqiao Zhang, Xianglin Qiu, Siqi Song, Xiaowei Huang, Fei Ma, Jimin Xiao

Main category: cs.CV

TL;DR: CPC is a novel WSSS framework that uses LLMs to create category clusters capturing inter-class relationships and employs class-aware patch-level contrastive loss for better intra-class consistency and inter-class separation.

DetailsMotivation: Existing WSSS methods focus too much on inter-class separation while neglecting shared semantics among related categories and lack fine-grained discrimination, leading to confusion among visually similar categories.

Method: Uses Large Language Models to derive category clusters encoding intrinsic inter-class relationships, and introduces class-aware patch-level contrastive loss to enforce intra-class consistency and inter-class separation in a hierarchical design.

Result: Experiments on PASCAL VOC 2012 and MS COCO 2014 show that CPC surpasses existing state-of-the-art methods in Weakly Supervised Semantic Segmentation.

Conclusion: CPC effectively addresses the limitations of previous WSSS methods by leveraging hierarchical semantic priors from LLMs and contrastive learning, achieving superior performance while maintaining cost-effectiveness of image-level labels.

Abstract: Weakly Supervised Semantic Segmentation (WSSS) with image-level labels has gained attention for its cost-effectiveness. Most existing methods emphasize inter-class separation, often neglecting the shared semantics among related categories and lacking fine-grained discrimination. To address this, we propose Contrastive Prompt Clustering (CPC), a novel WSSS framework. CPC exploits Large Language Models (LLMs) to derive category clusters that encode intrinsic inter-class relationships, and further introduces a class-aware patch-level contrastive loss to enforce intra-class consistency and inter-class separation. This hierarchical design leverages clusters as coarse-grained semantic priors while preserving fine-grained boundaries, thereby reducing confusion among visually similar categories. Experiments on PASCAL VOC 2012 and MS COCO 2014 demonstrate that CPC surpasses existing state-of-the-art methods in WSSS.

[567] Lightweight Joint Optimization of General-Purpose Vision-Language Models and Retrievers for Medical Diagnosis

Nir Mazor, Tom Hope

Main category: cs.CV

TL;DR: A joint optimization model combining multimodal retriever with LVLM for medical diagnosis, outperforming standard RAG and achieving competitive results with medical-pretrained models using only general-purpose backbones.

DetailsMotivation: To enhance diagnostic accuracy by retrieving relevant visual information from medical literature and hospital records, addressing limitations of standard RAG where error signals don't propagate to the retriever.

Method: Joint optimization of multimodal retriever with LVLM for medical diagnosis, using general-purpose backbones with lightweight fine-tuning only.

Result: Competitive performance with medically-pretrained models across clinical multi-label classification and visual question answering tasks. Significant improvement over standard RAG for challenging cases where different retrieved images lead to different predictions.

Conclusion: While correct diagnosis is frequently achievable with top retrieved images, there remains a large performance gap from oracle performance, and current rerankers using frontier LVLMs don’t close this gap, indicating substantial room for improvement.

Abstract: Clinical decision-making often involves interpreting images (e.g., radiology) for making diagnoses. Retrieving relevant visual information from medical literature and hospital records could enhance diagnostic accuracy. In this paper, we develop a model in which a multimodal retriever is jointly optimized with an LVLM for medical diagnosis, unlike standard RAG where LVLM error signal is not propagated down to the retriever. We show that using only general-purpose backbones, with only lightweight fine-tuning, our model is able to achieve competitive results with medically-pretrained models across clinical multi-label classification and visual question answering tasks. In a novel analysis, we additionally find that in many cases different top retrieved images each lead to different predictions for a given target, and that these cases are empirically challenging for all models, even for non-retrieval models. Our joint retrieval optimization significantly improves these challenging cases over standard RAG. However, oracle analysis reveals that while the correct diagnosis is frequently achievable using one of the top retrieved images, in practice there is a large performance gap from the oracle, and rerankers using frontier LVLMs do not close this gap – leaving ample room for improvement by future methods. Code will be made publicly available.

[568] Investigating Domain Gaps for Indoor 3D Object Detection

Zijing Zhao, Zhu Xu, Qingchao Chen, Yuxin Peng, Yang Liu

Main category: cs.CV

TL;DR: This paper introduces a comprehensive benchmark for domain adaptive indoor 3D object detection, addressing the problem of detectors overfitting to specific dataset characteristics like point cloud quality, layout, and instance features.

DetailsMotivation: Existing 3D object detection research has been limited to datasets with identical training and testing distributions, but real-world applications require detectors that can generalize across different indoor environments and data collection methods.

Method: The authors create a benchmark using ScanNet, SUN RGB-D, 3D Front datasets, plus new synthetic datasets ProcTHOR-OD and ProcFront. They conduct experiments across various adaptation scenarios including synthetic-to-real, point cloud quality, layout, and instance feature adaptation.

Result: The paper analyzes the impact of different domain gaps on 3D object detectors and provides baseline approaches to improve adaptation performance across datasets.

Conclusion: This work establishes foundational benchmarks for domain adaptive indoor 3D object detection, aiming to inspire future development of detectors with stronger cross-domain generalization capabilities.

Abstract: As a fundamental task for indoor scene understanding, 3D object detection has been extensively studied, and the accuracy on indoor point cloud data has been substantially improved. However, existing researches have been conducted on limited datasets, where the training and testing sets share the same distribution. In this paper, we consider the task of adapting indoor 3D object detectors from one dataset to another, presenting a comprehensive benchmark with ScanNet, SUN RGB-D and 3D Front datasets, as well as our newly proposed large-scale datasets ProcTHOR-OD and ProcFront generated by a 3D simulator. Since indoor point cloud datasets are collected and constructed in different ways, the object detectors are likely to overfit to specific factors within each dataset, such as point cloud quality, bounding box layout and instance features. We conduct experiments across datasets on different adaptation scenarios including synthetic-to-real adaptation, point cloud quality adaptation, layout adaptation and instance feature adaptation, analyzing the impact of different domain gaps on 3D object detectors. We also introduce several approaches to improve adaptation performances, providing baselines for domain adaptive indoor 3D object detection, hoping that future works may propose detectors with stronger generalization ability across domains. Our project homepage can be found in https://jeremyzhao1998.github.io/DAVoteNet-release/.

[569] Occlusion Robustness of CLIP for Military Vehicle Classification

Jan Erik van Woerden, Gertjan Burghouts, Lotte Nijskens, Alma M. Liezenga, Sabina van Rooij, Frank Ruis, Hugo J. Kuijf

Main category: cs.CV

TL;DR: CLIP’s robustness to occlusion in military environments is tested, showing Transformer models outperform CNNs, dispersed occlusions are more damaging than contiguous ones, and finetuning backbone significantly improves occlusion resilience.

DetailsMotivation: Evaluate CLIP's zero-shot classification robustness in challenging military environments with occlusion and degraded SNR, which remains underexplored despite its potential for defense applications with limited labeled data.

Method: Used custom dataset of 18 military vehicle classes to test CLIP variants’ occlusion robustness, measuring performance with Normalized Area Under the Curve (NAUC) across different occlusion percentages and types.

Result: Transformer-based CLIP models outperform CNNs; fine-grained dispersed occlusions degrade performance more than larger contiguous ones; linear-probed models show sharp performance drop at ~35% occlusion; backbone finetuning pushes performance drop threshold to >60% occlusion.

Conclusion: Occlusion-specific augmentations during training are crucial, and further research is needed on patch-level sensitivity and architectural resilience for real-world CLIP deployment in military applications.

Abstract: Vision-language models (VLMs) like CLIP enable zero-shot classification by aligning images and text in a shared embedding space, offering advantages for defense applications with scarce labeled data. However, CLIP’s robustness in challenging military environments, with partial occlusion and degraded signal-to-noise ratio (SNR), remains underexplored. We investigate CLIP variants’ robustness to occlusion using a custom dataset of 18 military vehicle classes and evaluate using Normalized Area Under the Curve (NAUC) across occlusion percentages. Four key insights emerge: (1) Transformer-based CLIP models consistently outperform CNNs, (2) fine-grained, dispersed occlusions degrade performance more than larger contiguous occlusions, (3) despite improved accuracy, performance of linear-probed models sharply drops at around 35% occlusion, (4) by finetuning the model’s backbone, this performance drop occurs at more than 60% occlusion. These results underscore the importance of occlusion-specific augmentations during training and the need for further exploration into patch-level sensitivity and architectural resilience for real-world deployment of CLIP.

[570] EventTracer: Fast Path Tracing-based Event Stream Rendering

Zhenyang Li, Xiaoyang Bai, Jinfan Lu, Pengfei Shen, Edmund Y. Lam, Yifan Peng

Main category: cs.CV

TL;DR: EventTracer is a path tracing-based rendering pipeline that efficiently simulates high-fidelity event sequences from 3D scenes using low SPP path tracing and a lightweight event spiking network with BiLIF units and EMD loss.

DetailsMotivation: Existing event stream simulation methods use expensive noiseless RGB frames and achieve only 100-300 FPS, which is far lower than real-world event data temporal resolution, creating a need for more efficient and physics-aware simulation.

Method: Uses low sample-per-pixel path tracing for efficient rendering, trains a lightweight event spiking network with bipolar leaky integrate-and-fired units and bidirectional earth mover distance loss to denoise RGB videos into realistic event sequences.

Result: EventTracer runs at about 4 minutes per second of 720p video, captures better scene details, shows greater similarity to real-world event data than other simulators, and enables accurate spatiotemporal modeling.

Conclusion: Establishes EventTracer as a promising tool for creating large-scale event-RGB datasets at low cost, narrowing the sim-to-real gap in event-based vision, and boosting applications in robotics, autonomous driving, and VR/AR.

Abstract: Simulating event streams from 3D scenes has become a common practice in event-based vision research, as it meets the demand for large-scale, high temporal frequency data without setting up expensive hardware devices or undertaking extensive data collections. Yet existing methods in this direction typically work with noiseless RGB frames that are costly to render, and therefore they can only achieve a temporal resolution equivalent to 100-300 FPS, far lower than that of real-world event data. In this work, we propose EventTracer, a path tracing-based rendering pipeline that simulates high-fidelity event sequences from complex 3D scenes in an efficient and physics-aware manner. Specifically, we speed up the rendering process via low sample-per-pixel (SPP) path tracing, and train a lightweight event spiking network to denoise the resulting RGB videos into realistic event sequences. To capture the physical properties of event streams, the network is equipped with a bipolar leaky integrate-and-fired (BiLIF) spiking unit and trained with a bidirectional earth mover distance (EMD) loss. Our EventTracer pipeline runs at a speed of about 4 minutes per second of 720p video, and it inherits the merit of accurate spatiotemporal modeling from its path tracing backbone. We show in two downstream tasks that EventTracer captures better scene details and demonstrates a greater similarity to real-world event data than other event simulators, which establishes it as a promising tool for creating large-scale event-RGB datasets at a low cost, narrowing the sim-to-real gap in event-based vision, and boosting various application scenarios such as robotics, autonomous driving, and VRAR.

[571] A Novel Deep Hybrid Framework with Ensemble-Based Feature Optimization for Robust Real-Time Human Activity Recognition

Wasi Ullah, Yasir Noman Khalid, Saddam Hussain Khan

Main category: cs.CV

TL;DR: An optimized hybrid deep learning framework for Human Activity Recognition that combines customized InceptionV3, LSTM, and ensemble feature selection to achieve high accuracy with minimal features for real-time deployment.

DetailsMotivation: Address challenges in HAR systems including high computational costs, redundant features, and limited scalability in real-time scenarios for applications like smart surveillance, healthcare, and assistive technologies.

Method: Integrates customized InceptionV3 for spatial feature extraction, LSTM for temporal dependency modeling, and ensemble-based genetic algorithm with ADFSA for optimized feature selection.

Result: Achieves 99.65% recognition accuracy on UCF-YouTube dataset, reduces features to as few as 7, and enhances inference time for real-time deployment on edge devices.

Conclusion: The lightweight and scalable framework enables practical real-time HAR applications in resource-aware environments including public safety, assistive technology, and autonomous monitoring systems.

Abstract: Human Activity Recognition (HAR) plays a pivotal role in various applications, including smart surveillance, healthcare, assistive technologies, sports analytics, etc. However, HAR systems still face critical challenges, including high computational costs, redundant features, and limited scalability in real-time scenarios. An optimized hybrid deep learning framework is introduced that integrates a customized InceptionV3, an LSTM architecture, and a novel ensemble-based feature selection strategy. The proposed framework first extracts spatial descriptors using the customized InceptionV3 model, which captures multilevel contextual patterns, region homogeneity, and fine-grained localization cues. The temporal dependencies across frames are then modeled using LSTMs to effectively encode motion dynamics. Finally, an ensemble-based genetic algorithm with Adaptive Dynamic Fitness Sharing and Attention (ADFSA) is employed to select a compact and optimized feature set by dynamically balancing objectives such as accuracy, redundancy, uniqueness, and complexity reduction. Consequently, the selected feature subsets, which are both diverse and discriminative, enable various lightweight machine learning classifiers to achieve accurate and robust HAR in heterogeneous environments. Experimental results on the robust UCF-YouTube dataset, which presents challenges such as occlusion, cluttered backgrounds, motion dynamics, and poor illumination, demonstrate good performance. The proposed approach achieves 99.65% recognition accuracy, reduces features to as few as 7, and enhances inference time. The lightweight and scalable nature of the HAR system supports real-time deployment on edge devices such as Raspberry Pi, enabling practical applications in intelligent, resource-aware environments, including public safety, assistive technology, and autonomous monitoring systems.

[572] R-4B: Incentivizing General-Purpose Auto-Thinking Capability in MLLMs via Bi-Mode Annealing and Reinforce Learning

Qi Yang, Bolin Ni, Shiming Xiang, Han Hu, Houwen Peng, Jie Jiang

Main category: cs.CV

TL;DR: R-4B is an auto-thinking MLLM that adaptively decides when to use step-by-step thinking based on problem complexity, achieving state-of-the-art performance with lower computational cost.

DetailsMotivation: Current MLLMs use step-by-step thinking for all problems, which is inefficient for simple problems that don't require complex reasoning.

Method: Uses bi-mode annealing to enable both thinking and non-thinking capabilities, followed by Bi-mode Policy Optimization (BPO) to improve decision-making on when to activate thinking. Trained on curated dataset with both modes and improved GRPO framework.

Result: Achieves SOTA performance across 25 benchmarks, outperforms Qwen2.5-VL-7B in most tasks, and matches larger models like Kimi-VL-A3B-Thinking-2506 (16B) with lower computational cost.

Conclusion: R-4B provides an efficient solution by adaptively using thinking only when needed, demonstrating superior performance with reduced computational requirements.

Abstract: Multimodal Large Language Models (MLLMs) equipped with step-by-step thinking capabilities have demonstrated remarkable performance on complex reasoning problems. However, this thinking process is redundant for simple problems solvable without complex reasoning. To address this inefficiency, we propose R-4B, an auto-thinking MLLM, which can adaptively decide when to think based on problem complexity. The central idea of R-4B is to empower the model with both thinking and non-thinking capabilities using bi-mode annealing, and apply Bi-mode Policy Optimization (BPO) to improve the model’s accuracy in determining whether to activate the thinking process. Specifically, we first train the model on a carefully curated dataset spanning various topics, which contains samples from both thinking and non-thinking modes. Then it undergoes a second phase of training under an improved GRPO framework, where the policy model is forced to generate responses from both modes for each input query. Experimental results show that R-4B achieves state-of-the-art performance across 25 challenging benchmarks. It outperforms Qwen2.5-VL-7B in most tasks and achieves performance comparable to larger models such as Kimi-VL-A3B-Thinking-2506 (16B) on reasoning-intensive benchmarks with lower computational cost.

[573] Drawing2CAD: Sequence-to-Sequence Learning for CAD Generation from Vector Drawings

Feiwei Qin, Shichao Lu, Junhao Hou, Changmiao Wang, Meie Fang, Ligang Liu

Main category: cs.CV

TL;DR: Drawing2CAD is a framework that converts 2D engineering drawings to parametric CAD models using sequence-to-sequence learning with transformer architecture.

DetailsMotivation: Traditional CAD generative methods don't align with industrial workflows that start from 2D engineering drawings, creating a gap in automated parametric CAD generation from vector drawings.

Method: Uses a dual-decoder transformer architecture with network-friendly vector representation, decoupling command type and parameter generation while maintaining geometric precision through soft target distribution loss.

Result: Developed CAD-VGDrawing dataset and demonstrated effective conversion of 2D drawings to parametric CAD models with preserved geometric precision.

Conclusion: The framework successfully bridges the gap between traditional 2D engineering workflows and modern parametric CAD generation, enabling automated conversion while maintaining design intent.

Abstract: Computer-Aided Design (CAD) generative modeling is driving significant innovations across industrial applications. Recent works have shown remarkable progress in creating solid models from various inputs such as point clouds, meshes, and text descriptions. However, these methods fundamentally diverge from traditional industrial workflows that begin with 2D engineering drawings. The automatic generation of parametric CAD models from these 2D vector drawings remains underexplored despite being a critical step in engineering design. To address this gap, our key insight is to reframe CAD generation as a sequence-to-sequence learning problem where vector drawing primitives directly inform the generation of parametric CAD operations, preserving geometric precision and design intent throughout the transformation process. We propose Drawing2CAD, a framework with three key technical components: a network-friendly vector primitive representation that preserves precise geometric information, a dual-decoder transformer architecture that decouples command type and parameter generation while maintaining precise correspondence, and a soft target distribution loss function accommodating inherent flexibility in CAD parameters. To train and evaluate Drawing2CAD, we create CAD-VGDrawing, a dataset of paired engineering drawings and parametric CAD models, and conduct thorough experiments to demonstrate the effectiveness of our method. Code and dataset are available at https://github.com/lllssc/Drawing2CAD.

[574] ELV-Halluc: Benchmarking Semantic Aggregation Hallucinations in Long Video Understanding

Hao Lu, Jiahao Wang, Yaolun Zhang, Ruohui Wang, Xuanyu Zheng, Yepeng Tang, Dahua Lin, Lewei Lu

Main category: cs.CV

TL;DR: ELV-Halluc is the first benchmark for long-video hallucination in Video-MLLMs, focusing on Semantic Aggregation Hallucination (SAH) where models generate incorrect outputs despite correct frame-level semantics.

DetailsMotivation: Current video hallucination benchmarks focus on short videos and oversimplify hallucination causes. SAH occurs during frame-to-event semantic aggregation and becomes critical in long videos due to increased semantic complexity.

Method: Introduce ELV-Halluc benchmark, analyze SAH patterns, experiment with positional encoding strategies, and use DPO (Direct Preference Optimization) with 8K adversarial data pairs to enhance semantic distinction capabilities.

Result: Confirmed SAH existence showing it increases with semantic complexity, especially with rapidly changing semantics. Achieved 27.7% SAH reduction and improvements on both ELV-Halluc and Video-MME benchmarks.

Conclusion: SAH is a distinct hallucination type in long videos that requires specialized mitigation approaches. Positional encoding and DPO strategies effectively reduce SAH, highlighting the need for targeted solutions for long-video understanding challenges.

Abstract: Video multimodal large language models (Video-MLLMs) have achieved remarkable progress in video understanding. However, they remain vulnerable to hallucination-producing content inconsistent with or unrelated to video inputs. Previous video hallucination benchmarks primarily focus on short-videos. They attribute hallucinations to factors such as strong language priors, missing frames, or vision-language biases introduced by the visual encoder. While these causes indeed account for most hallucinations in short videos, they still oversimplify the cause of hallucinations. Sometimes, models generate incorrect outputs but with correct frame-level semantics. We refer to this type of hallucination as Semantic Aggregation Hallucination (SAH), which arises during the process of aggregating frame-level semantics into event-level semantic groups. Given that SAH becomes particularly critical in long videos due to increased semantic complexity across multiple events, it is essential to separate and thoroughly investigate the causes of this type of hallucination. To address the above issues, we introduce ELV-Halluc, the first benchmark dedicated to long-video hallucination, enabling a systematic investigation of SAH. Our experiments confirm the existence of SAH and show that it increases with semantic complexity. Additionally, we find that models are more prone to SAH on rapidly changing semantics. Moreover, we discuss potential approaches to mitigate SAH. We demonstrate that positional encoding strategy contributes to alleviating SAH, and further adopt DPO strategy to enhance the model’s ability to distinguish semantics within and across events. To support this, we curate a dataset of 8K adversarial data pairs and achieve improvements on both ELV-Halluc and Video-MME, including a substantial 27.7% reduction in SAH ratio.

[575] FusionCounting: Robust visible-infrared image fusion guided by crowd counting via multi-task learning

He Li, Xinyu Liu, Weihang Kong, Xingchen Zhang

Main category: cs.CV

TL;DR: FusionCounting integrates visible and infrared image fusion with crowd counting in a unified multi-task framework, using crowd counting’s minimal annotation requirements to guide fusion in dense scenes.

DetailsMotivation: Existing VIF methods focus on image quality or use semantic segmentation/detection which require extensive annotations. Crowd counting provides quantitative density measurement with minimal annotation, making it suitable for dense scenes where detection struggles with occlusion.

Method: Multi-task learning framework that jointly optimizes VIF and crowd counting. Uses dynamic loss weighting for task balance and incorporates adversarial training for robustness. Leverages population density information to guide fusion process.

Result: Experimental results show improved image fusion quality and superior crowd counting performance compared to existing methods on public datasets.

Conclusion: Integrating crowd counting with VIF creates a mutually beneficial framework that enhances both tasks, particularly effective in dense crowded scenes with minimal annotation requirements.

Abstract: Visible and infrared image fusion (VIF) is an important multimedia task in computer vision. Most VIF methods focus primarily on optimizing fused image quality. Recent studies have begun incorporating downstream tasks, such as semantic segmentation and object detection, to provide semantic guidance for VIF. However, semantic segmentation requires extensive annotations, while object detection, despite reducing annotation efforts compared with segmentation, faces challenges in highly crowded scenes due to overlapping bounding boxes and occlusion. Moreover, although RGB-T crowd counting has gained increasing attention in recent years, no studies have integrated VIF and crowd counting into a unified framework. To address these challenges, we propose FusionCounting, a novel multi-task learning framework that integrates crowd counting into the VIF process. Crowd counting provides a direct quantitative measure of population density with minimal annotation, making it particularly suitable for dense scenes. Our framework leverages both input images and population density information in a mutually beneficial multi-task design. To accelerate convergence and balance tasks contributions, we introduce a dynamic loss function weighting strategy. Furthermore, we incorporate adversarial training to enhance the robustness of both VIF and crowd counting, improving the model’s stability and resilience to adversarial attacks. Experimental results on public datasets demonstrate that FusionCounting not only enhances image fusion quality but also achieves superior crowd counting performance.

cs.AI

[576] A Comparative Study of Controllability, Explainability, and Performance in Dysfluency Detection Models

Eric Zhang, Li Wei, Sarah Chen, Michael Wang

Main category: cs.AI

TL;DR: Systematic comparison of 4 dysfluency detection approaches (YOLO-Stutter, FluentNet, UDM, SSDM) across performance, controllability, and explainability dimensions, finding UDM offers best balance for clinical use.

DetailsMotivation: Clinical adoption of dysfluency detection models requires more than accuracy - models must be controllable and explainable for real-world clinical applications.

Method: Comparative analysis of four representative approaches through comprehensive evaluation on multiple datasets and expert clinician assessment.

Result: YOLO-Stutter and FluentNet provide efficiency but limited transparency; UDM achieves best balance of accuracy and interpretability; SSDM could not be fully reproduced.

Conclusion: The analysis highlights trade-offs among approaches and identifies future directions for clinically viable dysfluency modeling, with UDM showing the most promise for clinical adoption.

Abstract: Recent advances in dysfluency detection have introduced a variety of modeling paradigms, ranging from lightweight object-detection inspired networks (YOLOStutter) to modular interpretable frameworks (UDM). While performance on benchmark datasets continues to improve, clinical adoption requires more than accuracy: models must be controllable and explainable. In this paper, we present a systematic comparative analysis of four representative approaches–YOLO-Stutter, FluentNet, UDM, and SSDM–along three dimensions: performance, controllability, and explainability. Through comprehensive evaluation on multiple datasets and expert clinician assessment, we find that YOLO-Stutter and FluentNet provide efficiency and simplicity, but with limited transparency; UDM achieves the best balance of accuracy and clinical interpretability; and SSDM, while promising, could not be fully reproduced in our experiments. Our analysis highlights the trade-offs among competing approaches and identifies future directions for clinically viable dysfluency modeling. We also provide detailed implementation insights and practical deployment considerations for each approach.

[577] Beyond Memorization: Reasoning-Driven Synthesis as a Mitigation Strategy Against Benchmark Contamination

Terry Jingchen Zhang, Gopal Dev, Ning Wang, Nicole Ni, Wenyuan Jiang, Yinya Huang, Bernhard Schölkopf, Mrinmaya Sachan, Zhijing Jin

Main category: cs.AI

TL;DR: Empirical study shows no significant performance decay near knowledge cutoff dates in LLMs when tested on multi-step reasoning questions synthesized from arXiv papers, suggesting reasoning-driven benchmarks mitigate contamination concerns.

DetailsMotivation: Address growing concerns about data contamination in LLM evaluation, where static benchmarks may measure memorization rather than genuine reasoning capabilities.

Method: Used infinitely scalable framework to synthesize research-level QA from arXiv papers, leveraging temporal structure to detect performance decay after knowledge cutoffs. Evaluated 4 frontier models on 1,643 multi-step reasoning questions from 20,277 papers across 26 months.

Result: Consistent lack of significant performance decay near knowledge cutoff dates across models of various sizes, developers, and release dates. Multi-step reasoning complexity appears to mitigate benchmark contamination.

Conclusion: Advocates for paradigm shift towards reasoning-driven synthesis for benchmark construction rather than periodic collection of new questions, and open-sources code/dataset for reproducibility.

Abstract: Capability evaluation of large language models (LLMs) is increasingly shadowed by rising concerns of data contamination that cast doubts on whether static benchmarks measure genuine reasoning or mere memorization. We present an empirical study using an infinitely scalable framework to synthesize research-level QA directly from arXiv papers, harnessing the natural temporal structure of research publications where performance decay after knowledge cutoffs may indicate potential contamination. We evaluated 4 frontier model represented by 2 models of different knowledge cutoff dates per family on 1,643 multi-step reasoning questions synthesized from 20,277 arXiv papers stratified over 26 months, covering at least 6 months before and after all cutoff dates. Our results consistently showed a lack of significant performance decay near knowledge cutoff dates for models of various sizes, developers, and release dates. We further performed a comparative analysis with previous longitudinal studies that reported significant post-cutoff performance decay using directly retrieved questions based on public data. we hypothesize that the multi-step reasoning required by our synthesis pipeline offered additional complexity that goes deeper than shallow memorization, which effectively serves a mitigation strategy against benchmark contamination. We fully open source our code and dataset to aid reproducibility and advocate for a paradigm shift that prioritize reasoning-driven synthesis to construct benchmarks over simply collecting newly released questions periodically.

[578] Language and Experience: A Computational Model of Social Learning in Complex Tasks

Cédric Colas, Tracey Mills, Ben Prystawski, Michael Henry Tessler, Noah Goodman, Jacob Andreas, Joshua Tenenbaum

Main category: cs.AI

TL;DR: A computational framework for social learning that integrates linguistic guidance with direct experience using joint probabilistic inference over executable world models, enabling both humans and AI to accelerate learning through language-based knowledge transfer.

DetailsMotivation: To understand how humans combine linguistic guidance from others with direct experience for safe and rapid learning, and to develop AI systems that can similarly integrate these knowledge sources for collaborative learning.

Method: Developed a framework that models social learning as joint probabilistic inference over structured, executable world models. Used pretrained language models as probabilistic models of human advice sharing, enabling agents to generate advice and interpret linguistic input during Bayesian inference across 10 video games.

Result: Linguistic guidance shaped exploration and accelerated learning by reducing risky interactions and speeding up key discoveries in both humans and models. Successful knowledge transfer between humans and models was demonstrated through iterated learning experiments.

Conclusion: Structured, language-compatible representations enable effective human-machine collaborative learning, showing how linguistic guidance can be integrated with direct experience to accelerate learning and facilitate knowledge accumulation across generations.

Abstract: The ability to combine linguistic guidance from others with direct experience is central to human development, enabling safe and rapid learning in new environments. How do people integrate these two sources of knowledge, and how might AI systems? We present a computational framework that models social learning as joint probabilistic inference over structured, executable world models given sensorimotor and linguistic data. We make this possible by turning a pretrained language model into a probabilistic model of how humans share advice conditioned on their beliefs, allowing our agents both to generate advice for others and to interpret linguistic input as evidence during Bayesian inference. Using behavioral experiments and simulations across 10 video games, we show how linguistic guidance can shape exploration and accelerate learning by reducing risky interactions and speeding up key discoveries in both humans and models. We further explore how knowledge can accumulate across generations through iterated learning experiments and demonstrate successful knowledge transfer between humans and models – revealing how structured, language-compatible representations might enable human-machine collaborative learning.

[579] Entropy-Guided Loop: Achieving Reasoning through Uncertainty-Aware Generation

Andrew G. A. Correa, Ana C. H de Matos

Main category: cs.AI

TL;DR: Entropy-guided refinement is a lightweight test-time loop that uses token-level uncertainty to trigger targeted refinement passes, achieving 95% of reasoning model quality at 1/3 the cost.

DetailsMotivation: Reasoning models provide better performance but at 3-5x higher cost and latency, creating a need for cost-effective alternatives that maintain quality.

Method: Extracts logprobs, computes Shannon entropy on top-k alternatives, and uses OR-logic trigger (perplexity + max token entropy + low-confidence token count) to trigger refinement. Passes compact uncertainty report back to model for corrective edits.

Result: Small model with this loop achieves 95% of reference reasoning model quality at ~1/3 cost. Selective refinement on ~31% of responses improves accuracy by 16 percentage points over single-pass inference.

Conclusion: The uncertainty-aware loop provides an effective middle ground between single-pass inference and expensive reasoning chains, making it practical for production deployments where both quality and cost matter.

Abstract: Reasoning models often outperform smaller models but at 3–5$\times$ higher cost and added latency. We present entropy-guided refinement: a lightweight, test-time loop that uses token-level uncertainty to trigger a single, targeted refinement pass. We extract logprobs, compute Shannon entropy on top-$k$ alternatives, and apply a simple OR-logic trigger over perplexity, maximum token entropy, and low-confidence-token count. Unlike approaches that use entropy only for measurement or decoding, we pass a compact uncertainty report (tokens, confidences, alternatives, context) back to the model to guide corrective edits. On representative technical queries across reasoning, mathematics, and code generation tasks, a small model with our loop approaches 95% of a reference reasoning model’s quality at approximately one-third of the cost. The method achieves selective refinement on ~31% of responses while improving accuracy by 16 percentage points over single-pass inference. We demonstrate that this uncertainty-aware loop provides an effective middle ground between single-pass inference and expensive reasoning chains, making it practical for production deployments where both quality and cost matter.

[580] Adaptive Monitoring and Real-World Evaluation of Agentic AI Systems

Manish Shukla

Main category: cs.AI

TL;DR: This paper introduces an Adaptive Multi-Dimensional Monitoring (AMDM) algorithm for agentic AI systems, addressing the gap in human-centered evaluation metrics and providing empirical validation through simulations and real-world experiments.

DetailsMotivation: Current evaluations of agentic AI systems predominantly focus on technical metrics (83% of papers) while neglecting human-centered and economic considerations (only 30%). The authors aim to address this imbalance by developing a comprehensive monitoring framework.

Method: The paper formalizes an AMDM algorithm that: 1) normalizes heterogeneous metrics, 2) applies per-axis exponentially weighted moving-average thresholds, and 3) performs joint anomaly detection via Mahalanobis distance. They conduct simulations and real-world experiments to validate the approach.

Result: AMDM significantly improves performance: reduces anomaly-detection latency from 12.3s to 5.6s on simulated goal drift, and cuts false-positive rates from 4.5% to 0.9% compared to static thresholds. The paper includes comparison tables, ROC/PR curves, and case study reanalysis.

Conclusion: The AMDM algorithm provides an effective solution for comprehensive monitoring of agentic AI systems, addressing both technical and human-centered metrics. The authors provide code, data, and reproducibility materials to facilitate adoption and further research.

Abstract: Agentic artificial intelligence (AI) – multi-agent systems that combine large language models with external tools and autonomous planning – are rapidly transitioning from research laboratories into high-stakes domains. Our earlier “Basic” paper introduced a five-axis framework and proposed preliminary metrics such as goal drift and harm reduction but did not provide an algorithmic instantiation or empirical evidence. This “Advanced” sequel fills that gap. First, we revisit recent benchmarks and industrial deployments to show that technical metrics still dominate evaluations: a systematic review of 84 papers from 2023–2025 found that 83% report capability metrics while only 30% consider human-centred or economic axes [2]. Second, we formalise an Adaptive Multi-Dimensional Monitoring (AMDM) algorithm that normalises heterogeneous metrics, applies per-axis exponentially weighted moving-average thresholds and performs joint anomaly detection via the Mahalanobis distance. Third, we conduct simulations and real-world experiments. AMDM cuts anomaly-detection latency from 12.3 s to 5.6 s on simulated goal drift and reduces false-positive rates from 4.5% to 0.9% compared with static thresholds. We present a comparison table and ROC/PR curves, and we reanalyse case studies to surface missing metrics. Code, data and a reproducibility checklist accompany this paper to facilitate replication.

[581] OmniDPO: A Preference Optimization Framework to Address Omni-Modal Hallucination

Junzhe Chen, Tianshu Zhang, Shiyu Huang, Yuwei Niu, Chao Sun, Rongzhou Zhang, Guanyu Zhou, Lijie Wen, Xuming Hu

Main category: cs.AI

TL;DR: OmniDPO is a preference-alignment framework that reduces hallucinations in omni-modal LLMs by enhancing audio-video interaction understanding and strengthening attention to visual/auditory information.

DetailsMotivation: Current OLLMs suffer from hallucination issues where text priors dominate, causing neglect of visual/audio information. Existing models ignore intrinsic correlations between video and audio, leading to hallucinations when interpreting hidden audio cues in video content.

Method: OmniDPO uses two strategies: (1) text-preference sample pairs to enhance understanding of audio-video interactions, and (2) multimodal-preference sample pairs to strengthen attention to visual and auditory information.

Result: Experiments on two OLLMs show OmniDPO effectively mitigates multimodal hallucinations and significantly enhances reasoning capabilities across modalities.

Conclusion: OmniDPO successfully addresses hallucination challenges in OLLMs by improving multimodal grounding through preference alignment strategies.

Abstract: Recently, Omni-modal large language models (OLLMs) have sparked a new wave of research, achieving impressive results in tasks such as audio-video understanding and real-time environment perception. However, hallucination issues still persist. Similar to the bimodal setting, the priors from the text modality tend to dominate, leading OLLMs to rely more heavily on textual cues while neglecting visual and audio information. In addition, fully multimodal scenarios introduce new challenges. Most existing models align visual or auditory modalities with text independently during training, while ignoring the intrinsic correlations between video and its corresponding audio. This oversight results in hallucinations when reasoning requires interpreting hidden audio cues embedded in video content. To address these challenges, we propose OmniDPO, a preference-alignment framework designed to mitigate hallucinations in OLLMs. Specifically, OmniDPO incorporates two strategies: (1) constructing text-preference sample pairs to enhance the model’s understanding of audio-video interactions; and (2) constructing multimodal-preference sample pairs to strengthen the model’s attention to visual and auditory information. By tackling both challenges, OmniDPO effectively improves multimodal grounding and reduces hallucination. Experiments conducted on two OLLMs demonstrate that OmniDPO not only effectively mitigates multimodal hallucinations but also significantly enhances the models’ reasoning capabilities across modalities. All code and datasets will be released upon paper acceptance.

[582] Wrong Face, Wrong Move: The Social Dynamics of Emotion Misperception in Agent-Based Models

David Freire-Obregón

Main category: cs.AI

TL;DR: Study shows that emotion recognition accuracy significantly impacts social behavior - low accuracy leads to emotional disintegration and social disorder, while high accuracy enables emotional resilience and stable clusters.

DetailsMotivation: To understand how perceptual accuracy in emotion detection affects emergent emotional and spatial behavior in social agents, examining the impact of systematic misperception on social dynamics.

Method: Agents with emotion classifiers of varying accuracy (poor/medium/high) placed on 2D toroidal lattice, responding to perceived emotions by moving toward positive and away from negative emotions. Experiments conducted on homogeneous/heterogeneous populations with emotional shocks.

Result: Low-accuracy classifiers reliably caused diminished trust, emotional disintegration into sadness, and disordered social organization. High-accuracy agents developed hardy emotional clusters and resilience to disruptions. Misperception alone generated segregation even in neutral scenarios.

Conclusion: Biases or imprecision in emotion recognition significantly warp social processes and disrupt emotional integration, highlighting the critical importance of accurate emotion perception for stable social organization.

Abstract: The ability of humans to detect and respond to others’ emotions is fundamental to understanding social behavior. Here, agents are instantiated with emotion classifiers of varying accuracy to study the impact of perceptual accuracy on emergent emotional and spatial behavior. Agents are visually represented with face photos from the KDEF database and endowed with one of three classifiers trained on the JAFFE (poor), CK+ (medium), or KDEF (high) datasets. Agents communicate locally on a 2D toroidal lattice, perceiving neighbors’ emotional state based on their classifier and responding with movement toward perceived positive emotions and away from perceived negative emotions. Note that the agents respond to perceived, instead of ground-truth, emotions, introducing systematic misperception and frustration. A battery of experiments is carried out on homogeneous and heterogeneous populations and scenarios with repeated emotional shocks. Results show that low-accuracy classifiers on the part of the agent reliably result in diminished trust, emotional disintegration into sadness, and disordered social organization. By contrast, the agent that develops high accuracy develops hardy emotional clusters and resilience to emotional disruptions. Even in emotionally neutral scenarios, misperception is enough to generate segregation and disintegration of cohesion. These findings underscore the fact that biases or imprecision in emotion recognition may significantly warp social processes and disrupt emotional integration.

[583] Virtual Group Knowledge and Group Belief in Topological Evidence Models (Extended Version)

Alexandru Baltag, Malvin Gattinger, Djanira Gomes

Main category: cs.AI

TL;DR: Axiomatization and decidability results for logics of group knowledge and belief in multi-agent evidence models, including dynamic evidence-sharing extensions.

DetailsMotivation: To extend topological semantics of evidence-based belief and fallible knowledge from individuals to groups, formalizing notions of group knowledge and belief.

Method: Extend topological evidence models to groups, develop logical systems for group evidence, knowledge and belief, and add dynamic evidence-sharing operators.

Result: Complete axiomatization and decidability proofs for logics of group evidence, knowledge, and belief, showing dynamic extensions are co-expressive with static bases.

Conclusion: Successfully formalized group knowledge and belief in evidence models with complete logical systems, demonstrating the feasibility of dynamic evidence-sharing while maintaining expressivity.

Abstract: We study notions of (virtual) group knowledge and group belief within multi-agent evidence models, obtained by extending the topological semantics of evidence-based belief and fallible knowledge from individuals to groups. We completely axiomatize and show the decidability of the logic of (“hard” and “soft”) group evidence, and do the same for an especially interesting fragment of it: the logic of group knowledge and group belief. We also extend these languages with dynamic evidence-sharing operators, and completely axiomatize the corresponding logics, showing that they are co-expressive with their static bases.

[584] Ensemble Debates with Local Large Language Models for AI Alignment

Ephraiem Sarabamoun

Main category: cs.AI

TL;DR: Open-source ensemble debates improve alignment reasoning, outperforming single models with significant gains in reasoning depth and argument quality, especially for truthfulness.

DetailsMotivation: As LLMs are increasingly used in high-stakes decisions, alignment with human values is crucial. Reliance on proprietary APIs limits reproducibility and broad participation in alignment research.

Method: Conducted 150 debates across 15 scenarios using five ensemble configurations of local open-source models, comparing against single-model baselines using a 7-point evaluation rubric.

Result: Ensembles outperformed single models (3.48 vs 3.13 overall), with largest gains in reasoning depth (+19.4%) and argument quality (+34.1%). Strongest improvements in truthfulness (+1.25 points) and human enhancement (+0.80 points).

Conclusion: Local open-source ensemble debates provide an effective, accessible, and reproducible method for improving alignment-oriented reasoning, with code and dataset provided for broader research participation.

Abstract: As large language models (LLMs) take on greater roles in high-stakes decisions, alignment with human values is essential. Reliance on proprietary APIs limits reproducibility and broad participation. We study whether local open-source ensemble debates can improve alignmentoriented reasoning. Across 150 debates spanning 15 scenarios and five ensemble configurations, ensembles outperform single-model baselines on a 7-point rubric (overall: 3.48 vs. 3.13), with the largest gains in reasoning depth (+19.4%) and argument quality (+34.1%). Improvements are strongest for truthfulness (+1.25 points) and human enhancement (+0.80). We provide code, prompts, and a debate data set, providing an accessible and reproducible foundation for ensemble-based alignment evaluation.

[585] HiVA: Self-organized Hierarchical Variable Agent via Goal-driven Semantic-Topological Evolution

Jinzhou Tang, Jusheng Zhang, Qinhan Lv, Sidi Liu, Jing Yang, Chengpei Tang, Keze Wang

Main category: cs.AI

TL;DR: HiVA is a novel framework that models agent workflows as self-organized graphs using Semantic-Topological Evolution algorithm to optimize hybrid spaces with textual gradients, achieving 5-10% accuracy improvements over existing methods.

DetailsMotivation: Existing autonomous agent paradigms face a critical trade-off: fixed workflows require manual reconfiguration for environmental changes, while flexible reactive loops fail to create transferable reasoning structures.

Method: HiVA uses Semantic-Topological Evolution (STEV) algorithm that models workflows as self-organized graphs, optimizes hybrid semantic-topological spaces using textual gradients as discrete-domain surrogates for backpropagation, with iterative process including Multi-Armed Bandit routing, diagnostic gradient generation, and coordinated updates.

Result: Experiments across dialogue, coding, long-context Q&A, mathematical, and agentic benchmarks show 5-10% improvements in task accuracy and enhanced resource efficiency compared to existing baselines.

Conclusion: HiVA establishes effectiveness in autonomous task execution by enabling self-organized workflow optimization that adapts to unknown environments while maintaining transferable reasoning structures.

Abstract: Autonomous agents play a crucial role in advancing Artificial General Intelligence, enabling problem decomposition and tool orchestration through Large Language Models (LLMs). However, existing paradigms face a critical trade-off. On one hand, reusable fixed workflows require manual reconfiguration upon environmental changes; on the other hand, flexible reactive loops fail to distill reasoning progress into transferable structures. We introduce Hierarchical Variable Agent (HiVA), a novel framework modeling agentic workflows as self-organized graphs with the Semantic-Topological Evolution (STEV) algorithm, which optimizes hybrid semantic-topological spaces using textual gradients as discrete-domain surrogates for backpropagation. The iterative process comprises Multi-Armed Bandit-infused forward routing, diagnostic gradient generation from environmental feedback, and coordinated updates that co-evolve individual semantics and topology for collective optimization in unknown environments. Experiments on dialogue, coding, Long-context Q&A, mathematical, and agentic benchmarks demonstrate improvements of 5-10% in task accuracy and enhanced resource efficiency over existing baselines, establishing HiVA’s effectiveness in autonomous task execution.

[586] MODE: Mixture of Document Experts for RAG

Rahul Anand

Main category: cs.AI

TL;DR: MODE replaces traditional dense retrieval with cluster-and-route approach for small domain-specific collections, eliminating vector databases while maintaining accuracy with faster retrieval.

DetailsMotivation: Traditional RAG systems with large vector databases and cross-encoders are excessive for small, domain-specific document collections, creating unnecessary complexity and infrastructure requirements.

Method: Documents are embedded, grouped into semantically coherent clusters, and represented by cached centroids. At query time, routing goes to top centroid(s) and retrieves context only within those clusters.

Result: On HotpotQA and SQuAD corpora (100-500 chunks), MODE matches or exceeds dense-retrieval baseline in answer quality while reducing end-to-end retrieval time. Tighter clusters improve downstream accuracy.

Conclusion: MODE provides a practical solution for small and medium corpora where simplicity, speed, and topical focus are important, eliminating external vector-database infrastructure while maintaining performance.

Abstract: Retrieval-Augmented Generation (RAG) often relies on large vector databases and cross-encoders tuned for large-scale corpora, which can be excessive for small, domain-specific collections. We present MODE (Mixture of Document Experts), a lightweight alternative that replaces fine-grained nearest-neighbor search with cluster-and-route retrieval. Documents are embedded, grouped into semantically coherent clusters, and represented by cached centroids. At query time, we route to the top centroid(s) and retrieve context only within those clusters, eliminating external vector-database infrastructure and reranking while keeping latency low. On HotpotQA and SQuAD corpora with 100-500 chunks, MODE matches or exceeds a dense-retrieval baseline in answer quality while reducing end-to-end retrieval time. Ablations show that cluster granularity and multi-cluster routing control the recall/precision trade-off, and that tighter clusters improve downstream accuracy. MODE offers a practical recipe for small and medium corpora where simplicity, speed, and topical focus matter.

[587] Know When to Explore: Difficulty-Aware Certainty as a Guide for LLM Reinforcement Learning

Ang Li, Zhihang Yuan, Yang Zhang, Shouda Liu, Yisen Wang

Main category: cs.AI

TL;DR: DACE is a novel RL algorithm that uses LLM self-certainty to dynamically balance exploration-exploitation trade-off based on task difficulty, improving mathematical reasoning performance.

DetailsMotivation: Traditional RLVF relies on sparse outcome-based rewards that fail to provide granular guidance on reasoning processes, hindering efficient learning from different failure types.

Method: DACE assesses task difficulty online based on policy success rates and modulates intrinsic rewards: penalizing high certainty for difficult tasks to encourage exploration, and rewarding high certainty for easier tasks to promote learning efficiency.

Result: Experiments on AIME and MATH benchmarks show DACE significantly outperforms baselines, achieving higher accuracy and more robust performance with test-time compute scaling.

Conclusion: DACE’s adaptive approach effectively fosters exploration without sacrificing precision, demonstrating that leveraging LLM self-certainty based on task difficulty improves reinforcement learning for reasoning tasks.

Abstract: Reinforcement Learning with Verifiable Feedback (RLVF) has become a key technique for enhancing the reasoning abilities of Large Language Models (LLMs). However, its reliance on sparse, outcome based rewards, which only indicate if a final answer is correct or not, fails to provide granular guidance on the reasoning process itself. This limitation hinders efficient learning, as the model cannot distinguish between high quality and inefficient solutions, nor can it learn effectively from different types of failures. To address this, we observe that an LLMs self-certainty often correlates with task difficulty and solution quality. We introduce Difficulty Aware Certainty guided Exploration (DACE), a novel RL algorithm that leverages this insight to dynamically balance the exploration exploitation trade-off. DACE assesses task difficulty online based on the policys success rate. It then uses this signal to modulate an intrinsic reward: for difficult tasks where the model is struggling, DACE encourages exploration by penalizing high certainty; for easier tasks, it encourages learning efficiency by rewarding high certainty. Experiments on challenging mathematical reasoning benchmarks (AIME, MATH) show that DACE significantly outperforms strong baselines. The DACE-trained models not only achieve higher accuracy but also demonstrate more robust performance when scaling test-time compute, validating that our adaptive approach fosters effective exploration without sacrificing precision.

[588] Optimizing Health Coverage in Ethiopia: A Learning-augmented Approach and Persistent Proportionality Under an Online Budget

Davin Choo, Yohai Trabelsi, Fentabil Getnet, Samson Warkaye Lamma, Wondesen Nigatu, Kasahun Sime, Lisa Matay, Milind Tambe, Stéphane Verguet

Main category: cs.AI

TL;DR: Developed HARP optimization tool for prioritizing health facility planning in Ethiopia to maximize population coverage under budget constraints while meeting regional proportionality targets.

DetailsMotivation: Ethiopia needs to strategically prioritize limited health system strengthening resources across regions to expand healthcare access in alignment with UN Sustainable Development Goals.

Method: Created Health Access Resource Planner (HARP) with two algorithms: learning-augmented approach for single-step improvement over expert recommendations, and greedy algorithm for multi-step planning with worst-case approximation guarantees.

Result: Demonstrated empirical efficacy through collaboration with Ethiopian Public Health Institute and Ministry of Health across three regions and various planning scenarios.

Conclusion: HARP provides a principled optimization framework for sequential health facility planning that can effectively guide resource allocation decisions under budget uncertainty while ensuring regional equity.

Abstract: As part of nationwide efforts aligned with the United Nations’ Sustainable Development Goal 3 on Universal Health Coverage, Ethiopia’s Ministry of Health is strengthening health posts to expand access to essential healthcare services. However, only a fraction of this health system strengthening effort can be implemented each year due to limited budgets and other competing priorities, thus the need for an optimization framework to guide prioritization across the regions of Ethiopia. In this paper, we develop a tool, Health Access Resource Planner (HARP), based on a principled decision-support optimization framework for sequential facility planning that aims to maximize population coverage under budget uncertainty while satisfying region-specific proportionality targets at every time step. We then propose two algorithms: (i) a learning-augmented approach that improves upon expert recommendations at any single-step; and (ii) a greedy algorithm for multi-step planning, both with strong worst-case approximation estimation. In collaboration with the Ethiopian Public Health Institute and Ministry of Health, we demonstrated the empirical efficacy of our method on three regions across various planning scenarios.

[589] Universal Deep Research: Bring Your Own Model and Strategy

Peter Belcak, Pavlo Molchanov

Main category: cs.AI

TL;DR: Universal Deep Research (UDR) is a generalist agentic system that allows users to create custom deep research strategies without training, overcoming the limitation of hard-coded research agents.

DetailsMotivation: Current deep research agents are hard-coded with fixed strategies and tools, limiting flexibility and requiring users to adapt to predefined approaches rather than creating their own research methodologies.

Method: UDR wraps around any language model and provides a system for users to create, edit, and refine custom deep research strategies. It includes example strategies (minimal, expansive, intensive) and a user interface for experimentation.

Result: The system demonstrates generality by supporting diverse research strategies without additional training or finetuning, enabling flexible and customizable deep research workflows.

Conclusion: UDR provides a flexible framework for deep research that empowers users to define their own research strategies rather than being constrained by hard-coded approaches, making agentic research systems more accessible and adaptable.

Abstract: Deep research tools are among the most impactful and most commonly encountered agentic systems today. We observe, however, that each deep research agent introduced so far is hard-coded to carry out a particular research strategy using a fixed choice of tools. We introduce Universal Deep Research (UDR), a generalist agentic system that wraps around any language model and enables the user to create, edit, and refine their own entirely custom deep research strategies without any need for additional training or finetuning. To showcase the generality of our system, we equip UDR with example minimal, expansive, and intensive research strategies, and provide a user interface to facilitate experimentation with the system.

[590] Instruction-Level Weight Shaping: A Framework for Self-Improving AI Agents

Rimom Costa

Main category: cs.AI

TL;DR: ILWS is a method that uses system instructions as external pseudo-parameters updated via reflection and user feedback, enabling continuous LLM improvement without retrieval or fine-tuning, achieving 2.4-5x throughput gains and 80% fewer hallucinations.

DetailsMotivation: Traditional approaches like RAG increase latency and engineering overhead, prompt engineering is brittle, and fine-tuning risks catastrophic forgetting while being costly. There's a need for a method that allows LLMs to adapt to new knowledge without these drawbacks.

Method: Instruction-Level Weight Shaping (ILWS) uses curated system instructions as external pseudo-parameters updated after each session. A Reflection Engine analyzes conversations, diagnoses reasoning, and proposes typed deltas (ΔS, ΔU, ΔT) for instructions, user preferences, and tools. These deltas are version-controlled, evaluated with ratings, and distilled into model parameters when an edit budget threshold is crossed.

Result: In enterprise support, ILWS increased throughput 2.4-5.0x and reduced audited hallucinations by about 80% compared to frozen baseline. In Adobe Commerce Cloud proof of concept, it achieved 4-5x more tickets per hour and about 80% lower time per ticket with autonomous instruction updates.

Conclusion: ILWS provides an effective approach for continuous LLM improvement that preserves governance, removes per-call retrieval, and generalizes to dynamic domains requiring adaptive reasoning and low-latency deployment, converting prompt-space improvements into weight-space without downtime.

Abstract: Large language models (LLMs) are fluent but largely static after pre-training; new or shifting knowledge is typically added with retrieval-augmented generation (RAG) or fine-tuning. RAG raises latency and engineering overhead and often fails to integrate facts; prompt engineering is brittle and can conflict with prior knowledge; fine-tuning is costly and risks catastrophic forgetting. We propose Instruction-Level Weight Shaping (ILWS): curated system instructions act as external, auditable pseudo-parameters updated after each session via reflection and user feedback. A Reflection Engine inspects conversation traces, diagnoses reasoning successes and failures, and proposes typed deltas $\Delta K=(\Delta S,\Delta U,\Delta T)$ over instructions, user preferences, and tools. Deltas are version-controlled, evaluated with a sliding window of 1-5 star ratings, auto-repaired on first failure, and rolled back on repeated failure. When an edit budget crosses a threshold, the agent compiles a rating-weighted synthetic set and distills matured instruction-space gains into parameters, converting prompt-space improvements into weight-space without downtime. ILWS makes explicit the low-rank shaping induced by context in transformer blocks, preserves governance, and removes per-call retrieval. In enterprise support it increased throughput 2.4-5.0x and cut audited hallucinations by about 80% versus a frozen baseline. In an Adobe Commerce Cloud proof of concept “L0 Support”, it achieved 4-5x more tickets per hour and about 80% lower time per ticket, with autonomous instruction updates and optional tool synthesis. Because ILWS operates at the instruction layer until controlled distillation, it generalizes to dynamic domains (legal, medical, engineering) requiring adaptive reasoning, tool creation, and low-latency deployment.

[591] Symbolic Planning and Multi-Agent Path Finding in Extremely Dense Environments with Movable Obstacles

Bo Fu, Zhe Chen, Rahul Chandan, Alex Barbosa, Michael Caldara, Joey Durham, Federico Pecora

Main category: cs.AI

TL;DR: The paper introduces the Block Rearrangement Problem (BRaP) for warehouse management and proposes five search-based algorithms to solve it efficiently in large grids.

DetailsMotivation: Addressing the challenging problem of rearranging storage blocks in dense warehouse grids to achieve target configurations, which is a critical component of large warehouse management systems.

Method: Formally define BRaP as a graph search problem and propose five search-based solution algorithms using joint configuration space search, classical planning, multi-agent pathfinding, and expert heuristics.

Result: The methods demonstrate efficiency in creating rearrangement plans for deeply buried blocks in up to 80x80 grids, despite the exponential relationship between search space size and block number.

Conclusion: The proposed search-based approaches effectively solve the Block Rearrangement Problem and show scalability for large warehouse grid configurations.

Abstract: We introduce the Block Rearrangement Problem (BRaP), a challenging component of large warehouse management which involves rearranging storage blocks within dense grids to achieve a target state. We formally define the BRaP as a graph search problem. Building on intuitions from sliding puzzle problems, we propose five search-based solution algorithms, leveraging joint configuration space search, classical planning, multi-agent pathfinding, and expert heuristics. We evaluate the five approaches empirically for plan quality and scalability. Despite the exponential relation between search space size and block number, our methods demonstrate efficiency in creating rearrangement plans for deeply buried blocks in up to 80x80 grids.

[592] SHERPA: A Model-Driven Framework for Large Language Model Execution

Boqi Chen, Kua Chen, José Antonio Hernández López, Gunter Mussbacher, Dániel Varró, Amir Feizpour

Main category: cs.AI

TL;DR: SHERPA is a model-driven framework that uses hierarchical state machines to incorporate domain-specific best practices into LLM execution, improving performance on complex tasks like code generation and question answering.

DetailsMotivation: LLMs lack structured reasoning ability for complex tasks requiring domain-specific best practices that aren't available in training data. Existing multi-step prompting methods lack general control mechanisms.

Method: Proposes SHERPA framework that structures LLM execution using hierarchical state machines, enabling fine-grained control via rules or ML-based decisions including LLMs themselves.

Result: SHERPA significantly improves LLM output quality across code generation, class name generation, and question answering tasks, particularly beneficial for complex tasks with established human best practices but limited training data.

Conclusion: Integrating well-designed state machines provides an effective mechanism to control LLM behavior and incorporate domain expertise, demonstrating substantial performance improvements on various complex tasks.

Abstract: Recently, large language models (LLMs) have achieved widespread application across various fields. Despite their impressive capabilities, LLMs suffer from a lack of structured reasoning ability, particularly for complex tasks requiring domain-specific best practices, which are often unavailable in the training data. Although multi-step prompting methods incorporating human best practices, such as chain-of-thought and tree-of-thought, have gained popularity, they lack a general mechanism to control LLM behavior. In this paper, we propose SHERPA, a model-driven framework to improve the LLM performance on complex tasks by explicitly incorporating domain-specific best practices into hierarchical state machines. By structuring the LLM execution processes using state machines, SHERPA enables more fine-grained control over their behavior via rules or decisions driven by machine learning-based approaches, including LLMs. We show that SHERPA is applicable to a wide variety of tasks-specifically, code generation, class name generation, and question answering-replicating previously proposed approaches while further improving the performance. We demonstrate the effectiveness of SHERPA for the aforementioned tasks using various LLMs. Our systematic evaluation compares different state machine configurations against baseline approaches without state machines. Results show that integrating well-designed state machines significantly improves the quality of LLM outputs, and is particularly beneficial for complex tasks with well-established human best practices but lacking data used for training LLMs.

[593] SIGMUS: Semantic Integration for Knowledge Graphs in Multimodal Urban Spaces

Brian Wang, Mani Srivastava

Main category: cs.AI

TL;DR: SIGMUS system uses LLMs to automatically integrate multimodal urban sensor data and identify relationships with urban incidents through knowledge graphs, eliminating need for human-defined rules.

DetailsMotivation: Urban spaces have abundant multimodal sensor data but it's fragmented and difficult to integrate manually for incident analysis and forecasting.

Method: SIGMUS uses Large Language Models to generate world knowledge for identifying relationships between urban incidents and multimodal data, organizing evidence into knowledge graphs without human-encoded rules.

Result: The system successfully connects 5 different data sources (news articles, CCTV images, air quality, weather, traffic) with relevant incidents occurring at same time and location.

Conclusion: LLM-based semantic integration enables automated reasoning about urban incidents from fragmented multimodal data sources, providing scalable incident analysis without manual rule engineering.

Abstract: Modern urban spaces are equipped with an increasingly diverse set of sensors, all producing an abundance of multimodal data. Such multimodal data can be used to identify and reason about important incidents occurring in urban landscapes, such as major emergencies, cultural and social events, as well as natural disasters. However, such data may be fragmented over several sources and difficult to integrate due to the reliance on human-driven reasoning for identifying relationships between the multimodal data corresponding to an incident, as well as understanding the different components which define an incident. Such relationships and components are critical to identifying the causes of such incidents, as well as producing forecasting the scale and intensity of future incidents as they begin to develop. In this work, we create SIGMUS, a system for Semantic Integration for Knowledge Graphs in Multimodal Urban Spaces. SIGMUS uses Large Language Models (LLMs) to produce the necessary world knowledge for identifying relationships between incidents occurring in urban spaces and data from different modalities, allowing us to organize evidence and observations relevant to an incident without relying and human-encoded rules for relating multimodal sensory data with incidents. This organized knowledge is represented as a knowledge graph, organizing incidents, observations, and much more. We find that our system is able to produce reasonable connections between 5 different data sources (new article text, CCTV images, air quality, weather, and traffic measurements) and relevant incidents occurring at the same time and location.

[594] Question-to-Knowledge: Multi-Agent Generation of Inspectable Facts for Product Mapping

Wonduk Seo, Taesub Shin, Hyunjin An, Dokyun Kim, Seunghyun Lee

Main category: cs.AI

TL;DR: Q2K is a multi-agent LLM framework that improves SKU mapping accuracy by generating targeted questions, conducting focused web searches, and reusing validated reasoning to reduce redundancy.

DetailsMotivation: Traditional rule-based heuristics and keyword similarity methods often fail at SKU mapping due to subtle product variations in brand, specifications, and bundle configurations when explicit identifiers are missing.

Method: A three-agent framework: Reasoning Agent generates disambiguation questions, Knowledge Agent resolves them via focused web searches, and Deduplication Agent reuses validated reasoning traces to reduce redundancy. Includes human-in-the-loop for uncertain cases.

Result: Outperforms strong baselines on real-world consumer goods datasets, achieving higher accuracy and robustness in challenging scenarios like bundle identification and brand origin disambiguation.

Conclusion: Q2K provides a scalable and interpretable solution for product integration by balancing accuracy with efficiency through reasoning reuse rather than repeated searches.

Abstract: Identifying whether two product listings refer to the same Stock Keeping Unit (SKU) is a persistent challenge in ecommerce, especially when explicit identifiers are missing and product names vary widely across platforms. Rule based heuristics and keyword similarity often misclassify products by overlooking subtle distinctions in brand, specification, or bundle configuration. To overcome these limitations, we propose Question to Knowledge (Q2K), a multi agent framework that leverages Large Language Models (LLMs) for reliable SKU mapping. Q2K integrates: (1) a Reasoning Agent that generates targeted disambiguation questions, (2) a Knowledge Agent that resolves them via focused web searches, and (3) a Deduplication Agent that reuses validated reasoning traces to reduce redundancy and ensure consistency. A human in the loop mechanism further refines uncertain cases. Experiments on real world consumer goods datasets show that Q2K surpasses strong baselines, achieving higher accuracy and robustness in difficult scenarios such as bundle identification and brand origin disambiguation. By reusing retrieved reasoning instead of issuing repeated searches, Q2K balances accuracy with efficiency, offering a scalable and interpretable solution for product integration.

[595] NEWSAGENT: Benchmarking Multimodal Agents as Journalists with Real-World Newswriting Tasks

Yen-Che Chien, Kuang-Da Wang, Wei-Yao Wang, Wen-Chih Peng

Main category: cs.AI

TL;DR: NEWSAGENT benchmark evaluates autonomous agents’ ability to perform journalistic tasks like content search, information selection, and article generation from multimodal web data, showing current agents struggle with planning and narrative integration.

DetailsMotivation: To assess how well agent-based systems can improve multimodal web data productivity in journalism, which requires iterative planning, interpretation, and contextual reasoning from raw contents to form structured news.

Method: Introduces NEWSAGENT benchmark with 6k human-verified examples from real news, converted to text for broad compatibility. Evaluates open- and closed-source LLMs with agentic frameworks on tasks like narrative perspective identification, keyword querying, historical background retrieval, and article generation.

Result: Agents demonstrate capability in retrieving relevant facts but struggle significantly with planning and narrative integration tasks that require active information discovery and contextual reasoning.

Conclusion: NEWSAGENT provides a realistic testbed for evaluating agent capabilities in multimodal web data manipulation for real-world productivity applications, highlighting current limitations in complex reasoning and planning tasks.

Abstract: Recent advances in autonomous digital agents from industry (e.g., Manus AI and Gemini’s research mode) highlight potential for structured tasks by autonomous decision-making and task decomposition; however, it remains unclear to what extent the agent-based systems can improve multimodal web data productivity. We study this in the realm of journalism, which requires iterative planning, interpretation, and contextual reasoning from multimodal raw contents to form a well structured news. We introduce NEWSAGENT, a benchmark for evaluating how agents can automatically search available raw contents, select desired information, and edit and rephrase to form a news article by accessing core journalistic functions. Given a writing instruction and firsthand data as how a journalist initiates a news draft, agents are tasked to identify narrative perspectives, issue keyword-based queries, retrieve historical background, and generate complete articles. Unlike typical summarization or retrieval tasks, essential context is not directly available and must be actively discovered, reflecting the information gaps faced in real-world news writing. NEWSAGENT includes 6k human-verified examples derived from real news, with multimodal contents converted to text for broad model compatibility. We evaluate open- and closed-sourced LLMs with commonly-used agentic frameworks on NEWSAGENT, which shows that agents are capable of retrieving relevant facts but struggling with planning and narrative integration. We believe that NEWSAGENT serves a realistic testbed for iterating and evaluating agent capabilities in terms of multimodal web data manipulation to real-world productivity.

[596] Multi-Agent Data Visualization and Narrative Generation

Anton Wolter, Georgios Vidalakis, Michael Yu, Ankit Grover, Vaishali Dhanoa

Main category: cs.AI

TL;DR: A lightweight multi-agent system that automates data analysis workflow from exploration to visual narrative generation, using hybrid architecture with deterministic components for improved transparency and reliability.

DetailsMotivation: To enable greater automation and human-AI collaboration in data visualization by employing multi-agent systems throughout the data-to-communication pipeline.

Method: Combines hybrid multi-agent architecture with deterministic components, strategically externalizing critical logic from LLMs to improve transparency and reliability. Delivers granular, modular outputs for surgical modifications.

Result: Evaluated across 4 diverse datasets, demonstrating strong generalizability, narrative quality, and computational efficiency with minimal dependencies.

Conclusion: The system successfully automates data analysis workflow and generates coherent visual narratives while supporting sustainable human-AI collaboration through transparent and reliable architecture.

Abstract: Recent advancements in the field of AI agents have impacted the way we work, enabling greater automation and collaboration between humans and agents. In the data visualization field, multi-agent systems can be useful for employing agents throughout the entire data-to-communication pipeline. We present a lightweight multi-agent system that automates the data analysis workflow, from data exploration to generating coherent visual narratives for insight communication. Our approach combines a hybrid multi-agent architecture with deterministic components, strategically externalizing critical logic from LLMs to improve transparency and reliability. The system delivers granular, modular outputs that enable surgical modifications without full regeneration, supporting sustainable human-AI collaboration. We evaluated our system across 4 diverse datasets, demonstrating strong generalizability, narrative quality, and computational efficiency with minimal dependencies.

[597] Towards Agentic OS: An LLM Agent Framework for Linux Schedulers

Yusheng Zheng, Yanpeng Hu, Wei Zhang, Andi Quinn

Main category: cs.AI

TL;DR: SchedCP is an autonomous LLM agent framework that optimizes Linux schedulers by separating AI reasoning from system execution, achieving up to 1.79x performance improvement and 13x cost reduction.

DetailsMotivation: Operating system schedulers suffer from a semantic gap where kernel policies fail to understand application-specific needs, leading to suboptimal performance.

Method: Architects a decoupled control plane with MCP server providing Workload Analysis Engine, Scheduler Policy Repository, and Execution Verifier. Uses multi-agent system to analyze workloads and synthesize custom eBPF scheduling policies.

Result: Achieves up to 1.79x performance improvement and 13x cost reduction compared to naive agentic approaches while maintaining high success rate.

Conclusion: Bridges the semantic gap, democratizes expert-level system optimization, and represents a step towards creating truly self-optimizing, application-aware operating systems.

Abstract: Operating system schedulers suffer from a fundamental semantic gap, where kernel policies fail to understand application-specific needs, leading to suboptimal performance. We introduce SchedCP, the first framework that enables fully autonomous Large Language Model (LLM) agents to safely and efficiently optimize Linux schedulers without human involvement. Our core insight is that the challenge is not merely to apply a better LLM, but to architect a decoupled control plane that separates the AI’s role of semantic reasoning (“what to optimize”) from the system’s role of execution (“how to observe and act”). Implemented as Model Context Protocol(MCP) server, SchedCP provides a stable interface with three key services: a Workload Analysis Engine, an evolving Scheduler Policy Repository, and an Execution Verifier that validates all AI-generated code and configure before deployment with static and dynamic analysis. We demonstrate this architecture’s power with sched-agent, a multi-agent system that autonomously analyzes workloads, synthesizes custom eBPF scheduling policies, and deploys them via the sched_ext infrastructure. Our evaluation shows that SchedCP achieves up to an 1.79x performance improvement, and a 13x cost reduction compared to naive agentic approaches, all while maintaining high success rate. By bridging the semantic gap, SchedCP democratizes expert-level system optimization and represents a step towards creating truly self-optimizing, application-aware operating systems. The code is open-sourced in https://github.com/eunomia-bpf/schedcp

[598] Artificial Intelligence-Based Analysis of Ice Cream Melting Behavior Under Various Ingredients

Zhang Lai Bin, Zhen Bin It

Main category: cs.AI

TL;DR: Study examines how locust bean gum, guar gum, maltodextrin, and carrageenan affect ice cream melting behavior and identifies cost-effective stabilizer formulations.

DetailsMotivation: Ice cream stability during melting is crucial for consumer acceptance and product quality. The research aims to understand how different stabilizers influence melting resistance and find more economical recipe formulations.

Method: Prepared ice cream samples with each additive, conducted melting tests under controlled conditions, used timelapse recordings to track melting progression, and employed Python and OpenCV for image processing and analysis.

Result: All samples maintained foam-like structure after melting, indicating stable air-cell matrix formation. Re-frozen and re-melted samples showed increased sturdiness. Different stabilizers varied in effectiveness, with some providing stronger melting resistance and structural support than others.

Conclusion: The study provides insights into functional roles of food additives in ice cream, demonstrating potential for developing recipes that balance durability with cost efficiency for both small-scale and commercial production.

Abstract: The stability of ice cream during melting is a critical factor for consumer’s acceptance and product quality. With the commonly added stabilizer to improve texture, structure and slower melting as the factors to analyze. This report explores the effects of locust bean gum, guar gum, maltodextrin, and carrageenan on the melting behavior of homemade ice cream. The main objective was to assess how these additives influence melting resistance and to identify a more cost-effective recipe formulation. Ice cream samples incorporating each additive were prepared and subjected to melting tests under controlled conditions. Timelapse recordings were used to capture and analyze the progression of melting over time. Python and OpenCV is used for process and analysis. Observations revealed that all samples retained a foam-like structure even after melting, suggesting the stabilizers contributed to the formation of a stable air-cell matrix. Furthermore, when the melted samples were re-frozen and subsequently melted again, they displayed increased sturdiness, indicating improved resilience of the ice cream structure. Comparative analysis of the different stabilizers highlighted variations in their effectiveness, with some offering stronger melting resistance and structural support than others. Overall, the findings provide insights into the functional roles of commonly used food additives in ice cream formulation. By evaluating both performance and cost, this study demonstrates the potential for developing recipes that balance durability with economic efficiency, contributing to practical applications in both small-scale and commercial ice cream production.

[599] LLM-empowered Agents Simulation Framework for Scenario Generation in Service Ecosystem Governance

Deyu Zhou, Yuqi Hou, Xiao Xue, Xudong Lu, Qingzhong Li, Lizhen Cui

Main category: cs.AI

TL;DR: A method using three LLM-powered agents to generate high-quality scenarios for service ecosystem governance, overcoming limitations of traditional rule-based scenario analysis.

DetailsMotivation: Traditional scenario analysis methods for service ecosystem governance rely on predefined rules and face challenges like limited information, numerous influencing factors, and difficulty measuring social elements, which limit scenario quality and efficiency.

Method: Proposes a scenario generator design method with three coordinated LLM agents: Environment Agent (EA) generates social environments including extremes, Social Agent (SA) generates social collaboration structures, and Planner Agent (PA) couples task-role relationships and plans solutions while adjusting schemes in real-time.

Result: Experiments on ProgrammableWeb dataset show the method generates more accurate scenarios more efficiently compared to traditional approaches.

Conclusion: The method provides an innovative and effective approach for constructing experimental systems for service ecosystem governance, enabling better scenario generation through adaptive coordination of multiple AI agents.

Abstract: As the social environment is growing more complex and collaboration is deepening, factors affecting the healthy development of service ecosystem are constantly changing and diverse, making its governance a crucial research issue. Applying the scenario analysis method and conducting scenario rehearsals by constructing an experimental system before managers make decisions, losses caused by wrong decisions can be largely avoided. However, it relies on predefined rules to construct scenarios and faces challenges such as limited information, a large number of influencing factors, and the difficulty of measuring social elements. These challenges limit the quality and efficiency of generating social and uncertain scenarios for the service ecosystem. Therefore, we propose a scenario generator design method, which adaptively coordinates three Large Language Model (LLM) empowered agents that autonomously optimize experimental schemes to construct an experimental system and generate high quality scenarios. Specifically, the Environment Agent (EA) generates social environment including extremes, the Social Agent (SA) generates social collaboration structure, and the Planner Agent (PA) couples task-role relationships and plans task solutions. These agents work in coordination, with the PA adjusting the experimental scheme in real time by perceiving the states of each agent and these generating scenarios. Experiments on the ProgrammableWeb dataset illustrate our method generates more accurate scenarios more efficiently, and innovatively provides an effective way for service ecosystem governance related experimental system construction.

[600] LLM-Assisted Iterative Evolution with Swarm Intelligence Toward SuperBrain

Li Weigang, Pedro Carvalho Brom, Lucas Ramson Siefert

Main category: cs.AI

TL;DR: SuperBrain framework enables collective intelligence through co-evolution of LLMs and humans, progressing from individual cognitive dyads to emergent meta-intelligence via GA-assisted evolution and swarm coordination.

DetailsMotivation: To move beyond static prompt engineering and isolated agent simulations by creating a dynamic, scalable collective intelligence system that evolves through human-LLM interaction.

Method: Four-stage approach: 1) Subclass Brain formation through personalized user-LLM interactions, 2) GA-assisted forward-backward evolution for prompt refinement, 3) Swarm Intelligence coordination of multiple Subclass Brains, 4) Integration into Superclass Brain meta-intelligence.

Result: Initial implementations demonstrated in UAV scheduling and keyword filtering tasks, with proposed registry for cross-dyad knowledge consolidation.

Conclusion: Provides conceptual foundation and architectural roadmap for scalable, explainable, and ethically aligned collective AI systems through human-LLM co-evolution.

Abstract: We propose a novel SuperBrain framework for collective intelligence, grounded in the co-evolution of large language models (LLMs) and human users. Unlike static prompt engineering or isolated agent simulations, our approach emphasizes a dynamic pathway from Subclass Brain to Superclass Brain: (1) A Subclass Brain arises from persistent, personalized interaction between a user and an LLM, forming a cognitive dyad with adaptive learning memory. (2) Through GA-assisted forward-backward evolution, these dyads iteratively refine prompts and task performance. (3) Multiple Subclass Brains coordinate via Swarm Intelligence, optimizing across multi-objective fitness landscapes and exchanging distilled heuristics. (4) Their standardized behaviors and cognitive signatures integrate into a Superclass Brain, an emergent meta-intelligence capable of abstraction, generalization and self-improvement. We outline the theoretical constructs, present initial implementations (e.g., UAV scheduling, KU/KI keyword filtering) and propose a registry for cross-dyad knowledge consolidation. This work provides both a conceptual foundation and an architectural roadmap toward scalable, explainable and ethically aligned collective AI.

[601] Text-to-Layout: A Generative Workflow for Drafting Architectural Floor Plans Using LLMs

Jayakrishna Duggempudi, Lu Gao, Ahmed Senouci, Zhe Han, Yunpeng Zhang

Main category: cs.AI

TL;DR: AI workflow using LLMs to generate architectural floor plans from text prompts, producing Revit-compatible designs with minimal manual effort.

DetailsMotivation: To automate the initial drafting phase of architectural design by converting natural language descriptions into structured floor plans, reducing manual effort and enabling rapid prototyping.

Method: Combines prompt engineering with LLMs, furniture placement refinement algorithms, and Python scripting to interpret text inputs and generate spatially coherent layouts including walls, doors, windows, and furniture arrangements.

Result: Successfully generates functional residential layouts that preserve Revit-native parametric attributes for direct BIM integration, demonstrated through a case study of a mid-sized residential layout.

Conclusion: The workflow provides an efficient method for automated architectural drafting with transparent replication capabilities, enabling other researchers to implement similar systems and integrate directly into professional BIM workflows.

Abstract: This paper presents the development of an AI-powered workflow that uses Large Language Models (LLMs) to assist in drafting schematic architectural floor plans from natural language prompts. The proposed system interprets textual input to automatically generate layout options including walls, doors, windows, and furniture arrangements. It combines prompt engineering, a furniture placement refinement algorithm, and Python scripting to produce spatially coherent draft plans compatible with design tools such as Autodesk Revit. A case study of a mid-sized residential layout demonstrates the approach’s ability to generate functional and structured outputs with minimal manual effort. The workflow is designed for transparent replication, with all key prompt specifications documented to enable independent implementation by other researchers. In addition, the generated models preserve the full range of Revit-native parametric attributes required for direct integration into professional BIM processes.

[602] Social World Models

Xuhui Zhou, Jiarui Liu, Akhila Yerukola, Hyunwoo Kim, Maarten Sap

Main category: cs.AI

TL;DR: S3AP is a novel structured social world representation that helps AI systems better understand and reason about social dynamics, achieving significant improvements in social reasoning tasks and social interaction benchmarks.

DetailsMotivation: Humans naturally navigate social interactions by simulating unspoken dynamics and reasoning about others' perspectives, but AI systems struggle with automatically structuring and reasoning about implicit social contexts.

Method: A POMDP-driven design that represents social interactions as structured tuples (state, observation, agent actions, mental states) which can be automatically induced from free-form narratives or other inputs.

Result: +51% improvement on FANToM’s theory-of-mind reasoning with OpenAI’s o1, reaching new SOTA performance; +18% improvement on SOTOPIA social interaction benchmark; ability to predict future social dynamics and improve agent decision-making.

Conclusion: S3AP shows promise as a powerful, general-purpose representation for social world states, enabling development of more socially-aware systems that better navigate social interactions.

Abstract: Humans intuitively navigate social interactions by simulating unspoken dynamics and reasoning about others’ perspectives, even with limited information. In contrast, AI systems struggle to automatically structure and reason about these implicit social contexts. In this paper, we introduce a novel structured social world representation formalism (S3AP), designed to help AI systems reason more effectively about social dynamics. Following a POMDP-driven design, S3AP represents social interactions as structured tuples, such as state, observation, agent actions, and mental states, which can be automatically induced from free-form narratives or other inputs. We first show S3AP can help LLMs better understand social narratives across 5 social reasoning tasks (e.g., +51% improvement on FANToM’s theory-of-mind reasoning with OpenAI’s o1), reaching new state-of-the-art (SOTA) performance. We then induce social world models from these structured representations, demonstrating their ability to predict future social dynamics and improve agent decision-making, yielding up to +18% improvement on the SOTOPIA social interaction benchmark. Our findings highlight the promise of S3AP as a powerful, general-purpose representation for social world states, enabling the development of more socially-aware systems that better navigate social interactions.

[603] How Real Is AI Tutoring? Comparing Simulated and Human Dialogues in One-on-One Instruction

Ruijia Li, Yuan-Hao Jiang, Jiatong Wang, Bo Jiang

Main category: cs.AI

TL;DR: AI-generated tutoring dialogues lack the pedagogical richness of human dialogues, showing structural simplification and less cognitive guidance compared to human “question-factual response-feedback” teaching loops.

DetailsMotivation: To systematically investigate structural and behavioral differences between AI-simulated and authentic human tutoring dialogues, as LLMs currently struggle to generate pedagogically rich interactions that foster higher-order thinking.

Method: Quantitative comparison using Initiation-Response-Feedback (IRF) coding scheme and Epistemic Network Analysis (ENA) to analyze dialogue patterns.

Result: Human dialogues significantly outperform AI in utterance length, questioning behaviors, and general feedback. Human interactions are more cognitively guided and diverse with clear pedagogical guidance, while AI shows structural simplification and behavioral convergence in simple information transfer patterns.

Conclusion: Current AI-generated tutoring has key limitations in pedagogical effectiveness, providing empirical guidance for designing better educational dialogue systems that can replicate human teaching patterns.

Abstract: Heuristic and scaffolded teacher-student dialogues are widely regarded as critical for fostering students’ higher-order thinking and deep learning. However, large language models (LLMs) currently face challenges in generating pedagogically rich interactions. This study systematically investigates the structural and behavioral differences between AI-simulated and authentic human tutoring dialogues. We conducted a quantitative comparison using an Initiation-Response-Feedback (IRF) coding scheme and Epistemic Network Analysis (ENA). The results show that human dialogues are significantly superior to their AI counterparts in utterance length, as well as in questioning (I-Q) and general feedback (F-F) behaviors. More importantly, ENA results reveal a fundamental divergence in interactional patterns: human dialogues are more cognitively guided and diverse, centered around a “question-factual response-feedback” teaching loop that clearly reflects pedagogical guidance and student-driven thinking; in contrast, simulated dialogues exhibit a pattern of structural simplification and behavioral convergence, revolving around an “explanation-simplistic response” loop that is essentially a simple information transfer between the teacher and student. These findings illuminate key limitations in current AI-generated tutoring and provide empirical guidance for designing and evaluating more pedagogically effective generative educational dialogue systems.

[604] BALM-TSF: Balanced Multimodal Alignment for LLM-Based Time Series Forecasting

Shiqiao Zhou, Holger Schöner, Huanbo Lyu, Edouard Fouché, Shuo Wang

Main category: cs.AI

TL;DR: BALM-TSF is a lightweight multimodal framework that balances time series and textual modalities for improved forecasting performance using contrastive alignment and learnable prompts.

DetailsMotivation: Current multimodal architectures for time series forecasting often over-emphasize one modality (text or time series) while neglecting the other, leading to information loss and suboptimal performance.

Method: Processes raw time series through a time series encoder, feeds descriptive statistics to an LLM with learnable prompts, uses scaling strategy and contrastive objective to align textual embeddings with time series embeddings, then integrates both for forecasting.

Result: Achieves state-of-the-art performance in both long-term and few-shot forecasting with minimal trainable parameters, demonstrating effective complementary information utilization.

Conclusion: BALM-TSF successfully addresses modality imbalance and harnesses the complementary strengths of both text and time series data for superior forecasting performance.

Abstract: Time series forecasting is a long-standing and highly challenging research topic. Recently, driven by the rise of large language models (LLMs), research has increasingly shifted from purely time series methods toward harnessing textual modalities to enhance forecasting performance. However, the vast discrepancy between text and temporal data often leads current multimodal architectures to over-emphasise one modality while neglecting the other, resulting in information loss that harms forecasting performance. To address this modality imbalance, we introduce BALM-TSF (Balanced Multimodal Alignment for LLM-Based Time Series Forecasting), a lightweight time series forecasting framework that maintains balance between the two modalities. Specifically, raw time series are processed by the time series encoder, while descriptive statistics of raw time series are fed to an LLM with learnable prompt, producing compact textual embeddings. To ensure balanced cross-modal context alignment of time series and textual embeddings, a simple yet effective scaling strategy combined with a contrastive objective then maps these textual embeddings into the latent space of the time series embeddings. Finally, the aligned textual semantic embeddings and time series embeddings are together integrated for forecasting. Extensive experiments on standard benchmarks show that, with minimal trainable parameters, BALM-TSF achieves state-of-the-art performance in both long-term and few-shot forecasting, confirming its ability to harness complementary information from text and time series. Code is available at https://github.com/ShiqiaoZhou/BALM-TSF.

[605] Dynamic Speculative Agent Planning

Yilin Guan, Wenyue Hua, Qingfeng Lan, Sun Fei, Dujian Ding, Devang Acharya, Chi Wang, William Yang Wang

Main category: cs.AI

TL;DR: DSP is an online reinforcement learning framework that provides lossless acceleration for LLM agents with 30% cost reduction and up to 60% reduction in unnecessary costs, while allowing flexible tradeoffs between latency and cost.

DetailsMotivation: Large language model agents face prohibitive latency and inference costs in deployment. Existing acceleration methods either fail to preserve performance, require extensive offline training, or incur excessive costs with minimal user control over acceleration-performance tradeoffs.

Method: Dynamic Speculative Planning (DSP) - an asynchronous online reinforcement learning framework that optimizes a joint objective balancing end-to-end latency against dollar cost without requiring additional pre-deployment preparation.

Result: Experiments on two standard agent benchmarks show DSP achieves comparable efficiency to the fastest lossless acceleration method while reducing total cost by 30% and unnecessary cost by up to 60%.

Conclusion: DSP provides a flexible, cost-effective solution for accelerating LLM agents with lossless performance, allowing practitioners to control the tradeoff between response speed and operational costs through a single parameter.

Abstract: Despite their remarkable success in complex tasks propelling widespread adoption, large language-model-based agents still face critical deployment challenges due to prohibitive latency and inference costs. While recent work has explored various methods to accelerate inference, existing approaches suffer from significant limitations: they either fail to preserve performance fidelity, require extensive offline training of router modules, or incur excessive operational costs. Moreover, they provide minimal user control over the tradeoff between acceleration and other performance metrics. To address these gaps, we introduce Dynamic Speculative Planning (DSP), an asynchronous online reinforcement learning framework that provides lossless acceleration with substantially reduced costs without requiring additional pre-deployment preparation. DSP explicitly optimizes a joint objective balancing end-to-end latency against dollar cost, allowing practitioners to adjust a single parameter that steers the system toward faster responses, cheaper operation, or any point along this continuum. Experiments on two standard agent benchmarks demonstrate that DSP achieves comparable efficiency to the fastest lossless acceleration method while reducing total cost by 30% and unnecessary cost up to 60%. Our code and data are available through https://github.com/guanyilin428/Dynamic-Speculative-Planning.

[606] NetGent: Agent-Based Automation of Network Application Workflows

Jaber Daneshamooz, Eugene Vuong, Laasya Koduru, Sanjay Chandrasekaran, Arpit Gupta

Main category: cs.AI

TL;DR: NetGent is an AI-agent framework that automates web application workflows using natural language rules to generate realistic network traffic datasets for ML model development.

DetailsMotivation: Developing generalizable ML models for networking requires diverse, realistic traffic data from real-world web applications, but existing browser automation tools are fragile, costly, and lack repeatability.

Method: Users specify workflows as natural-language rules that define state-dependent actions, which are compiled into nondeterministic finite automata (NFAs) and translated into reusable executable code with state synthesis and caching.

Result: NetGent automated 50+ workflows across various domains (video streaming, conferencing, social media, web scraping), producing realistic traffic traces while remaining robust to UI changes and enabling deterministic replay.

Conclusion: NetGent combines language-based agent flexibility with compiled execution reliability, providing a scalable foundation for generating diverse, repeatable network traffic datasets to advance ML in networking.

Abstract: We present NetGent, an AI-agent framework for automating complex application workflows to generate realistic network traffic datasets. Developing generalizable ML models for networking requires data collection from network environments with traffic that results from a diverse set of real-world web applications. However, using existing browser automation tools that are diverse, repeatable, realistic, and efficient remains fragile and costly. NetGent addresses this challenge by allowing users to specify workflows as natural-language rules that define state-dependent actions. These abstract specifications are compiled into nondeterministic finite automata (NFAs), which a state synthesis component translates into reusable, executable code. This design enables deterministic replay, reduces redundant LLM calls through state caching, and adapts quickly when application interfaces change. In experiments, NetGent automated more than 50+ workflows spanning video-on-demand streaming, live video streaming, video conferencing, social media, and web scraping, producing realistic traffic traces while remaining robust to UI variability. By combining the flexibility of language-based agents with the reliability of compiled execution, NetGent provides a scalable foundation for generating the diverse, repeatable datasets needed to advance ML in networking.

Albert Sadowski, Jarosław A. Chudziak

Main category: cs.AI

TL;DR: Modular multi-agent framework decomposes legal reasoning into knowledge acquisition and application stages, achieving 76.4% accuracy on tax tasks vs 18.8% baseline.

DetailsMotivation: Legal reasoning requires precise interpretation of statutes and consistent rule application, presenting challenges for AI systems that need transparency and verifiability.

Method: Two-stage framework: 1) Specialized agents extract legal concepts and formalize rules into verifiable intermediate representations 2) Three-step application: query analysis, symbolic inference, and programmatic answer generation bridging natural language with symbolic reasoning.

Result: Substantial improvement on statutory tax calculation tasks - foundational models achieved 76.4% accuracy compared to 18.8% baseline performance, effectively narrowing the performance gap.

Conclusion: Modular architectures with formalized knowledge representations make legal reasoning more accessible through computationally efficient models while enhancing consistency and explainability, establishing foundation for transparent and trustworthy AI legal systems.

Abstract: Legal reasoning requires both precise interpretation of statutory language and consistent application of complex rules, presenting significant challenges for AI systems. This paper introduces a modular multi-agent framework that decomposes legal reasoning into distinct knowledge acquisition and application stages. In the first stage, specialized agents extract legal concepts and formalize rules to create verifiable intermediate representations of statutes. The second stage applies this knowledge to specific cases through three steps: analyzing queries to map case facts onto the ontology schema, performing symbolic inference to derive logically entailed conclusions, and generating final answers using a programmatic implementation that operationalizes the ontological knowledge. This bridging of natural language understanding with symbolic reasoning provides explicit and verifiable inspection points, significantly enhancing transparency compared to end-to-end approaches. Evaluation on statutory tax calculation tasks demonstrates substantial improvements, with foundational models achieving 76.4% accuracy compared to 18.8% baseline performance, effectively narrowing the performance gap between reasoning and foundational models. These findings suggest that modular architectures with formalized knowledge representations can make sophisticated legal reasoning more accessible through computationally efficient models while enhancing consistency and explainability in AI legal reasoning, establishing a foundation for future research into more transparent, trustworthy, and effective AI systems for legal domain.

[608] Efficient Graph Understanding with LLMs via Structured Context Injection

Govind Waghmare, Sumedh BG, Sonia Gupta, Srikanta Bedathur

Main category: cs.AI

TL;DR: A framework for structured context injection that embeds task-specific information in LLM inputs to solve graph problems without fine-tuning, achieving performance comparable to more complex methods.

DetailsMotivation: Graph reasoning tasks remain challenging for LLMs unless mapped to conceptually grounded representations, but fine-tuning or multi-step querying approaches are expensive and inefficient.

Method: Systematically inject structured context directly into the input to guide LLMs, enabling implicit alignment of tasks with grounded conceptual spaces without requiring model fine-tuning.

Result: Consistent performance improvements across multiple graph tasks using both lightweight and large models, with structured input context rivaling or surpassing more complex approaches.

Conclusion: Structured context injection is an effective and scalable strategy for graph understanding with LLMs, offering a practical alternative to expensive fine-tuning methods.

Abstract: Large Language Models (LLMs) have shown strong capabilities in solving problems across domains, including graph-related tasks traditionally addressed by symbolic or algorithmic methods. In this work, we present a framework for structured context injection, where task-specific information is systematically embedded in the input to guide LLMs in solving a wide range of graph problems. Our method does not require fine-tuning of LLMs, making it cost-efficient and lightweight. We observe that certain graph reasoning tasks remain challenging for LLMs unless they are mapped to conceptually grounded representations. However, achieving such mappings through fine-tuning or repeated multi-step querying can be expensive and inefficient. Our approach offers a practical alternative by injecting structured context directly into the input, enabling the LLM to implicitly align the task with grounded conceptual spaces. We evaluate the approach on multiple graph tasks using both lightweight and large models, highlighting the trade-offs between accuracy and computational cost. The results demonstrate consistent performance improvements, showing that structured input context can rival or surpass more complex approaches. Our findings underscore the value of structured context injection as an effective and scalable strategy for graph understanding with LLMs.

Ziqi Wang, Boqin Yuan

Main category: cs.AI

TL;DR: L-MARS is a multi-agent legal QA system that reduces hallucinations through coordinated reasoning, targeted search across multiple sources, and iterative verification before answer synthesis.

DetailsMotivation: To address hallucination and uncertainty in legal question answering by moving beyond single-pass RAG systems and providing more reliable, grounded legal information retrieval.

Method: Decomposes queries into subproblems, performs targeted searches across heterogeneous sources (web, local RAG, case law), and uses a Judge Agent for verification of sufficiency, jurisdiction, and temporal validity in an iterative reasoning-search-verification loop.

Result: Substantially improves factual accuracy, reduces uncertainty, and achieves higher preference scores from both human experts and LLM-based judges on the LegalSearchQA benchmark of 200 up-to-date legal questions.

Conclusion: Multi-agent reasoning with agentic search provides a scalable and reproducible blueprint for deploying LLMs in high-stakes domains requiring precise legal retrieval and deliberation.

Abstract: We present L-MARS (Legal Multi-Agent Workflow with Orchestrated Reasoning and Agentic Search), a system that reduces hallucination and uncertainty in legal question answering through coordinated multi-agent reasoning and retrieval. Unlike single-pass retrieval-augmented generation (RAG), L-MARS decomposes queries into subproblems, issues targeted searches across heterogeneous sources (Serper web, local RAG, CourtListener case law), and employs a Judge Agent to verify sufficiency, jurisdiction, and temporal validity before answer synthesis. This iterative reasoning-search-verification loop maintains coherence, filters noisy evidence, and grounds answers in authoritative law. We evaluated L-MARS on LegalSearchQA, a new benchmark of 200 up-to-date multiple choice legal questions in 2025. Results show that L-MARS substantially improves factual accuracy, reduces uncertainty, and achieves higher preference scores from both human experts and LLM-based judges. Our work demonstrates that multi-agent reasoning with agentic search offers a scalable and reproducible blueprint for deploying LLMs in high-stakes domains requiring precise legal retrieval and deliberation.

[610] Aligning Reasoning LLMs for Materials Discovery with Physics-aware Rejection Sampling

Lee Hyun, Sohee Yoon, Jinwoo Park, Sue In Chae, Seongeon Park, Jooyeon Ahn, Yebin Jung, Youjung Chung, Hogeun Chang, Myeonginn Kang, Jina Kim, Ho-Gyeong Kim, Myeonghun Jeong

Main category: cs.AI

TL;DR: Physics-aware Rejection Sampling (PaRS) improves AI materials discovery by selecting reasoning traces that follow physical laws and match target values, enhancing accuracy and reducing physics violations.

DetailsMotivation: Current AI-driven materials discovery uses binary correctness or preference signals that don't adequately reflect physical admissibility, leading to unreliable predictions.

Method: Introduces Physics-aware Rejection Sampling (PaRS) that selects training traces based on consistency with fundamental physics and numerical closeness to targets, using a teacher-student model framework.

Result: Improves accuracy and calibration, reduces physics-violation rates, and lowers sampling cost compared to baseline rejection sampling methods.

Conclusion: Domain-aware constraints with trace-level selection provide a practical path toward reliable, efficient large reasoning models for process-aware property prediction and materials design.

Abstract: AI-driven materials discovery that couples automated experimentation with algorithmic decision-making requires process aware recipe to property predictors that are accurate, calibrated, and physically admissible. We approach this as a reasoning problem with large reasoning models (LRMs). To instill reasoning capability into language models, we curate reasoning traces from a teacher model to train a student model. However, most training pipelines select reasoning traces using binary correctness or learned preference signals that poorly reflect physical admissibility. We introduce Physics-aware Rejection Sampling (PaRS), a training-time trace selection scheme that favors traces consistent with fundamental physics and numerically close to targets, with lightweight halting to control compute. We instantiate our framework with a large student model fine-tuned on traces synthesized by a larger teacher model, and evaluate under matched token budgets against various rejection sampling baselines. Our method improves accuracy and calibration, reduces physics-violation rates, and lowers sampling cost relative to baselines. These results indicate that modest, domain-aware constraints combined with trace-level selection provide a practical path toward reliable, efficient LRMs for process-aware property prediction and closed-loop materials design.

[611] Sharpe Ratio Optimization in Markov Decision Processes

Shuai Ma, Guangwu Liu, Li Xia

Main category: cs.AI

TL;DR: This paper presents a novel dynamic programming approach to optimize Sharpe ratio in Markov decision processes by converting it to mean-squared-variance optimization using Dinkelbach’s transform and developing iterative algorithms that converge to optimal policies.

DetailsMotivation: Sharpe ratio optimization in MDPs is challenging due to two fundamental problems: dynamic programming doesn't work for fractional objectives and is invalid for risk metrics. Existing methods struggle with these limitations.

Method: Uses Dinkelbach’s transform to convert Sharpe ratio to mean-squared-variance objective. Develops iterative algorithms that solve M2V problems and update risk-sensitive parameters, proving convergence for both average and discounted MDP settings with policy iteration procedures.

Result: The proposed algorithm produces monotonically increasing Sharpe ratios that converge to the optimal value. Numerical experiments validate the approach, showing it’s the first dynamic programming type solution for Sharpe ratio optimization in MDPs.

Conclusion: The method successfully addresses both challenges of fractional objectives and risk metrics in MDPs, providing a convergent dynamic programming framework that can potentially be extended to other fractional objective problems in reinforcement learning.

Abstract: Sharpe ratio (also known as reward-to-variability ratio) is a widely-used metric in finance, which measures the additional return at the cost of per unit of increased risk (standard deviation of return). However, the optimization of Sharpe ratio in Markov decision processes (MDPs) is challenging, because there exist two difficulties hindering the application of dynamic programming. One is that dynamic programming does not work for fractional objectives, and the other is that dynamic programming is invalid for risk metrics. In this paper, we study the Sharpe ratio optimization in infinite-horizon MDPs, considering both the long-run average and discounted settings. We address the first challenge with the Dinkelbachs transform, which converts the Sharpe ratio objective to a mean-squared-variance (M2V) objective. It is shown that the M2V optimization and the original Sharpe ratio optimization share the same optimal policy when the risk-sensitive parameter is equal to the optimal Sharpe ratio. For the second challenge, we develop an iterative algorithm to solve the M2V optimization which is similar to a mean-variance optimization in MDPs. We iteratively solve the M2V problem and obtain the associated Sharpe ratio that is used to update the risk-sensitive parameter in the next iteration of M2V problems. We show that such a sequence of Sharpe ratios derived is monotonically increasing and converges to the optimal Sharpe ratio. For both average and discounted MDP settings, we develop a policy iteration procedure and prove its convergence to the optimum. Numerical experiments are conducted for validation. To the best of our knowledge, our approach is the first that solves the Sharpe ratio optimization in MDPs with dynamic programming type algorithms. We believe that the proposed algorithm can shed light on solving MDPs with other fractional objectives.

[612] Neuro-Symbolic Predictive Process Monitoring

Axel Mezini, Elena Umili, Ivan Donadello, Fabrizio Maria Maggi, Matteo Mancanelli, Fabio Patrizi

Main category: cs.AI

TL;DR: Neuro-Symbolic approach combining deep learning with temporal logic constraints for suffix prediction in BPM, improving accuracy and logical consistency.

DetailsMotivation: Current deep learning models for suffix prediction often fail to satisfy basic logical constraints due to lack of domain knowledge integration during training.

Method: Integrates Linear Temporal Logic over finite traces (LTLf) into autoregressive sequence predictors using differentiable logical loss function with Gumbel-Softmax trick and soft LTLf approximation.

Result: Experimental evaluation on three real-world datasets shows improved suffix prediction accuracy and compliance with temporal constraints.

Conclusion: Framework improves Neuro-Symbolic AI for sequence generation tasks beyond BPM, with effective local and global logic loss variants for noisy settings.

Abstract: This paper addresses the problem of suffix prediction in Business Process Management (BPM) by proposing a Neuro-Symbolic Predictive Process Monitoring (PPM) approach that integrates data-driven learning with temporal logic-based prior knowledge. While recent approaches leverage deep learning models for suffix prediction, they often fail to satisfy even basic logical constraints due to the absence of explicit integration of domain knowledge during training. We propose a novel method to incorporate Linear Temporal Logic over finite traces (LTLf) into the training process of autoregressive sequence predictors. Our approach introduces a differentiable logical loss function, defined using a soft approximation of LTLf semantics and the Gumbel-Softmax trick, which can be combined with standard predictive losses. This ensures the model learns to generate suffixes that are both accurate and logically consistent. Experimental evaluation on three real-world datasets shows that our method improves suffix prediction accuracy and compliance with temporal constraints. We also introduce two variants of the logic loss (local and global) and demonstrate their effectiveness under noisy and realistic settings. While developed in the context of BPM, our framework is applicable to any symbolic sequence generation task and contributes toward advancing Neuro-Symbolic AI.

[613] ChatCLIDS: Simulating Persuasive AI Dialogues to Promote Closed-Loop Insulin Adoption in Type 1 Diabetes Care

Zonghai Yao, Talha Chafekar, Junda Wang, Shuo Han, Feiyun Ouyang, Junhui Qian, Lingxi Li, Hong Yu

Main category: cs.AI

TL;DR: ChatCLIDS is the first benchmark for evaluating LLM-driven persuasive dialogue in healthcare, specifically for closed-loop insulin delivery system adoption in type 1 diabetes, featuring expert-validated virtual patients and multi-dimensional evaluation.

DetailsMotivation: Real-world adoption of closed-loop insulin delivery systems remains low due to behavioral, psychosocial, and social barriers rather than technical failures, highlighting the need for effective persuasive AI in healthcare.

Method: The framework includes a library of expert-validated virtual patients with clinically grounded profiles, simulates multi-turn interactions with nurse agents using evidence-based persuasive strategies, and supports longitudinal counseling and adversarial social influence scenarios.

Result: Larger and more reflective LLMs adapt strategies over time, but all models struggle to overcome resistance, particularly under realistic social pressure scenarios.

Conclusion: Current LLMs have critical limitations for behavior change applications, and ChatCLIDS provides a high-fidelity, scalable testbed for advancing trustworthy persuasive AI in healthcare and other domains.

Abstract: Real-world adoption of closed-loop insulin delivery systems (CLIDS) in type 1 diabetes remains low, driven not by technical failure, but by diverse behavioral, psychosocial, and social barriers. We introduce ChatCLIDS, the first benchmark to rigorously evaluate LLM-driven persuasive dialogue for health behavior change. Our framework features a library of expert-validated virtual patients, each with clinically grounded, heterogeneous profiles and realistic adoption barriers, and simulates multi-turn interactions with nurse agents equipped with a diverse set of evidence-based persuasive strategies. ChatCLIDS uniquely supports longitudinal counseling and adversarial social influence scenarios, enabling robust, multi-dimensional evaluation. Our findings reveal that while larger and more reflective LLMs adapt strategies over time, all models struggle to overcome resistance, especially under realistic social pressure. These results highlight critical limitations of current LLMs for behavior change, and offer a high-fidelity, scalable testbed for advancing trustworthy persuasive AI in healthcare and beyond.

[614] Robust Deep Monte Carlo Counterfactual Regret Minimization: Addressing Theoretical Risks in Neural Fictitious Self-Play

Zakaria El Jaafari

Main category: cs.AI

TL;DR: Analysis of neural MCCFR component effectiveness across game scales, proposing adaptive framework with 60% improvement on Kuhn Poker and 23.5% on Leduc Poker.

DetailsMotivation: MCCFR integration with deep neural networks faces scale-dependent challenges that manifest differently across game complexities, requiring adaptive mitigation strategies.

Method: Proposed Robust Deep MCCFR framework with target networks, uniform exploration mixing, variance-aware training objectives, and diagnostic monitoring. Systematic ablation studies on Kuhn and Leduc Poker.

Result: 60% improvement on Kuhn Poker (0.0628 vs 0.156 exploitability) and 23.5% improvement on Leduc Poker (0.2386 vs 0.3703). Identified scale-dependent component effectiveness and critical interactions.

Conclusion: Careful component selection is more important than comprehensive mitigation. Framework provides convergence guarantees and practical guidelines for larger games with scale-dependent risk patterns.

Abstract: Monte Carlo Counterfactual Regret Minimization (MCCFR) has emerged as a cornerstone algorithm for solving extensive-form games, but its integration with deep neural networks introduces scale-dependent challenges that manifest differently across game complexities. This paper presents a comprehensive analysis of how neural MCCFR component effectiveness varies with game scale and proposes an adaptive framework for selective component deployment. We identify that theoretical risks such as nonstationary target distribution shifts, action support collapse, variance explosion, and warm-starting bias have scale-dependent manifestation patterns, requiring different mitigation strategies for small versus large games. Our proposed Robust Deep MCCFR framework incorporates target networks with delayed updates, uniform exploration mixing, variance-aware training objectives, and comprehensive diagnostic monitoring. Through systematic ablation studies on Kuhn and Leduc Poker, we demonstrate scale-dependent component effectiveness and identify critical component interactions. The best configuration achieves final exploitability of 0.0628 on Kuhn Poker, representing a 60% improvement over the classical framework (0.156). On the more complex Leduc Poker domain, selective component usage achieves exploitability of 0.2386, a 23.5% improvement over the classical framework (0.3703) and highlighting the importance of careful component selection over comprehensive mitigation. Our contributions include: (1) a formal theoretical analysis of risks in neural MCCFR, (2) a principled mitigation framework with convergence guarantees, (3) comprehensive multi-scale experimental validation revealing scale-dependent component interactions, and (4) practical guidelines for deployment in larger games.

[615] SATQuest: A Verifier for Logical Reasoning Evaluation and Reinforcement Fine-Tuning of LLMs

Yanxiao Zhao, Yaqian Li, Zihao Bo, Rinyoichi Takezoe, Haojia Hui, Mo Guang, Lei Ren, Xiaolin Qin, Kaiwen Long

Main category: cs.AI

TL;DR: SATQuest is a systematic verifier that generates diverse logical reasoning problems from CNF instances to evaluate and enhance LLM reasoning capabilities through three dimensions: instance scale, problem type, and question format.

DetailsMotivation: Existing benchmarks lack controllable and scalable tools for fine-grained analysis of LLM reasoning capabilities, with insufficient variable control and narrow problem types/formats.

Method: Generates Satisfiability-based logical reasoning problems directly from CNF instances using randomized SAT-based problem generation and objective answer verification via PySAT.

Result: Evaluation revealed significant limitations in LLM logical reasoning, particularly in generalization beyond familiar mathematical formats. Reinforcement fine-tuning with SATQuest rewards substantially improved targeted task performance and generalized to more complex instances.

Conclusion: SATQuest demonstrates potential as a foundational tool for advancing LLM logical reasoning, though challenges remain in cross-format adaptation.

Abstract: Recent advances in Large Language Models (LLMs) have demonstrated remarkable general reasoning capabilities. However, systematically evaluating and enhancing these reasoning capabilities is challenging due to the lack of controllable and scalable tools for fine-grained analysis. Existing benchmarks and datasets often lack the necessary variable control for multi-dimensional, systematic analysis and training, or have narrow problem types and formats. To address these limitations, we introduce SATQuest, a systematic verifier designed to evaluate and enhance logical reasoning in LLMs by generating diverse, Satisfiability-based logical reasoning problems directly from Conjunctive Normal Form (CNF) instances. SATQuest structures these problems along three orthogonal dimensions: instance scale, problem type, and question format, employing randomized, SAT-based problem generation and objective answer verification via PySAT. This design mitigates memorization issues, allows for nuanced insights into reasoning performance, and enables effective reinforcement fine-tuning. Our extensive evaluation of various LLMs using SATQuest identified significant limitations in their logical reasoning, particularly in generalizing beyond familiar mathematical formats. Furthermore, we show that reinforcement fine-tuning with SATQuest rewards substantially improves targeted task performance and generalizes to more complex instances, while highlighting remaining challenges in cross-format adaptation. Through these demonstrations, we showcase SATQuest’s potential as a foundational tool and a valuable starting point for advancing LLM logical reasoning.

[616] Towards Efficient LLM Grounding for Embodied Multi-Agent Collaboration

Yang Zhang, Shixin Yang, Chenjia Bai, Fei Wu, Xiu Li, Zhen Wang, Xuelong Li

Main category: cs.AI

TL;DR: ReAd framework uses reinforced advantage feedback to improve LLM planning efficiency for multi-agent collaboration, reducing LLM queries while increasing success rates.

DetailsMotivation: Existing methods for grounding LLM reasoning in embodied tasks rely too heavily on physical verification and self-reflection, leading to excessive and inefficient LLM querying during multi-agent collaboration.

Method: Proposes Reinforced Advantage feedback (ReAd) framework that performs critic regression to learn a sequential advantage function from LLM-planned data, then treats the LLM planner as an optimizer to generate actions maximizing the advantage function.

Result: Experiments on Overcooked-AI and RoCoBench show ReAd surpasses baselines in success rate while significantly decreasing agent interaction steps and LLM query rounds.

Conclusion: ReAd provides an efficient framework for grounding LLMs in multi-agent collaboration by endowing LLMs with foresight to discern action contributions to final task completion, with theoretical foundation in advantage-weighted regression.

Abstract: Grounding the reasoning ability of large language models (LLMs) for embodied tasks is challenging due to the complexity of the physical world. Especially, LLM planning for multi-agent collaboration requires communication of agents or credit assignment as the feedback to re-adjust the proposed plans and achieve effective coordination. However, existing methods that overly rely on physical verification or self-reflection suffer from excessive and inefficient querying of LLMs. In this paper, we propose a novel framework for multi-agent collaboration that introduces Reinforced Advantage feedback (ReAd) for efficient self-refinement of plans. Specifically, we perform critic regression to learn a sequential advantage function from LLM-planned data, and then treat the LLM planner as an optimizer to generate actions that maximize the advantage function. It endows the LLM with the foresight to discern whether the action contributes to accomplishing the final task. We provide theoretical analysis by extending advantage-weighted regression in reinforcement learning to multi-agent systems. Experiments on Overcooked-AI and a difficult variant of RoCoBench show that ReAd surpasses baselines in success rate, and also significantly decreases the interaction steps of agents and query rounds of LLMs, demonstrating its high efficiency for grounding LLMs. More results are given at https://read-llm.github.io.

[617] UrbanInsight: A Distributed Edge Computing Framework with LLM-Powered Data Filtering for Smart City Digital Twins

Kishor Datta Gupta, Md Manjurul Ahsan, Mohd Ariful Haque, Roy George, Azmine Toushik Wasi

Main category: cs.AI

TL;DR: A framework combining physics-informed ML, knowledge graphs, and LLMs for real-time urban data analysis and adaptive decision-making in smart cities.

DetailsMotivation: Cities generate massive data streams but existing systems struggle with scale, latency, and fragmented insights, limiting their ability to improve urban life effectively.

Method: Blends physics-informed machine learning (for real-world constraint adherence), multimodal data fusion via knowledge graphs (semantic integration), and adaptive rule-based intelligence using LLMs for real-time filtering and decision-making at the edge.

Result: Creates a foundation for digital twin systems that move beyond passive monitoring to provide actionable insights, enabling efficient operation even under constrained resources.

Conclusion: This unified approach of physics-based reasoning, semantic data fusion, and adaptive rule generation enables responsive, trustworthy, and sustainable smart infrastructure development.

Abstract: Cities today generate enormous streams of data from sensors, cameras, and connected infrastructure. While this information offers unprecedented opportunities to improve urban life, most existing systems struggle with scale, latency, and fragmented insights. This work introduces a framework that blends physics-informed machine learning, multimodal data fusion, and knowledge graph representation with adaptive, rule-based intelligence powered by large language models (LLMs). Physics-informed methods ground learning in real-world constraints, ensuring predictions remain meaningful and consistent with physical dynamics. Knowledge graphs act as the semantic backbone, integrating heterogeneous sensor data into a connected, queryable structure. At the edge, LLMs generate context-aware rules that adapt filtering and decision-making in real time, enabling efficient operation even under constrained resources. Together, these elements form a foundation for digital twin systems that go beyond passive monitoring to provide actionable insights. By uniting physics-based reasoning, semantic data fusion, and adaptive rule generation, this approach opens new possibilities for creating responsive, trustworthy, and sustainable smart infrastructures.

[618] A Hybrid Ai Framework For Strategic Patent Portfolio Pruning: Integrating Learning To-Rank And Market Need Analysis For Technology Transfer Optimization

Manish Verma, Vivek Sharma, Vishal Singh

Main category: cs.AI

TL;DR: A hybrid AI framework combining Learning to Rank with Need-Seed agents to automate patent portfolio pruning for technology transfer by matching high-value patents to market needs.

DetailsMotivation: Current patent valuation methods rely on retrospective indicators and manual analysis, which are time-intensive and inefficient for identifying high-value assets for technology transfer.

Method: Multi-stage hybrid intelligence framework combining: 1) Learning to Rank model evaluating patents against 30+ legal/commercial parameters, 2) Need Agent using NLP to mine market needs from unstructured data, 3) Seed Agent using fine-tuned LLMs to analyze patent claims, and 4) Core Ontology Framework matching patents to market demands with HITL validation.

Result: The framework automates and deepens patent portfolio analysis, generating strategic rationale for divestment decisions by systematically matching high-potential patents to documented market demands.

Conclusion: The proposed hybrid intelligence framework provides an automated, adaptable, and credible solution for identifying high-value patent assets for technology transfer, overcoming limitations of traditional manual methods.

Abstract: This paper introduces a novel, multi stage hybrid intelligence framework for pruning patent portfolios to identify high value assets for technology transfer. Current patent valuation methods often rely on retrospective indicators or manual, time intensive analysis. Our framework automates and deepens this process by combining a Learning to Rank (LTR) model, which evaluates patents against over 30 legal and commercial parameters, with a unique “Need-Seed” agent-based system. The “Need Agent” uses Natural Language Processing (NLP) to mine unstructured market and industry data, identifying explicit technological needs. Concurrently, the “Seed Agent” employs fine tuned Large Language Models (LLMs) to analyze patent claims and map their technological capabilities. The system generates a “Core Ontology Framework” that matches high potential patents (Seeds) to documented market demands (Needs), providing a strategic rationale for divestment decisions. We detail the architecture, including a dynamic parameter weighting system and a crucial Human in the-Loop (HITL) validation protocol, to ensure both adaptability and real-world credibility.

[619] Ultra Strong Machine Learning: Teaching Humans Active Learning Strategies via Automated AI Explanations

Lun Ai, Johannes Langer, Ute Schmid, Stephen Muggleton

Main category: cs.AI

TL;DR: LENS is a neuro-symbolic method combining symbolic program synthesis with LLMs to automatically explain machine-learned logic programs in natural language, outperforming direct LLM prompting and templates but showing limited human learning improvements.

DetailsMotivation: To address the limitation of hand-crafted explanation templates in Ultra Strong Machine Learning (USML) systems by creating scalable automated natural language explanations that can teach humans.

Method: Combines symbolic program synthesis with large language models (LLMs) to generate explanations, evaluated through systematic testing with multiple LLM judges and human validation, plus human learning experiments across three domains.

Result: LENS generates superior explanations compared to direct LLM prompting and hand-crafted templates, but human learning experiments showed no significant performance improvements, suggesting comprehensive LLM responses may overwhelm users for simpler problems.

Conclusion: Provides a foundation for building effective USML systems to support human learning, though current approach may need refinement for optimal teaching effectiveness with human users.

Abstract: Ultra Strong Machine Learning (USML) refers to symbolic learning systems that not only improve their own performance but can also teach their acquired knowledge to quantifiably improve human performance. In this work, we present LENS (Logic Programming Explanation via Neural Summarisation), a neuro-symbolic method that combines symbolic program synthesis with large language models (LLMs) to automate the explanation of machine-learned logic programs in natural language. LENS addresses a key limitation of prior USML approaches by replacing hand-crafted explanation templates with scalable automated generation. Through systematic evaluation using multiple LLM judges and human validation, we demonstrate that LENS generates superior explanations compared to direct LLM prompting and hand-crafted templates. To investigate whether LENS can teach transferable active learning strategies, we carried out a human learning experiment across three related domains. Our results show no significant human performance improvements, suggesting that comprehensive LLM responses may overwhelm users for simpler problems rather than providing learning support. Our work provides a solid foundation for building effective USML systems to support human learning. The source code is available on: https://github.com/lun-ai/LENS.git.

[620] CoreThink: A Symbolic Reasoning Layer to reason over Long Horizon Tasks with LLMs

Jay Vaghasiya, Omkar Ghugarkar, Vishvesh Bhat, Vipul Dholaria, Julian McAuley

Main category: cs.AI

TL;DR: CoreThink introduces a novel General Symbolics reasoning method that achieves SOTA performance on multiple benchmarks without fine-tuning or training costs, providing pure performance uplift for reasoning tasks.

DetailsMotivation: Existing reasoning methods like test-time scaling, SFT, and RLVR are reaching diminishing returns in LLM performance, necessitating new reasoning techniques that don't negatively impact model accuracy.

Method: General Symbolics reasoning approach structured around three key use cases: tool-calling, code generation, and planning. The method operates without any finetuning or training costs.

Result: Achieved SOTA scores: 66.66% on Livecodebench v6, 89% on Instruction-Following Evals, 24.4% on ARC-AGI-2, and 62.3% on SWE-Bench Lite through an agentic coding IDE.

Conclusion: CoreThink’s General Symbolics provides a pure performance uplift for reasoning tasks and represents a necessary advancement beyond incumbent methods that face diminishing returns.

Abstract: We introduce CoreThink, a state-of-the-art Reasoning Layer built upon a novel reasoning method called General Symbolics. This approach diverges from reasoning paradigms such as test-time scaling, Supervised Fine-Tuning (SFT), and Reinforcement Learning with Verifiable Rewards (RLVR). CoreThink General Symbolic Reasoner (GSR) is specifically structured around three key use cases: tool-calling, code generation, and planning, demonstrating exemplary performance across a total of seven benchmarks in their respective areas. Notably, we are achieving SOTA scores of 66.66% on Livecodebench v6, 89% on Instruction-Following Evals, and 24.4% on ARC-AGI-2. We also present an agentic coding IDE, developed using the principles of General Symbolics, which achieves a state-of-the-art accuracy of 62.3% on \texttt{SWE-Bench Lite}. We are able to achieve these improvements without any finetuning or training costs. Our Reasoning Layer is designed to provide a pure performance uplift, ensuring that a model’s accuracy on reasoning tasks is never negatively impacted. We argue that incumbent methods will eventually lead to diminishing returns in LLM performance, necessitating the development of new reasoning techniques. This technical report details our approach at a high level and the availability of the CoreThink models for reasoning-intensive use cases.

Zifeng Ding, Shenyang Huang, Zeyu Cao, Emma Kondrup, Zachary Yang, Xingyue Huang, Yuan Sui, Zhangdie Yuan, Yuqicheng Zhu, Xianglong Hu, Yuan He, Farimah Poursafaei, Michael Bronstein, Andreas Vlachos

Main category: cs.AI

TL;DR: ReaL-TG is a reinforcement learning framework that fine-tunes LLMs for explainable link forecasting on temporal graphs, outperforming larger models while producing high-quality explanations.

DetailsMotivation: Traditional temporal graph neural networks lack explainability and cannot handle unseen graphs without retraining. Existing LLM approaches are limited to static graphs or small synthetic datasets and lack proper evaluation of reasoning traces.

Method: Reinforcement learning framework that fine-tunes LLMs using outcome-based rewards to encourage self-exploration of reasoning strategies from graph structure and production of explanations that justify predictions.

Result: ReaL-TG-4B (fine-tuned Qwen3-4B) outperforms much larger frontier LLMs including GPT-5 mini on ranking metrics, while producing high-quality explanations validated by both LLM judge and human evaluation.

Conclusion: The framework successfully enables LLMs to perform explainable temporal graph reasoning with strong performance and verifiable explanation quality, addressing key limitations of existing approaches.

Abstract: Forecasting future links is a central task in temporal graph (TG) reasoning, requiring models to leverage historical interactions to predict upcoming ones. Traditional neural approaches, such as temporal graph neural networks, achieve strong performance but lack explainability and cannot be applied to unseen graphs without retraining. Recent studies have begun to explore using large language models (LLMs) for graph reasoning, but most of them are constrained to static graphs or small synthetic TGs and lack the evaluation of the quality of reasoning traces generated by LLMs. In this work, we present Reasoning-Enhanced Learning for Temporal Graphs (ReaL-TG), a reinforcement learning framework that fine-tunes LLMs to perform explainable link forecasting on real-world TGs. ReaL-TG uses outcome-based reward to encourage models to self-explore reasoning strategies from graph structure and to produce explanations that directly justify their predictions. To enable evaluation on LLM-generated reasoning traces, we propose a new evaluation protocol combining ranking metrics with an LLM-as-a-Judge system that assesses both the quality of reasoning and the impact of hallucinations. Experiments with ReaL-TG-4B, obtained by fine-tuning Qwen3-4B under our framework, show that it outperforms much larger frontier LLMs, including GPT-5 mini, on ranking metrics, while producing high-quality explanations confirmed by both the LLM judge and human evaluation.

[622] Causal MAS: A Survey of Large Language Model Architectures for Discovery and Effect Estimation

Adib Bazgir, Amir Habibdoust, Yuwen Zhang, Xing Song

Main category: cs.AI

TL;DR: Review paper on causal multi-agent LLM systems that use collaborative LLM agents to overcome limitations in causal reasoning, discovery, and estimation.

DetailsMotivation: LLMs struggle with complex causal reasoning due to hallucination, spurious correlations, and difficulties with nuanced/personalized causal relationships. Multi-agent systems offer a promising solution.

Method: Examines diverse architectural patterns including pipeline processing, debate frameworks, simulation environments, and iterative refinement loops using multiple LLM-based agents.

Result: Provides comprehensive overview of how multi-agent LLM systems are being applied to causal reasoning, discovery, and effect estimation across various domains.

Conclusion: Causal multi-agent LLMs represent a synergistic field with significant potential, though challenges remain that require further research and development.

Abstract: Large Language Models (LLMs) have demonstrated remarkable capabilities in various reasoning and generation tasks. However, their proficiency in complex causal reasoning, discovery, and estimation remains an area of active development, often hindered by issues like hallucination, reliance on spurious correlations, and difficulties in handling nuanced, domain-specific, or personalized causal relationships. Multi-agent systems, leveraging the collaborative or specialized abilities of multiple LLM-based agents, are emerging as a powerful paradigm to address these limitations. This review paper explores the burgeoning field of causal multi-agent LLMs. We examine how these systems are designed to tackle different facets of causality, including causal reasoning and counterfactual analysis, causal discovery from data, and the estimation of causal effects. We delve into the diverse architectural patterns and interaction protocols employed, from pipeline-based processing and debate frameworks to simulation environments and iterative refinement loops. Furthermore, we discuss the evaluation methodologies, benchmarks, and diverse application domains where causal multi-agent LLMs are making an impact, including scientific discovery, healthcare, fact-checking, and personalized systems. Finally, we highlight the persistent challenges, open research questions, and promising future directions in this synergistic field, aiming to provide a comprehensive overview of its current state and potential trajectory.

[623] A Comprehensive Survey of Self-Evolving AI Agents: A New Paradigm Bridging Foundation Models and Lifelong Agentic Systems

Jinyuan Fang, Yanwen Peng, Xi Zhang, Yingxu Wang, Xinhao Yi, Guibin Zhang, Yi Xu, Bin Wu, Siwei Liu, Zihao Li, Zhaochun Ren, Nikos Aletras, Xi Wang, Han Zhou, Zaiqiao Meng

Main category: cs.AI

TL;DR: Survey paper on self-evolving AI agents that automatically adapt using interaction data and environmental feedback, presenting a unified framework and reviewing techniques across different domains.

DetailsMotivation: Existing AI agent systems rely on static manual configurations that limit adaptability to dynamic environments, creating need for self-evolving agents that can continuously improve.

Method: Proposes a unified conceptual framework with four key components (System Inputs, Agent System, Environment, Optimisers) and systematically reviews self-evolving techniques targeting different agent system components across various domains.

Result: Comprehensive review of self-evolving agent techniques, domain-specific evolution strategies for biomedicine, programming, and finance, plus evaluation, safety, and ethical considerations.

Conclusion: Provides foundation for developing more adaptive, autonomous, and lifelong agentic systems by systematizing understanding of self-evolving AI agents and their implementation considerations.

Abstract: Recent advances in large language models have sparked growing interest in AI agents capable of solving complex, real-world tasks. However, most existing agent systems rely on manually crafted configurations that remain static after deployment, limiting their ability to adapt to dynamic and evolving environments. To this end, recent research has explored agent evolution techniques that aim to automatically enhance agent systems based on interaction data and environmental feedback. This emerging direction lays the foundation for self-evolving AI agents, which bridge the static capabilities of foundation models with the continuous adaptability required by lifelong agentic systems. In this survey, we provide a comprehensive review of existing techniques for self-evolving agentic systems. Specifically, we first introduce a unified conceptual framework that abstracts the feedback loop underlying the design of self-evolving agentic systems. The framework highlights four key components: System Inputs, Agent System, Environment, and Optimisers, serving as a foundation for understanding and comparing different strategies. Based on this framework, we systematically review a wide range of self-evolving techniques that target different components of the agent system. We also investigate domain-specific evolution strategies developed for specialised fields such as biomedicine, programming, and finance, where optimisation objectives are tightly coupled with domain constraints. In addition, we provide a dedicated discussion on the evaluation, safety, and ethical considerations for self-evolving agentic systems, which are critical to ensuring their effectiveness and reliability. This survey aims to provide researchers and practitioners with a systematic understanding of self-evolving AI agents, laying the foundation for the development of more adaptive, autonomous, and lifelong agentic systems.

[624] Supporting Our AI Overlords: Redesigning Data Systems to be Agent-First

Shu Liu, Soujanya Ponnapalli, Shreya Shankar, Sepanta Zeighami, Alan Zhu, Shubham Agarwal, Ruiqi Chen, Samion Suwito, Shuo Yuan, Ion Stoica, Matei Zaharia, Alvin Cheung, Natacha Crooks, Joseph E. Gonzalez, Aditya G. Parameswaran

Main category: cs.AI

TL;DR: LLM agents will become dominant data workloads, requiring new data systems architecture to handle agentic speculation characteristics like scale, heterogeneity, redundancy, and steerability.

DetailsMotivation: LLM agents acting on users' behalf for data manipulation will become the primary workload for data systems, but current systems are challenged by the volume and inefficiencies of agentic speculation processes.

Method: The paper identifies key characteristics of agentic speculation (scale, heterogeneity, redundancy, steerability) and proposes adapting data systems with new query interfaces, processing techniques, and agentic memory stores.

Result: The analysis reveals that current data systems are inadequate for handling LLM agent workloads and outlines research opportunities for agent-first data system architecture.

Conclusion: Data systems need fundamental architectural changes to natively support LLM agent workloads, with new research directions spanning query interfaces, processing techniques, and specialized memory stores for agentic operations.

Abstract: Large Language Model (LLM) agents, acting on their users’ behalf to manipulate and analyze data, are likely to become the dominant workload for data systems in the future. When working with data, agents employ a high-throughput process of exploration and solution formulation for the given task, one we call agentic speculation. The sheer volume and inefficiencies of agentic speculation can pose challenges for present-day data systems. We argue that data systems need to adapt to more natively support agentic workloads. We take advantage of the characteristics of agentic speculation that we identify, i.e., scale, heterogeneity, redundancy, and steerability - to outline a number of new research opportunities for a new agent-first data systems architecture, ranging from new query interfaces, to new query processing techniques, to new agentic memory stores.

[625] Analysis of Error Sources in LLM-based Hypothesis Search for Few-Shot Rule Induction

Aishni Parab, Hongjing Lu, Ying Nian Wu, Sumit Gulwani

Main category: cs.AI

TL;DR: LLM-based hypothesis search matches human performance in few-shot rule induction, while direct program generation lags significantly behind.

DetailsMotivation: To compare different approaches for modeling human-like inductive reasoning capabilities in AI systems, particularly for inferring abstract rules from limited examples.

Method: Compared LLM-based hypothesis search framework with direct program generation approaches on few-shot rule induction tasks, including error analysis of hypothesis generation bottlenecks.

Result: Hypothesis search achieved performance comparable to humans, while direct program generation fell notably behind in rule induction tasks.

Conclusion: LLM-based hypothesis search shows strong potential for modeling inductive reasoning, but current program induction methods face significant challenges that need addressing for more efficient systems.

Abstract: Inductive reasoning enables humans to infer abstract rules from limited examples and apply them to novel situations. In this work, we compare an LLM-based hypothesis search framework with direct program generation approaches on few-shot rule induction tasks. Our findings show that hypothesis search achieves performance comparable to humans, while direct program generation falls notably behind. An error analysis reveals key bottlenecks in hypothesis generation and suggests directions for advancing program induction methods. Overall, this paper underscores the potential of LLM-based hypothesis search for modeling inductive reasoning and the challenges in building more efficient systems.

[626] Quantum-like Coherence Derived from the Interaction between Chemical Reaction and Its Environment

Yukio-Pegio Gunji, Andrew Adamatzky, Panagiotis Mougkogiannis, Andrei Khrenikov

Main category: cs.AI

TL;DR: The paper introduces open computing implemented through chemical reactions, distinguishing between Token computing (individual molecular behavior) and Type computing (normative behavior), showing self-organizing critical phenomena and quantum logic respectively, with their interplay enabling quantum-like coherence and spike wave formation.

DetailsMotivation: To explore the contrast between Artificial Intelligence and Natural-born Intelligence by defining and implementing open computing within chemical reaction systems, aiming to understand how computational processes can adapt to environmental fluctuations.

Method: Modeling chemical reactions where computation is the reaction itself and execution environment is molecular aggregation degree. Implementing open computing through Token computing (individual molecular behavior) and Type computing (normative behavior) with their interplay.

Result: Token computing exhibits self-organizing critical phenomena, Type computing demonstrates quantum logic. Their interplay enables recruitment of fluctuations, quantum coherence across Hilbert spaces, and formation of spike waves for signal transmission.

Conclusion: The system achieves quantum-like coherence through the interplay of Token and Type computing, potentially explaining the source of enzymes controlling spike waves and biochemical rhythms, representing a novel approach to open computing in chemical systems.

Abstract: By uncovering the contrast between Artificial Intelligence and Natural-born Intelligence as a computational process, we define closed computing and open computing, and implement open computing within chemical reactions. This involves forming a mixture and invalidation of the computational process and the execution environment, which are logically distinct, and coalescing both to create a system that adjusts fluctuations. We model chemical reactions by considering the computation as the chemical reaction and the execution environment as the degree of aggregation of molecules that interact with the reactive environment. This results in a chemical reaction that progresses while repeatedly clustering and de-clustering, where concentration no longer holds significant meaning. Open computing is segmented into Token computing, which focuses on the individual behavior of chemical molecules, and Type computing, which focuses on normative behavior. Ultimately, both are constructed as an interplay between the two. In this system, Token computing demonstrates self-organizing critical phenomena, while Type computing exhibits quantum logic. Through their interplay, the recruitment of fluctuations is realized, giving rise to interactions between quantum logical subspaces corresponding to quantum coherence across different Hilbert spaces. As a result, spike waves are formed, enabling signal transmission. This occurrence may be termed quantum-like coherence, implying the source of enzymes responsible for controlling spike waves and biochemical rhythms.

[627] FlashAdventure: A Benchmark for GUI Agents Solving Full Story Arcs in Diverse Adventure Games

Jaewoo Ahn, Junseo Kim, Heeseung Yun, Jaehyeon Son, Dongmin Park, Jaewoong Cho, Gunhee Kim

Main category: cs.AI

TL;DR: FlashAdventure benchmark with 34 Flash adventure games tests full story completion and addresses observation-behavior gap. COAST framework with clue memory improves performance but still lags behind humans.

DetailsMotivation: Existing game benchmarks lack diversity and don't evaluate full storyline completion. Adventure games pose unique challenges with complex narrative-driven interactions and the observation-behavior gap problem.

Method: Introduces FlashAdventure benchmark, CUA-as-a-Judge automated evaluator, and COAST agentic framework that uses long-term clue memory for better planning and sequential task solving.

Result: Current GUI agents struggle with complete story arcs. COAST improves milestone completion by addressing the observation-behavior gap, but significant performance gap remains between best agents and humans.

Conclusion: The benchmark reveals limitations in current GUI agents for complex narrative tasks. While COAST shows improvement, continued research is needed to bridge the human-agent performance gap in adventure game completion.

Abstract: GUI agents powered by LLMs show promise in interacting with diverse digital environments. Among these, video games offer a valuable testbed due to their varied interfaces, with adventure games posing additional challenges through complex, narrative-driven interactions. Existing game benchmarks, however, lack diversity and rarely evaluate agents on completing entire storylines. To address this, we introduce FlashAdventure, a benchmark of 34 Flash-based adventure games designed to test full story arc completion and tackle the observation-behavior gap: the challenge of remembering and acting on earlier gameplay information. We also propose CUA-as-a-Judge, an automated gameplay evaluator, and COAST, an agentic framework leveraging long-term clue memory to better plan and solve sequential tasks. Experiments show current GUI agents struggle with full story arcs, while COAST improves milestone completion by bridging the observation-behavior gap. Nonetheless, a marked discrepancy between humans and best-performing agents warrants continued research efforts to narrow this divide.

[628] VerlTool: Towards Holistic Agentic Reinforcement Learning with Tool Use

Dongfu Jiang, Yi Lu, Zhuofeng Li, Zhiheng Lyu, Ping Nie, Haozhe Wang, Alex Su, Hui Chen, Kai Zou, Chao Du, Tianyu Pang, Wenhu Chen

Main category: cs.AI

TL;DR: VerlTool is a unified framework that addresses limitations in Agentic Reinforcement Learning with Tool use (ARLT) by providing modular tool integration, asynchronous execution, and standardized APIs across multiple domains.

DetailsMotivation: Existing ARLT approaches suffer from fragmentation, synchronous execution bottlenecks, and limited extensibility across domains, hindering community adoption and algorithmic innovation.

Method: VerlTool introduces systematic design principles including upstream alignment with VeRL, unified tool management via standardized APIs, asynchronous rollout execution, and modular plugin architecture for rapid tool integration.

Result: The framework achieves near 2x speedup through asynchronous execution and demonstrates competitive performance across 6 ARLT domains including mathematical reasoning, knowledge QA, SQL generation, visual reasoning, web search, and software engineering.

Conclusion: VerlTool provides a scalable foundation for tool-augmented RL research with reduced development overhead, formalizing ARLT as multi-turn trajectories with multi-modal observations beyond single-turn RLVR paradigms.

Abstract: Reinforcement Learning with Verifiable Rewards (RLVR) has demonstrated success in enhancing LLM reasoning capabilities, but remains limited to single-turn interactions without tool integration. While recent Agentic Reinforcement Learning with Tool use (ARLT) approaches have emerged to address multi-turn tool interactions, existing works develop task-specific codebases that suffer from fragmentation, synchronous execution bottlenecks, and limited extensibility across domains. These inefficiencies hinder broader community adoption and algorithmic innovation. We introduce VerlTool, a unified and modular framework that addresses these limitations through systematic design principles. VerlTool provides four key contributions: (1) upstream alignment with VeRL ensuring compatibility and simplified maintenance, (2) unified tool management via standardized APIs supporting diverse modalities including code execution, search, SQL databases, and vision processing, (3) asynchronous rollout execution achieving near 2$\times$ speedup by eliminating synchronization bottlenecks, and (4) comprehensive evaluation demonstrating competitive performance across 6 ARLT domains. Our framework formalizes ARLT as multi-turn trajectories with multi-modal observation tokens (text/image/video), extending beyond single-turn RLVR paradigms. We train and evaluate models on mathematical reasoning, knowledge QA, SQL generation, visual reasoning, web search, and software engineering tasks, achieving results comparable to specialized systems while providing unified training infrastructure. The modular plugin architecture enables rapid tool integration requiring only lightweight Python definitions, significantly reducing development overhead and providing a scalable foundation for tool-augmented RL research. Our code is open-sourced at https://github.com/TIGER-AI-Lab/verl-tool.

[629] Robix: A Unified Model for Robot Interaction, Reasoning and Planning

Huang Fang, Mengxi Zhang, Heng Dong, Wei Li, Zixuan Wang, Qifeng Zhang, Xueyun Tian, Yucheng Hu, Hang Li

Main category: cs.AI

TL;DR: Robix is a unified vision-language model that integrates robot reasoning, task planning, and natural language interaction in a single architecture, outperforming commercial baselines like GPT-4o and Gemini 2.5 Pro in interactive task execution.

DetailsMotivation: To create a unified cognitive layer for robots that can handle complex instructions, plan long-horizon tasks, and interact naturally with humans within an end-to-end framework, addressing the need for more capable and interactive robotic systems.

Method: Uses chain-of-thought reasoning with a three-stage training strategy: (1) continued pretraining for embodied reasoning abilities, (2) supervised finetuning to model human-robot interaction and task planning as unified sequences, and (3) reinforcement learning for reasoning-action consistency.

Result: Robix outperforms both open-source and commercial baselines in interactive task execution, demonstrating strong generalization across diverse instruction types and various user-involved tasks like table bussing, grocery shopping, and dietary filtering.

Conclusion: Robix successfully integrates reasoning, planning, and interaction capabilities within a single vision-language architecture, enabling robots to handle complex real-world scenarios with natural human interaction and demonstrating superior performance compared to state-of-the-art models.

Abstract: We introduce Robix, a unified model that integrates robot reasoning, task planning, and natural language interaction within a single vision-language architecture. Acting as the high-level cognitive layer in a hierarchical robot system, Robix dynamically generates atomic commands for the low-level controller and verbal responses for human interaction, enabling robots to follow complex instructions, plan long-horizon tasks, and interact naturally with human within an end-to-end framework. Robix further introduces novel capabilities such as proactive dialogue, real-time interruption handling, and context-aware commonsense reasoning during task execution. At its core, Robix leverages chain-of-thought reasoning and adopts a three-stage training strategy: (1) continued pretraining to enhance foundational embodied reasoning abilities including 3D spatial understanding, visual grounding, and task-centric reasoning; (2) supervised finetuning to model human-robot interaction and task planning as a unified reasoning-action sequence; and (3) reinforcement learning to improve reasoning-action consistency and long-horizon task coherence. Extensive experiments demonstrate that Robix outperforms both open-source and commercial baselines (e.g., GPT-4o and Gemini 2.5 Pro) in interactive task execution, demonstrating strong generalization across diverse instruction types (e.g., open-ended, multi-stage, constrained, invalid, and interrupted) and various user-involved tasks such as table bussing, grocery shopping, and dietary filtering.

[630] Heads or Tails: A Simple Example of Causal Abstractive Simulation

Gabriel Simmons

Main category: cs.AI

TL;DR: This paper formalizes language model simulation using causal abstraction, specifically demonstrating how to simulate a fair coin toss with language models and showing both failure cases and success proofs.

DetailsMotivation: To provide a formal causal framework for language model simulation practices, connecting statistical benchmarking to causal foundations and offering precise operationalization for concepts like role-playing in AI.

Method: Uses causal abstractive simulation (a variation of causal abstraction) to analyze language model simulation, with concrete examples of simulating fair coin tosses and proving simulation capabilities.

Result: Demonstrates both failure modes and successful cases of language model simulation, showing how the formalism can prove when a language model properly simulates another system given causal descriptions.

Conclusion: Causal abstractive simulation provides a rigorous foundation for language model simulation that benefits practitioners, philosophers, and mathematicians by formalizing simulation concepts and connecting to causal theory.

Abstract: This note illustrates how a variety of causal abstraction arXiv:1707.00819 arXiv:1812.03789, defined here as causal abstractive simulation, can be used to formalize a simple example of language model simulation. This note considers the case of simulating a fair coin toss with a language model. Examples are presented illustrating the ways language models can fail to simulate, and a success case is presented, illustrating how this formalism may be used to prove that a language model simulates some other system, given a causal description of the system. This note may be of interest to three groups. For practitioners in the growing field of language model simulation, causal abstractive simulation is a means to connect ad-hoc statistical benchmarking practices to the solid formal foundation of causality. Philosophers of AI and philosophers of mind may be interested as causal abstractive simulation gives a precise operationalization to the idea that language models are role-playing arXiv:2402.12422. Mathematicians and others working on causal abstraction may be interested to see a new application of the core ideas that yields a new variation of causal abstraction.

[631] Towards Open-World Retrieval-Augmented Generation on Knowledge Graph: A Multi-Agent Collaboration Framework

Jiasheng Xu, Mingda Li, Yongqiang Tang, Peijie Wang, Wensheng Zhang

Main category: cs.AI

TL;DR: AnchorRAG is a multi-agent framework that enables open-world retrieval-augmented generation without predefined anchor entities, using dynamic entity prediction and parallel multi-hop exploration to overcome limitations of traditional KG-based RAG systems.

DetailsMotivation: Traditional KG-based RAG approaches require predefined anchor entities for graph traversal, which limits robustness in open-world settings where entity linking is unreliable. The authors aim to overcome this limitation by developing a system that can dynamically identify and work with candidate entities without pre-specified anchors.

Method: A multi-agent collaboration framework with three components: 1) Predictor agent dynamically identifies candidate anchor entities by aligning query terms with KG nodes, 2) Retriever agents conduct parallel multi-hop explorations from each candidate, and 3) Supervisor agent formulates retrieval strategy and synthesizes knowledge paths for final answer generation.

Result: Extensive experiments on four public benchmarks show that AnchorRAG significantly outperforms existing baselines and establishes new state-of-the-art results on real-world question answering tasks.

Conclusion: AnchorRAG provides a robust solution for open-world RAG by eliminating the dependency on predefined anchor entities through multi-agent collaboration, demonstrating superior performance and addressing the limitations of traditional KG-based retrieval approaches.

Abstract: Large Language Models (LLMs) have demonstrated strong capabilities in language understanding and reasoning. However, their dependence on static training corpora makes them prone to factual errors and knowledge gaps. Retrieval-Augmented Generation (RAG) addresses this limitation by incorporating external knowledge sources, especially structured Knowledge Graphs (KGs), which provide explicit semantics and efficient retrieval. Existing KG-based RAG approaches, however, generally assume that anchor entities are accessible to initiate graph traversal, which limits their robustness in open world settings where accurate linking between the query and the entity is unreliable. To overcome this limitation, we propose AnchorRAG, a novel multi-agent collaboration framework for open-world RAG without the predefined anchor entities. Specifically, a predictor agent dynamically identifies candidate anchor entities by aligning user query terms with KG nodes and initializes independent retriever agents to conduct parallel multi-hop explorations from each candidate. Then a supervisor agent formulates the iterative retrieval strategy for these retriever agents and synthesizes the resulting knowledge paths to generate the final answer. This multi-agent collaboration framework improves retrieval robustness and mitigates the impact of ambiguous or erroneous anchors. Extensive experiments on four public benchmarks demonstrate that AnchorRAG significantly outperforms existing baselines and establishes new state-of-the-art results on the real-world question answering tasks.

[632] Communicative Agents for Slideshow Storytelling Video Generation based on LLMs

Jingxing Fan, Jinrong Shen, Yusheng Yao, Shuangqing Wang, Qian Wang, Yuling Wang

Main category: cs.AI

TL;DR: VGTeam is a novel slide show video generation system that uses LLM-based communicative agents to create videos from text prompts with high efficiency and low cost ($0.103 per video, 98.4% success rate).

DetailsMotivation: Traditional text-to-video models suffer from high computational costs, limiting accessibility to video production. The paper aims to democratize video creation through an efficient LLM-based approach.

Method: VGTeam uses a team of communicative agents (scriptwriting, scene creation, audio design) working collaboratively in a chat tower workflow to transform text prompts into slide-style narrative videos, emulating traditional video production stages.

Result: The system achieves remarkable efficiency improvements with substantially reduced computational overhead - generates videos at $0.103 average cost with 98.4% success rate while maintaining high creative fidelity and customization.

Conclusion: VGTeam democratizes video production by enabling high-quality content creation without extensive resources, demonstrating the transformative potential of language models in creative domains as a pioneering next-generation content creation system.

Abstract: With the rapid advancement of artificial intelligence (AI), the proliferation of AI-generated content (AIGC) tasks has significantly accelerated developments in text-to-video generation. As a result, the field of video production is undergoing a transformative shift. However, conventional text-to-video models are typically constrained by high computational costs. In this study, we propose Video-Generation-Team (VGTeam), a novel slide show video generation system designed to redefine the video creation pipeline through the integration of large language models (LLMs). VGTeam is composed of a suite of communicative agents, each responsible for a distinct aspect of video generation, such as scriptwriting, scene creation, and audio design. These agents operate collaboratively within a chat tower workflow, transforming user-provided textual prompts into coherent, slide-style narrative videos. By emulating the sequential stages of traditional video production, VGTeam achieves remarkable improvements in both efficiency and scalability, while substantially reducing computational overhead. On average, the system generates videos at a cost of only $0.103, with a successful generation rate of 98.4%. Importantly, this framework maintains a high degree of creative fidelity and customization. The implications of VGTeam are far-reaching. It democratizes video production by enabling broader access to high-quality content creation without the need for extensive resources. Furthermore, it highlights the transformative potential of language models in creative domains and positions VGTeam as a pioneering system for next-generation content creation.

[633] GradeSQL: Outcome Reward Models for Ranking SQL Queries from Large Language Models

Mattia Tritto, Giuseppe Farano, Dario Di Palma, Gaetano Rossiello, Fedelucio Narducci, Dharmashankar Subramanian, Tommaso Di Noia

Main category: cs.AI

TL;DR: ORMs outperform traditional test-time strategies (ex-BoN and Maj) for Text-to-SQL by using semantic correctness scoring, achieving significant accuracy gains on BIRD and Spider benchmarks.

DetailsMotivation: Current LLMs struggle with complex Text-to-SQL queries requiring precise intent-schema alignment. Traditional methods like ex-BoN and Maj rely on surface-level heuristics rather than semantic understanding.

Method: Introduced a framework for training Outcome Reward Models (ORMs) that assign utility scores based on semantic correctness. Evaluated on BIRD and Spider benchmarks using various open-source LLMs including Qwen2, Granite3, and Llama3 families.

Result: ORMs achieved execution accuracy gains of +4.33% (BIRD) and +2.10% (Spider) over ex-BoN, and +2.91% (BIRD) and +0.93% (Spider) over Maj. Fine-tuned models like OmniSQL showed superior ORM performance.

Conclusion: ORMs provide a more effective heuristic for Best-of-N selection in Text-to-SQL, outperforming traditional methods by better aligning with semantic correctness and benefiting more from increased candidate numbers.

Abstract: Text-to-SQL, the task of translating natural language questions into SQL queries, has significantly advanced with the introduction of Large Language Models (LLMs), broadening database accessibility for a wide range of users. Despite substantial progress in generating valid SQL, current LLMs still struggle with complex queries that require precise alignment between user intent and the database schema. To mitigate this, test-time strategies such as Best-of-N (BoN) and Majority Voting (Maj) are often employed, based on the assumption that LLMs can generate correct answers but may require multiple attempts. However, these methods rely on surface-level heuristics, selecting either the syntactically correct query through execution-based BoN (ex-BoN) or the most frequently generated query with Maj. Recently, Outcome Reward Models (ORMs), which assign utility scores to generated outputs based on semantic correctness, have emerged as a promising approach for better aligning model predictions with user intent. Nevertheless, their application to Text-to-SQL remains largely underexplored. In this work, we evaluate ORMs as an effective heuristic for BoN, compare them with ex-BoN and Maj, and introduce a framework for training ORMs for the Text-to-SQL task. We evaluate our ORMs on the BIRD and SPIDER benchmarks, finetuning various open-source LLMs, including the Qwen2, Granite3, and Llama3 model families. Our results show that ORMs outperform ex-BoN and Maj, achieving execution accuracy gains of +4.33% (BIRD) and +2.10% (Spider) over ex-BoN, and +2.91% (BIRD) and +0.93% (Spider) over Maj. We further demonstrate that finetuning models already aligned with SQL generation, such as OmniSQL, yields superior ORM performance. Additionally, we observe that ORMs achieve competitive results on simple queries and benefit more from an increased number of candidates compared to ex-BoN and Maj.

[634] Conformal Predictive Monitoring for Multi-Modal Scenarios

Francesca Cairoli, Luca Bortolussi, Jyotirmoy V. Deshmukh, Lars Lindemann, Nicola Paoletti

Main category: cs.AI

TL;DR: GenQPM is a quantitative predictive monitoring method that uses deep generative models and mode classification to provide more informative, mode-specific prediction intervals for stochastic systems with multi-modal dynamics.

DetailsMotivation: Existing QPM methods produce overly conservative prediction intervals when systems exhibit multi-modal dynamics, as they treat all modes equally and lack mode-specific information.

Method: Leverages score-based diffusion models to approximate probabilistic multi-modal system dynamics, employs a mode classifier to partition trajectories by dynamical mode, and applies conformal inference for each mode to produce statistically valid prediction intervals.

Result: GenQPM produces significantly more informative (less conservative) prediction intervals compared to mode-agnostic baselines on agent navigation and autonomous driving benchmarks.

Conclusion: The proposed GenQPM method effectively addresses the limitations of existing QPM approaches by incorporating mode-specific analysis through deep generative models, resulting in more precise and useful prediction intervals for systems with complex multi-modal dynamics.

Abstract: We consider the problem of quantitative predictive monitoring (QPM) of stochastic systems, i.e., predicting at runtime the degree of satisfaction of a desired temporal logic property from the current state of the system. Since computational efficiency is key to enable timely intervention against predicted violations, several state-of-the-art QPM approaches rely on fast machine-learning surrogates to provide prediction intervals for the satisfaction values, using conformal inference to offer statistical guarantees. However, these QPM methods suffer when the monitored agent exhibits multi-modal dynamics, whereby certain modes may yield high satisfaction values while others critically violate the property. Existing QPM methods are mode-agnostic and so would yield overly conservative and uninformative intervals that lack meaningful mode-specific satisfaction information. To address this problem, we present GenQPM, a method that leverages deep generative models, specifically score-based diffusion models, to reliably approximate the probabilistic and multi-modal system dynamics without requiring explicit model access. GenQPM employs a mode classifier to partition the predicted trajectories by dynamical mode. For each mode, we then apply conformal inference to produce statistically valid, mode-specific prediction intervals. We demonstrate the effectiveness of GenQPM on a benchmark of agent navigation and autonomous driving tasks, resulting in prediction intervals that are significantly more informative (less conservative) than mode-agnostic baselines.

[635] Error Notebook-Guided, Training-Free Part Retrieval in 3D CAD Assemblies via Vision-Language Models

Yunqing Liu, Nan Zhang, Zhiming Tan

Main category: cs.AI

TL;DR: A novel part retrieval framework using Error Notebooks + RAG for improved CAD assembly part retrieval without extra training, achieving up to 23.4% accuracy improvement with GPT-4o.

DetailsMotivation: LLMs/VLMs face challenges with token limits and unsatisfactory performance in CAD part retrieval, plus fine-tuning is computationally expensive and unavailable for proprietary models.

Method: Error Notebooks construction (collecting erroneous CoTs with corrections) + RAG to retrieve specification-relevant records for refined prompt engineering.

Result: Substantial gains with proprietary models, GPT-4o achieving 23.4% absolute accuracy improvement; CoT reasoning benefits challenging cases with >10 parts.

Conclusion: The framework effectively handles 3D models with lengthy metadata without training, demonstrating significant performance improvements in CAD part retrieval.

Abstract: Effective specification-aware part retrieval within complex CAD assemblies is essential for automated design verification and downstream engineering tasks. However, directly using LLMs/VLMs to this task presents some challenges: the input sequences may exceed model token limits, and even after processing, performance remains unsatisfactory. Moreover, fine-tuning LLMs/VLMs requires significant computational resources, and for many high-performing general-use proprietary models (e.g., GPT or Gemini), fine-tuning access is not available. In this paper, we propose a novel part retrieval framework that requires no extra training, but using Error Notebooks + RAG for refined prompt engineering to help improve the existing general model’s retrieval performance. The construction of Error Notebooks consists of two steps: (1) collecting historical erroneous CoTs and their incorrect answers, and (2) connecting these CoTs through reflective corrections until the correct solutions are obtained. As a result, the Error Notebooks serve as a repository of tasks along with their corrected CoTs and final answers. RAG is then employed to retrieve specification-relevant records from the Error Notebooks and incorporate them into the inference process. Another major contribution of our work is a human-in-the-loop CAD dataset, which is used to evaluate our method. In addition, the engineering value of our novel framework lies in its ability to effectively handle 3D models with lengthy, non-natural language metadata. Experiments with proprietary models, including GPT-4o and the Gemini series, show substantial gains, with GPT-4o (Omni) achieving up to a 23.4% absolute accuracy improvement on the human preference dataset. Moreover, ablation studies confirm that CoT reasoning provides benefits especially in challenging cases with higher part counts (>10).

[636] DeepResearch Arena: The First Exam of LLMs’ Research Abilities via Seminar-Grounded Tasks

Haiyuan Wan, Chen Yang, Junchi Yu, Meiqi Tu, Jiaxuan Lu, Di Yu, Jianbao Cao, Ben Gao, Jiaqing Xie, Aoran Wang, Wenlong Zhang, Philip Torr, Dongzhan Zhou

Main category: cs.AI

TL;DR: DeepResearch Arena benchmark for evaluating deep research agents using academic seminar transcripts to create realistic research tasks across 12 disciplines.

DetailsMotivation: Current evaluation of deep research agents is challenging due to difficulty collecting frontier research questions that genuinely capture researchers' attention and intellectual curiosity.

Method: Multi-Agent Hierarchical Task Generation (MAHTG) system that extracts research-worthy inspirations from seminar transcripts and translates them into high-quality research tasks.

Result: Curated over 10,000 high-quality research tasks from 200+ academic seminars across 12 disciplines, presenting substantial challenges for state-of-the-art agents with clear performance gaps.

Conclusion: DeepResearch Arena provides a faithful benchmark grounded in real academic discourse that better reflects real-world research environments and reduces data leakage risks.

Abstract: Deep research agents have attracted growing attention for their potential to orchestrate multi-stage research workflows, spanning literature synthesis, methodological design, and empirical verification. Despite these strides, evaluating their research capability faithfully is rather challenging due to the difficulty of collecting frontier research questions that genuinely capture researchers’ attention and intellectual curiosity. To address this gap, we introduce DeepResearch Arena, a benchmark grounded in academic seminars that capture rich expert discourse and interaction, better reflecting real-world research environments and reducing the risk of data leakage. To automatically construct DeepResearch Arena, we propose a Multi-Agent Hierarchical Task Generation (MAHTG) system that extracts research-worthy inspirations from seminar transcripts. The MAHTG system further translates research-worthy inspirations into high-quality research tasks, ensuring the traceability of research task formulation while filtering noise. With the MAHTG system, we curate DeepResearch Arena with over 10,000 high-quality research tasks from over 200 academic seminars, spanning 12 disciplines, such as literature, history, and science. Our extensive evaluation shows that DeepResearch Arena presents substantial challenges for current state-of-the-art agents, with clear performance gaps observed across different models.

[637] The Need for Verification in AI-Driven Scientific Discovery

Cristina Cornelio, Takuya Ito, Ryan Cory-Wright, Sanjeeb Dash, Lior Horesh

Main category: cs.AI

TL;DR: AI accelerates scientific discovery by generating hypotheses at scale, but requires rigorous verification mechanisms to ensure scientific validity and prevent progress hindrance.

DetailsMotivation: The rapid generation of hypotheses by AI systems creates a critical need for scalable verification methods to ensure that AI-assisted discovery actually advances rather than hinders scientific progress.

Method: The paper examines historical scientific discovery development, reviews AI approaches including data-driven methods, knowledge-aware neural architectures, symbolic reasoning frameworks, and LLM agents for hypothesis generation and pattern discovery.

Result: AI systems can successfully uncover patterns and propose candidate scientific laws at unprecedented scale and speed, but their scientific value remains dependent on verification processes.

Conclusion: Rigorous and transparent verification must be the cornerstone of AI-assisted scientific discovery to ensure that the abundance of AI-generated hypotheses translates into meaningful scientific advancement rather than creating verification bottlenecks.

Abstract: Artificial intelligence (AI) is transforming the practice of science. Machine learning and large language models (LLMs) can generate hypotheses at a scale and speed far exceeding traditional methods, offering the potential to accelerate discovery across diverse fields. However, the abundance of hypotheses introduces a critical challenge: without scalable and reliable mechanisms for verification, scientific progress risks being hindered rather than being advanced. In this article, we trace the historical development of scientific discovery, examine how AI is reshaping established practices for scientific discovery, and review the principal approaches, ranging from data-driven methods and knowledge-aware neural architectures to symbolic reasoning frameworks and LLM agents. While these systems can uncover patterns and propose candidate laws, their scientific value ultimately depends on rigorous and transparent verification, which we argue must be the cornerstone of AI-assisted discovery.

[638] Counterfactual Sensitivity for Faithful Reasoning in Language Models

Ibne Farabi Shihab, Sanjeda Akter, Anuj Sharma

Main category: cs.AI

TL;DR: CSR is a lightweight training method that improves LLM reasoning faithfulness by penalizing models that produce correct answers from flawed reasoning traces through counterfactual interventions.

DetailsMotivation: LLMs often generate correct answers using flawed reasoning, which undermines trustworthiness in high-stakes applications where reliable reasoning processes are crucial.

Method: Counterfactual Sensitivity Regularization (CSR) introduces automated operator-level counterfactual interventions during training and penalizes models that maintain the same answer under logically invalid reasoning traces.

Result: CSR improves faithfulness by up to 70 percentage points over standard methods across arithmetic, logical deduction, and planning tasks, with minimal accuracy loss, and generalizes to larger models.

Conclusion: CSR effectively enhances reasoning faithfulness in LLMs through lightweight counterfactual training, and can be extended with semantic perturbations for commonsense reasoning tasks.

Abstract: Large language models (LLMs) often produce correct answers while relying on flawed or irrelevant reasoning traces, undermining their trustworthiness in high-stakes domains. We propose Counterfactual Sensitivity Regularization (CSR), a lightweight training objective that enforces dependence between intermediate reasoning and final outputs. CSR introduces automated, operator-level counterfactual interventions (e.g., swapping “+” with “-”) during training and penalizes models that preserve the same answer under logically invalid traces. This requires only one additional forward pass per sample. To measure faithfulness, we introduce Counterfactual Outcome Sensitivity (COS), which quantifies the impact of such perturbations on model predictions. Across structured reasoning tasks - arithmetic (GSM8K), logical deduction (PrOntoQA), and planning (Blocks World) - CSR improves faithfulness by up to 70 percentage points over standard fine-tuning and process supervision, with only minor accuracy loss. The learned sensitivity generalizes to larger models and synergizes with inference-time methods such as self-consistency. A pilot study on HellaSwag further demonstrates that extending CSR with semantic perturbations can enhance faithfulness in commonsense reasoning.

[639] Structured AI Decision-Making in Disaster Management

Julian Gerald Dcruz, Argyrios Zolotas, Niall Ross Greenwood, Miguel Arana-Catania

Main category: cs.AI

TL;DR: Proposes a structured decision-making framework for responsible AI in safety-critical domains, achieving 60.94% greater stability and 38.93% higher accuracy than human operators in disaster management scenarios.

DetailsMotivation: Address ethical implications and ensure reliable, justifiable decision-making in safety-critical AI applications like aerospace and emergency-response services where human lives are at stake.

Method: Developed a structured decision-making framework using Enabler agents, Levels and Scenarios, implemented in autonomous decision-making for disaster management. Compared against judgement-based systems and experienced human operators (victims, volunteers, stakeholders).

Result: Framework achieved 60.94% greater stability in consistently accurate decisions across multiple scenarios compared to judgement-based systems, and outperformed human operators with 38.93% higher accuracy across various scenarios.

Conclusion: The structured decision-making framework shows promise for building more reliable autonomous AI applications in safety-critical contexts, demonstrating significant improvements in decision stability and accuracy over both automated and human alternatives.

Abstract: With artificial intelligence (AI) being applied to bring autonomy to decision-making in safety-critical domains such as the ones typified in the aerospace and emergency-response services, there has been a call to address the ethical implications of structuring those decisions, so they remain reliable and justifiable when human lives are at stake. This paper contributes to addressing the challenge of decision-making by proposing a structured decision-making framework as a foundational step towards responsible AI. The proposed structured decision-making framework is implemented in autonomous decision-making, specifically within disaster management. By introducing concepts of Enabler agents, Levels and Scenarios, the proposed framework’s performance is evaluated against systems relying solely on judgement-based insights, as well as human operators who have disaster experience: victims, volunteers, and stakeholders. The results demonstrate that the structured decision-making framework achieves 60.94% greater stability in consistently accurate decisions across multiple Scenarios, compared to judgement-based systems. Moreover, the study shows that the proposed framework outperforms human operators with a 38.93% higher accuracy across various Scenarios. These findings demonstrate the promise of the structured decision-making framework for building more reliable autonomous AI applications in safety-critical contexts.

[640] Throttling Web Agents Using Reasoning Gates

Abhinav Kumar, Jaechul Roh, Ali Naseh, Amir Houmansadr, Eugene Bagdasarian

Main category: cs.AI

TL;DR: Web Agent Throttling framework imposes computational costs on AI agents through reasoning puzzles to prevent service overload and bypassing of web defenses.

DetailsMotivation: AI web agents can overload content providers, bypass CAPTCHAs, and flood authentication systems, requiring new protection mechanisms that don't rely on traditional human verification methods.

Method: Developed Throttling Gates using rebus-based Reasoning Gates - synthetic text puzzles requiring multi-hop reasoning over world knowledge, with scalable generation and verification protocols.

Result: Achieved 9.2x computational asymmetry (response cost vs generation cost for SOTA models), successfully deployed on custom websites and MCP servers, and evaluated with real-world web agents.

Conclusion: The framework effectively throttles web agents through computational costs but has limitations and environmental impact concerns that need consideration for real-world deployment.

Abstract: AI web agents use Internet resources at far greater speed, scale, and complexity – changing how users and services interact. Deployed maliciously or erroneously, these agents could overload content providers. At the same time, web agents can bypass CAPTCHAs and other defenses by mimicking user behavior or flood authentication systems with fake accounts. Yet providers must protect their services and content from denial-of-service attacks and scraping by web agents. In this paper, we design a framework that imposes tunable costs on agents before providing access to resources; we call this Web Agent Throttling. We start by formalizing Throttling Gates as challenges issued to an agent that are asymmetric, scalable, robust, and compatible with any agent. Focusing on a common component – the language model – we require the agent to solve reasoning puzzles, thereby incurring excessive token-generation costs. However, we find that using existing puzzles, e.g., coding or math, as throttling gates fails to satisfy our properties. To address this, we introduce rebus-based Reasoning Gates, synthetic text puzzles that require multi-hop reasoning over world knowledge (thereby throttling an agent’s model). We design a scalable generation and verification protocol for such reasoning gates. Our framework achieves computational asymmetry, i.e., the response-generation cost is 9.2x higher than the generation cost for SOTA models. We further deploy reasoning gates on a custom website and Model Context Protocol (MCP) servers and evaluate with real-world web agents. Finally, we discuss the limitations and environmental impact of real-world deployment of our framework.

[641] Unraveling LLM Jailbreaks Through Safety Knowledge Neurons

Chongwen Zhao, Kaizhu Huang

Main category: cs.AI

TL;DR: A neuron-level interpretability method that identifies safety-related knowledge neurons in LLMs and uses this insight to develop SafeTuning, a fine-tuning strategy that reinforces safety-critical neurons to defend against jailbreak attacks with over 97% effectiveness.

DetailsMotivation: Growing concern about malicious exploitation of LLMs for harmful purposes like synthesizing controlled substances and spreading disinformation through jailbreak attacks, with existing defense mechanisms lacking clear understanding of their rationale.

Method: Developed a novel neuron-level interpretability method that projects model’s internal representation into interpretable vocabulary space to identify safety-related knowledge neurons, then created SafeTuning - a fine-tuning strategy that reinforces these safety-critical neurons.

Result: Achieved mean Attack Success Rate (ASR) higher than 97% by adjusting safety-related neuron activations. SafeTuning consistently reduced attack success rates across multiple LLMs and outperformed all four baseline defenses.

Conclusion: Provides a new perspective on understanding and defending against jailbreak attacks by focusing on safety-critical neurons, offering an effective fine-tuning approach to improve model robustness against malicious exploitation.

Abstract: Large Language Models (LLMs) are increasingly attracting attention in various applications. Nonetheless, there is a growing concern as some users attempt to exploit these models for malicious purposes, including the synthesis of controlled substances and the propagation of disinformation, a technique known as “Jailbreak.” While some studies have achieved defenses against jailbreak attacks by modifying output distributions or detecting harmful content, the exact rationale still remains elusive. In this work, we present a novel neuron-level interpretability method that focuses on the role of safety-related knowledge neurons. Unlike existing approaches, our method projects the model’s internal representation into a more consistent and interpretable vocabulary space. We then show that adjusting the activation of safety-related neurons can effectively control the model’s behavior with a mean ASR higher than 97%. Building on this insight, we propose SafeTuning, a fine-tuning strategy that reinforces safety-critical neurons to improve model robustness against jailbreaks. SafeTuning consistently reduces attack success rates across multiple LLMs and outperforms all four baseline defenses. These findings offer a new perspective on understanding and defending against jailbreak attacks.

[642] Physics Supernova: AI Agent Matches Elite Gold Medalists at IPhO 2025

Jiahao Qiu, Jingzhe Shi, Xinzhe Juan, Zelin Zhao, Jiayi Geng, Shilong Liu, Hongru Wang, Sanfeng Wu, Mengdi Wang

Main category: cs.AI

TL;DR: Physics Supernova is an AI agent system that achieves elite-level physics problem-solving abilities comparable to top International Physics Olympiad gold medalists, scoring 23.5/30 points and ranking 14th among 406 contestants.

DetailsMotivation: AI systems need strong physics problem-solving capabilities to demonstrate real-world intelligence, as physics provides fundamental laws that describe and predict the natural world. The International Physics Olympiad serves as a rigorous benchmark for this purpose.

Method: The researchers developed Physics Supernova as an AI agent system with principled tool integration, designed to formulate and apply physical laws for explaining and predicting physical processes.

Result: Physics Supernova achieved 23.5/30 points in IPhO 2025 theory problems, ranking 14th of 406 contestants and surpassing the median performance of human gold medalists. Extensive analysis showed strong capabilities across diverse physics tasks.

Conclusion: Principled tool integration within agent systems can deliver competitive improvements in solving challenging science problems, demonstrating that AI can achieve elite-level physics problem-solving abilities comparable to top human performers.

Abstract: Physics provides fundamental laws that describe and predict the natural world. AI systems aspiring toward more general, real-world intelligence must therefore demonstrate strong physics problem-solving abilities: to formulate and apply physical laws for explaining and predicting physical processes. The International Physics Olympiad (IPhO)–the world’s most prestigious physics competition–offers a rigorous benchmark for this purpose. We introduce Physics Supernova, an AI agent system with superior physics problem-solving abilities that match elite IPhO gold medalists. In IPhO 2025 theory problems, Physics Supernova attains 23.5/30 points, ranking 14th of 406 contestants and surpassing the median performance of human gold medalists. We extensively analyzed Physics Supernova’s capabilities and flexibility across diverse physics tasks. These results show that principled tool integration within agent systems can deliver competitive improvements in solving challenging science problems. The codes are available at https://github.com/CharlesQ9/Physics-Supernova.

[643] An LLM-enabled semantic-centric framework to consume privacy policies

Rui Zhao, Vladyslav Melnychuk, Jun Zhao, Jesse Wright, Nigel Shadbolt

Main category: cs.AI

TL;DR: A semantic approach using LLMs to automatically extract privacy practices from policies and build knowledge graphs, enabling large-scale privacy policy analysis and formal policy generation.

DetailsMotivation: Users rarely read privacy policies due to complexity, creating barriers for user-centered web approaches and data sharing. Existing methods focus on compliance verification but lack scalable ways to create formal policies.

Method: Uses state-of-the-art LLMs to identify key privacy information, constructs Pr²Graph knowledge graph grounded in Data Privacy Vocabulary (DPV), and demonstrates conversion to formal policy representations like ODRL and psDToU.

Result: Created Pr²Graph for top-100 websites as public resource, enriched Policy-IE dataset with expert annotations, benchmarked LLM performance, and verified technology capabilities for privacy practice extraction.

Conclusion: Shows promise for large-scale analysis of online privacy practices, enabling web auditing. All datasets and source code released publicly to facilitate reuse and improvement.

Abstract: In modern times, people have numerous online accounts, but they rarely read the Terms of Service or Privacy Policy of those sites, despite claiming otherwise, due to the practical difficulty in comprehending them. The mist of data privacy practices forms a major barrier for user-centred Web approaches, and for data sharing and reusing in an agentic world. Existing research proposed methods for using formal languages and reasoning for verifying the compliance of a specified policy, as a potential cure for ignoring privacy policies. However, a critical gap remains in the creation or acquisition of such formal policies at scale. We present a semantic-centric approach for using state-of-the-art large language models (LLM), to automatically identify key information about privacy practices from privacy policies, and construct $\mathit{Pr}^2\mathit{Graph}$, knowledge graph with grounding from Data Privacy Vocabulary (DPV) for privacy practices, to support downstream tasks. Along with the pipeline, the $\mathit{Pr}^2\mathit{Graph}$ for the top-100 popular websites is also released as a public resource, by using the pipeline for analysis. We also demonstrate how the $\mathit{Pr}^2\mathit{Graph}$ can be used to support downstream tasks by constructing formal policy representations such as Open Digital Right Language (ODRL) or perennial semantic Data Terms of Use (psDToU). To evaluate the technology capability, we enriched the Policy-IE dataset by employing legal experts to create custom annotations. We benchmarked the performance of different large language models for our pipeline and verified their capabilities. Overall, they shed light on the possibility of large-scale analysis of online services’ privacy practices, as a promising direction to audit the Web and the Internet. We release all datasets and source code as public resources to facilitate reuse and improvement.

[644] Oyster-I: Beyond Refusal – Constructive Safety Alignment for Responsible Language Models

Ranjie Duan, Jiexi Liu, Xiaojun Jia, Shiji Zhao, Ruoxi Cheng, Fengxiang Wang, Cheng Wei, Yong Xie, Chang Liu, Defeng Li, Yinpeng Dong, Yichi Zhang, Yuefeng Chen, Chongwen Wang, Xingjun Ma, Xingxing Wei, Yang Liu, Hang Su, Jun Zhu, Xinfeng Li, Yitong Sun, Jie Zhang, Jinzhao Hu, Sha Xu, Yitong Yang, Jialing Tao, Hui Xue

Main category: cs.AI

TL;DR: CSA is a new safety paradigm that shifts from refusal-based responses to constructive guidance, especially for vulnerable users, achieving state-of-the-art safety while maintaining helpfulness.

DetailsMotivation: Current LLM safety approaches focus on adversarial risks but fail vulnerable users seeking help during psychological distress, where simple refusals can worsen outcomes.

Method: Constructive Safety Alignment (CSA) combines game-theoretic anticipation of user reactions, fine-grained risk boundary discovery, and interpretable reasoning control.

Result: Oy1 achieves SOTA safety among open models, strong constructive engagement close to GPT-5, and unmatched robustness on jailbreak datasets nearing GPT-o1 levels.

Conclusion: CSA redefines model-user relationships by shifting from refusal-first to guidance-first safety, creating systems that are both safe and meaningfully helpful.

Abstract: Large language models (LLMs) typically deploy safety mechanisms to prevent harmful content generation. Most current approaches focus narrowly on risks posed by malicious actors, often framing risks as adversarial events and relying on defensive refusals. However, in real-world settings, risks also come from non-malicious users seeking help while under psychological distress (e.g., self-harm intentions). In such cases, the model’s response can strongly influence the user’s next actions. Simple refusals may lead them to repeat, escalate, or move to unsafe platforms, creating worse outcomes. We introduce Constructive Safety Alignment (CSA), a human-centric paradigm that protects against malicious misuse while actively guiding vulnerable users toward safe and helpful results. Implemented in Oyster-I (Oy1), CSA combines game-theoretic anticipation of user reactions, fine-grained risk boundary discovery, and interpretable reasoning control, turning safety into a trust-building process. Oy1 achieves state-of-the-art safety among open models while retaining high general capabilities. On our Constructive Benchmark, it shows strong constructive engagement, close to GPT-5, and unmatched robustness on the Strata-Sword jailbreak dataset, nearing GPT-o1 levels. By shifting from refusal-first to guidance-first safety, CSA redefines the model-user relationship, aiming for systems that are not just safe, but meaningfully helpful. We release Oy1, code, and the benchmark to support responsible, user-centered AI.

[645] EigenBench: A Comparative Behavioral Measure of Value Alignment

Jonathn Chang, Leonard Piff, Suvadip Sana, Jasmine X. Li, Lionel Levine

Main category: cs.AI

TL;DR: EigenBench is a black-box benchmarking method that quantifies language models’ value alignment by having models judge each other’s outputs using EigenTrust aggregation, without requiring ground truth labels.

DetailsMotivation: Addressing the lack of quantitative metrics for AI value alignment by creating a method that can benchmark language models' values when reasonable judges may disagree on correct labels.

Method: Uses an ensemble of models, a constitution describing values, and scenario datasets. Models judge each other’s outputs, and EigenTrust aggregates these judgments to produce alignment scores without ground truth data.

Result: EigenBench scores are mostly sensitive to the prompt rather than the model itself, but a small residual quantifies the model’s inherent disposition.

Conclusion: EigenBench provides a practical method for quantitatively benchmarking value alignment in language models, demonstrating that prompts dominate alignment scores while still capturing model-specific dispositions.

Abstract: Aligning AI with human values is a pressing unsolved problem. To address the lack of quantitative metrics for value alignment, we propose EigenBench: a black-box method for comparatively benchmarking language models’ values. Given an ensemble of models, a constitution describing a value system, and a dataset of scenarios, our method returns a vector of scores quantifying each model’s alignment to the given constitution. To produce these scores, each model judges the outputs of other models across many scenarios, and these judgments are aggregated with EigenTrust (Kamvar et al, 2003), yielding scores that reflect a weighted-average judgment of the whole ensemble. EigenBench uses no ground truth labels, as it is designed to quantify traits for which reasonable judges may disagree on the correct label. Using prompted personas, we test whether EigenBench scores are more sensitive to the model or the prompt: we find that most of the variance is explained by the prompt, but a small residual quantifies the disposition of the model itself.

[646] mFARM: Towards Multi-Faceted Fairness Assessment based on HARMs in Clinical Decision Support

Shreyash Adappanavar, Krithi Shailya, Gokul S Krishnan, Sriraam Natarajan, Balaraman Ravindran

Main category: cs.AI

TL;DR: Proposes mFARM framework for multi-dimensional fairness assessment in medical LLMs, with benchmarks showing models maintain fairness under quantization but degrade with reduced context.

DetailsMotivation: Address AI alignment challenges in medical LLMs where existing fairness metrics overlook multi-dimensional medical harms and promote clinically inert models.

Method: Construct two large benchmarks from MIMIC-IV (ED-Triage and Opioid Recommendation) with 50k+ prompts across race/gender variants, and develop mFARM framework to assess allocational, stability, and latent disparities with aggregated fairness-accuracy scores.

Result: Evaluation of 4 open-source LLMs shows mFARM captures subtle biases effectively; models maintain robust fairness under quantization but deteriorate significantly with reduced context.

Conclusion: The mFARM framework provides comprehensive fairness assessment for medical LLMs, with benchmarks and code released to advance aligned AI in healthcare.

Abstract: The deployment of Large Language Models (LLMs) in high-stakes medical settings poses a critical AI alignment challenge, as models can inherit and amplify societal biases, leading to significant disparities. Existing fairness evaluation methods fall short in these contexts as they typically use simplistic metrics that overlook the multi-dimensional nature of medical harms. This also promotes models that are fair only because they are clinically inert, defaulting to safe but potentially inaccurate outputs. To address this gap, our contributions are mainly two-fold: first, we construct two large-scale, controlled benchmarks (ED-Triage and Opioid Analgesic Recommendation) from MIMIC-IV, comprising over 50,000 prompts with twelve race x gender variants and three context tiers. Second, we propose a multi-metric framework - Multi-faceted Fairness Assessment based on hARMs ($mFARM$) to audit fairness for three distinct dimensions of disparity (Allocational, Stability, and Latent) and aggregate them into an $mFARM$ score. We also present an aggregated Fairness-Accuracy Balance (FAB) score to benchmark and observe trade-offs between fairness and prediction accuracy. We empirically evaluate four open-source LLMs (Mistral-7B, BioMistral-7B, Qwen-2.5-7B, Bio-LLaMA3-8B) and their finetuned versions under quantization and context variations. Our findings showcase that the proposed $mFARM$ metrics capture subtle biases more effectively under various settings. We find that most models maintain robust performance in terms of $mFARM$ score across varying levels of quantization but deteriorate significantly when the context is reduced. Our benchmarks and evaluation code are publicly released to enhance research in aligned AI for healthcare.

[647] Generative KI für TA

Wolfgang Eppler, Reinhard Heil

Main category: cs.AI

TL;DR: Technology assessment professionals use generative AI both as a tool for their work and as a subject of research, while addressing its structural risks and proposing solutions.

DetailsMotivation: To explore how technology assessment (TA) professionals can effectively utilize generative AI while addressing its inherent structural problems and risks.

Method: The paper outlines the phenomenon of generative AI, formulates requirements for its use in TA, analyzes structural causes of associated problems, and proposes solutions with feasibility assessments.

Result: The analysis reveals that while generative AI continues to develop, structurally induced risks persist, requiring specific solutions for safe and effective implementation in technology assessment work.

Conclusion: Generative AI presents both opportunities and challenges for technology assessment, with persistent structural risks that need targeted solutions, though the technology shows promise when implemented with appropriate safeguards.

Abstract: Many scientists use generative AI in their scientific work. People working in technology assessment (TA) are no exception. TA’s approach to generative AI is twofold: on the one hand, generative AI is used for TA work, and on the other hand, generative AI is the subject of TA research. After briefly outlining the phenomenon of generative AI and formulating requirements for its use in TA, the following article discusses in detail the structural causes of the problems associated with it. Although generative AI is constantly being further developed, the structurally induced risks remain. The article concludes with proposed solutions and brief notes on their feasibility, as well as some examples of the use of generative AI in TA work.

[648] AGI as Second Being: The Structural-Generative Ontology of Intelligence

Maijunxian Wang, Ran Ji

Main category: cs.AI

TL;DR: The paper proposes a Structural-Generative Ontology of Intelligence, arguing that true intelligence requires three conditions: generativity (creating new structures), coordination (organizing them into reasons), and sustaining (maintaining identity over time). Current AI systems lack this depth despite broad capabilities.

DetailsMotivation: To address the limitation that current AI systems, while capable of performing many tasks, remain superficial simulations without genuine intelligence. The paper aims to define what constitutes real intelligence beyond mere functional breadth.

Method: The paper proposes a theoretical framework called Structural-Generative Ontology of Intelligence, which establishes three essential conditions for true intelligence: generativity, coordination, and sustaining of identity over time.

Result: The framework demonstrates that current AI systems lack the depth required for genuine intelligence despite their broad functional capabilities. It provides criteria to distinguish between surface-level simulation and true intelligence.

Conclusion: True intelligence emerges from depth (generativity, coordination, and sustaining) rather than breadth of function. Future systems meeting these conditions could represent a “Second Being” distinct from human existence, moving beyond being mere tools.

Abstract: Artificial intelligence is often measured by the range of tasks it can perform. Yet wide ability without depth remains only an imitation. This paper proposes a Structural-Generative Ontology of Intelligence: true intelligence exists only when a system can generate new structures, coordinate them into reasons, and sustain its identity over time. These three conditions – generativity, coordination, and sustaining – define the depth that underlies real intelligence. Current AI systems, however broad in function, remain surface simulations because they lack this depth. Breadth is not the source of intelligence but the growth that follows from depth. If future systems were to meet these conditions, they would no longer be mere tools, but could be seen as a possible Second Being, standing alongside yet distinct from human existence.

Strahinja Klem, Noura Al Moubayed

Main category: cs.AI

TL;DR: Structured prompting methodology using QWEN-2 LLM for legal document information retrieval, achieving 9% improvement over previous methods with chunking, prompt engineering, and novel heuristics.

DetailsMotivation: Address reliability and transparency challenges in applying LLMs to legal domains by developing a structured alternative to expensive fine-tuning that can handle long legal documents.

Method: Document chunking and augmentation to handle long documents, engineered prompts fed into QWEN-2 model, and Distribution-based Localisation + Inverse Cardinality Weighting heuristics for candidate selection.

Result: Achieved state-of-the-art performance with up to 9% improvement over previous methods on CUAD dataset for legal information retrieval.

Conclusion: Structured prompt engineering shows promise as an under-explored tool for ensuring AI accountability in legal domains, though current automatic evaluation metrics remain limiting.

Abstract: The rise of Large Language Models (LLMs) has had a profoundly transformative effect on a number of fields and domains. However, their uptake in Law has proven more challenging due to the important issues of reliability and transparency. In this study, we present a structured prompting methodology as a viable alternative to the often expensive fine-tuning, with the capability of tacking long legal documents from the CUAD dataset on the task of information retrieval. Each document is first split into chunks via a system of chunking and augmentation, addressing the long document problem. Then, alongside an engineered prompt, the input is fed into QWEN-2 to produce a set of answers for each question. Finally, we tackle the resulting candidate selection problem with the introduction of the Distribution-based Localisation and Inverse Cardinality Weighting heuristics. This approach leverages a general purpose model to promote long term scalability, prompt engineering to increase reliability and the two heuristic strategies to reduce the impact of the black box effect. Whilst our model performs up to 9% better than the previously presented method, reaching state-of-the-art performance, it also highlights the limiting factor of current automatic evaluation metrics for question answering, serving as a call to action for future research. However, the chief aim of this work is to underscore the potential of structured prompt engineering as a useful, yet under-explored, tool in ensuring accountability and responsibility of AI in the legal domain, and beyond.

[650] An Epidemiological Knowledge Graph extracted from the World Health Organization’s Disease Outbreak News

Sergio Consoli, Pietro Coletti, Peter V. Markov, Lia Orfei, Indaco Biazzo, Lea Schuh, Nicolas Stefanovitch, Lorenzo Bertolini, Mario Ceresa, Nikolaos I. Stilianakis

Main category: cs.AI

TL;DR: Researchers developed an AI-powered system using multiple LLMs to extract epidemiological data from WHO Disease Outbreak News, creating a daily-updated dataset and knowledge graph (eKG) for enhanced disease surveillance and research.

DetailsMotivation: To leverage AI advancements and available social media/news data to improve epidemiological surveillance and extract actionable information from WHO outbreak reports for better public health decision-making.

Method: Used an ensemble approach incorporating multiple Large Language Models to process WHO Disease Outbreak News reports, extracting valuable epidemiological information and constructing a knowledge graph (eKG) with daily updates.

Result: Created a comprehensive daily-updated dataset and knowledge graph (eKG) that provides nuanced representation of public health domain knowledge from WHO outbreak reports, enabling new epidemiological research opportunities.

Conclusion: The developed AI-powered system and data resources open new possibilities for epidemiological research, disease outbreak analysis, and enhanced surveillance capabilities through automated extraction of actionable information from authoritative health reports.

Abstract: The rapid evolution of artificial intelligence (AI), together with the increased availability of social media and news for epidemiological surveillance, are marking a pivotal moment in epidemiology and public health research. Leveraging the power of generative AI, we use an ensemble approach which incorporates multiple Large Language Models (LLMs) to extract valuable actionable epidemiological information from the World Health Organization (WHO) Disease Outbreak News (DONs). DONs is a collection of regular reports on global outbreaks curated by the WHO and the adopted decision-making processes to respond to them. The extracted information is made available in a daily-updated dataset and a knowledge graph, referred to as eKG, derived to provide a nuanced representation of the public health domain knowledge. We provide an overview of this new dataset and describe the structure of eKG, along with the services and tools used to access and utilize the data that we are building on top. These innovative data resources open altogether new opportunities for epidemiological research, and the analysis and surveillance of disease outbreaks.

[651] Rewarding Explainability in Drug Repurposing with Knowledge Graphs

Susana Nunes, Samy Badreddine, Catia Pesquita

Main category: cs.AI

TL;DR: REx is a reinforcement learning approach that generates scientific explanations for knowledge graph link predictions, particularly in drug repurposing, by identifying explanatory paths enriched with biomedical ontologies.

DetailsMotivation: Knowledge graphs are powerful for modeling complex data but need to provide meaningful scientific explanations alongside accurate predictions to gain acceptance as credible scientific tools.

Method: Uses reward and policy mechanisms in reinforcement learning to identify explanatory paths within knowledge graphs, enriched with domain-specific ontologies to ensure explanations are grounded in biomedical knowledge.

Result: Outperforms state-of-the-art approaches in predictive performance on three knowledge graph benchmarks and generates explanations that validate predictive insights against biomedical knowledge.

Conclusion: REx represents a significant contribution to advancing AI-driven scientific discovery by providing both accurate predictions and scientifically meaningful explanations.

Abstract: Knowledge graphs (KGs) are powerful tools for modelling complex, multi-relational data and supporting hypothesis generation, particularly in applications like drug repurposing. However, for predictive methods to gain acceptance as credible scientific tools, they must ensure not only accuracy but also the capacity to offer meaningful scientific explanations. This paper presents a novel approach REx, for generating scientific explanations based in link prediction in knowledge graphs. It employs reward and policy mechanisms that consider desirable properties of scientific explanation to guide a reinforcement learning agent in the identification of explanatory paths within a KG. The approach further enriches explanatory paths with domain-specific ontologies, ensuring that the explanations are both insightful and grounded in established biomedical knowledge. We evaluate our approach in drug repurposing using three popular knowledge graph benchmarks. The results clearly demonstrate its ability to generate explanations that validate predictive insights against biomedical knowledge and that outperform the state-of-the-art approaches in predictive performance, establishing REx as a relevant contribution to advance AI-driven scientific discovery.

[652] Re-evaluating LLM-based Heuristic Search: A Case Study on the 3D Packing Problem

Guorui Quan, Mingfei Sun, Manuel López-Ibáñez

Main category: cs.AI

TL;DR: LLMs can generate heuristic code but struggle with complete solver design. With constraint scaffolding and iterative self-correction, LLMs focused mainly on scoring functions, producing results comparable to human-designed algorithms but showing limitations in complex constraint handling.

DetailsMotivation: To investigate whether LLMs can go beyond simple code generation and demonstrate broader innovation in automated heuristic design, specifically for complex problems like the constrained 3D Packing Problem.

Method: Used constraint scaffolding (prewritten constraint-checking code) and iterative self-correction (refinement cycles to repair bugs) to support LLM in building a complete solver. The LLM primarily focused on refining scoring functions within a greedy process.

Result: The LLM-generated heuristic performed comparably to human-designed greedy algorithms. When integrated into a human-crafted metaheuristic, it rivaled established solvers, though performance decreased with tighter constraints.

Conclusion: Current LLMs face two major barriers in automated heuristic design: engineering requirements to mitigate fragility in complex reasoning, and pretrained biases that prematurely narrow solution search, limiting broader innovation.

Abstract: The art of heuristic design has traditionally been a human pursuit. While Large Language Models (LLMs) can generate code for search heuristics, their application has largely been confined to adjusting simple functions within human-crafted frameworks, leaving their capacity for broader innovation an open question. To investigate this, we tasked an LLM with building a complete solver for the constrained 3D Packing Problem. Direct code generation quickly proved fragile, prompting us to introduce two supports: constraint scaffolding–prewritten constraint-checking code–and iterative self-correction–additional refinement cycles to repair bugs and produce a viable initial population. Notably, even within a vast search space in a greedy process, the LLM concentrated its efforts almost exclusively on refining the scoring function. This suggests that the emphasis on scoring functions in prior work may reflect not a principled strategy, but rather a natural limitation of LLM capabilities. The resulting heuristic was comparable to a human-designed greedy algorithm, and when its scoring function was integrated into a human-crafted metaheuristic, its performance rivaled established solvers, though its effectiveness waned as constraints tightened. Our findings highlight two major barriers to automated heuristic design with current LLMs: the engineering required to mitigate their fragility in complex reasoning tasks, and the influence of pretrained biases, which can prematurely narrow the search for novel solutions.

[653] Exploring Diffusion Models for Generative Forecasting of Financial Charts

Taegyeong Lee, Jiwon Park, Kyunga Bang, Seunghyun Hwang, Ung-Jin Jang

Main category: cs.AI

TL;DR: Using text-to-image diffusion models to predict stock price trends by treating time-series data as images and generating future chart patterns from current charts with instruction prompts.

DetailsMotivation: Financial domain relies heavily on time-series data and transformer models, lacking diverse generative model applications. Current methods focus on classifying chart patterns rather than generating future trends.

Method: Proposes treating time-series data as image patterns and using diffusion models to generate next chart image from current chart with instruction prompts. Introduces evaluation method for generated charts against ground truth.

Result: Demonstrates potential of text-to-image generative models in financial applications for stock price trend prediction.

Conclusion: Highlights the viability of using generative models in finance and motivates further research to address limitations and expand applicability beyond traditional methods.

Abstract: Recent advances in generative models have enabled significant progress in tasks such as generating and editing images from text, as well as creating videos from text prompts, and these methods are being applied across various fields. However, in the financial domain, there may still be a reliance on time-series data and a continued focus on transformer models, rather than on diverse applications of generative models. In this paper, we propose a novel approach that leverages text-to-image model by treating time-series data as a single image pattern, thereby enabling the prediction of stock price trends. Unlike prior methods that focus on learning and classifying chart patterns using architectures such as ResNet or ViT, we experiment with generating the next chart image from the current chart image and an instruction prompt using diffusion models. Furthermore, we introduce a simple method for evaluating the generated chart image against ground truth image. We highlight the potential of leveraging text-to-image generative models in the financial domain, and our findings motivate further research to address the current limitations and expand their applicability.

[654] Explainability-Driven Dimensionality Reduction for Hyperspectral Imaging

Salma Haidar, José Oramas

Main category: cs.AI

TL;DR: Using explainability methods to guide band selection in hyperspectral imaging, achieving comparable accuracy with only 30 selected bands instead of full spectrum.

DetailsMotivation: Hyperspectral imaging has high dimensionality causing computational burden and redundancy, making dimensionality reduction essential while preserving predictive performance.

Method: Apply post-hoc explainability methods to quantify each band’s contribution to classifier decisions, perform deletion-insertion evaluations to aggregate influence scores, and select highest-influence bands.

Result: Classifiers trained on 30 selected bands match or exceed full-spectrum baselines on Pavia University and Salinas benchmarks, reducing computational requirements while maintaining accuracy.

Conclusion: Model-aligned, explanation-guided band selection is a principled approach for effective dimensionality reduction in hyperspectral imaging, producing compact spectral subsets that align with physically meaningful wavelength regions.

Abstract: Hyperspectral imaging (HSI) provides rich spectral information for precise material classification and analysis; however, its high dimensionality introduces a computational burden and redundancy, making dimensionality reduction essential. We present an exploratory study into the application of post-hoc explainability methods in a model–driven framework for band selection, which reduces the spectral dimension while preserving predictive performance. A trained classifier is probed with explanations to quantify each band’s contribution to its decisions. We then perform deletion–insertion evaluations, recording confidence changes as ranked bands are removed or reintroduced, and aggregate these signals into influence scores. Selecting the highest–influence bands yields compact spectral subsets that maintain accuracy and improve efficiency. Experiments on two public benchmarks (Pavia University and Salinas) demonstrate that classifiers trained on as few as 30 selected bands match or exceed full–spectrum baselines while reducing computational requirements. The resulting subsets align with physically meaningful, highly discriminative wavelength regions, indicating that model–aligned, explanation-guided band selection is a principled route to effective dimensionality reduction for HSI.

[655] When Agents go Astray: Course-Correcting SWE Agents with PRMs

Shubham Gandhi, Jason Tsay, Jatin Ganhotra, Kiran Kate, Yara Rizk

Main category: cs.AI

TL;DR: SWE-PRM is a process reward model that detects and corrects inefficiencies in LLM agent trajectories during execution, improving software engineering task success rates by 10.6 percentage points with minimal added cost.

DetailsMotivation: LLM agents deployed for complex software engineering tasks often exhibit costly inefficiencies like redundant exploration, looping, and failure to terminate, which prior work only addressed post-execution.

Method: Introduces SWE-PRM, an inference-time Process Reward Model that intervenes during execution using a taxonomy of common inefficiencies to provide lightweight, interpretable feedback without modifying the underlying policy.

Result: On SWE-bench Verified, closed-source PRMs improved resolution from 40.0% to 50.6% (+10.6 p.p.), with largest gains on medium/hard tasks. Taxonomy-guided PRMs outperformed unguided variants, increasing success while reducing trajectory length.

Conclusion: PRMs provide a practical and scalable mechanism for improving SWE agents’ reliability and efficiency with acceptable added inference cost (as low as $0.2), making them suitable for real-world deployment.

Abstract: Large Language Model (LLM) agents are increasingly deployed for complex, multi-step software engineering (SWE) tasks. However, their trajectories often contain costly inefficiencies, such as redundant exploration, looping, and failure to terminate once a solution is reached. Prior work has largely treated these errors in a post-hoc manner, diagnosing failures only after execution. In this paper, we introduce SWE-PRM, an inference-time Process Reward Model (PRM) that intervenes during execution to detect and course-correct trajectory-level errors. Our PRM design leverages a taxonomy of common inefficiencies and delivers lightweight, interpretable feedback without modifying the underlying policy. On SWE-bench Verified, closed-source PRMs improve resolution from 40.0% to 50.6% (+10.6 p.p.), with the largest gains on medium and hard tasks. Among feedback strategies, taxonomy-guided PRMs outperform unguided or explicit action-prescriptive variants, increasing success rate while reducing trajectory length. These benefits come at an acceptable added inference cost of as low as $0.2, making PRMs a practical and scalable mechanism for improving SWE agents’ reliability and efficiency.

[656] Towards Agents That Know When They Don’t Know: Uncertainty as a Control Signal for Structured Reasoning

Josefa Lia Stoisser, Marc Boubnovski Martell, Lawrence Phillips, Gianluca Mazzoni, Lea Mørch Harder, Philip Torr, Jesper Ferkinghoff-Borg, Kaspar Martens, Julien Fauqueur

Main category: cs.AI

TL;DR: Uncertainty-aware LLM agent improves biomedical data summarization by leveraging retrieval and summary uncertainty signals, achieving significant improvements in factuality and downstream prediction tasks.

DetailsMotivation: LLM agents often produce overconfident outputs when reasoning over complex multi-table biomedical data, requiring better uncertainty quantification to improve reliability.

Method: Uses two uncertainty signals: retrieval uncertainty (entropy over table-selection rollouts) and summary uncertainty (self-consistency + perplexity). Incorporates summary uncertainty into RL with Group Relative Policy Optimization (GRPO) and uses both for inference-time filtering and dataset construction.

Result: Nearly triples correct claims per summary (3.0→8.4 internal; 3.6→9.9 cancer multi-omics) and substantially improves survival prediction (C-index 0.32→0.63).

Conclusion: Uncertainty serves as an effective control signal, enabling agents to abstain, communicate confidence, and become more reliable tools for complex structured-data environments.

Abstract: Large language model (LLM) agents are increasingly deployed in structured biomedical data environments, yet they often produce fluent but overconfident outputs when reasoning over complex multi-table data. We introduce an uncertainty-aware agent for query-conditioned multi-table summarization that leverages two complementary signals: (i) retrieval uncertainty–entropy over multiple table-selection rollouts–and (ii) summary uncertainty–combining self-consistency and perplexity. Summary uncertainty is incorporated into reinforcement learning (RL) with Group Relative Policy Optimization (GRPO), while both retrieval and summary uncertainty guide inference-time filtering and support the construction of higher-quality synthetic datasets. On multi-omics benchmarks, our approach improves factuality and calibration, nearly tripling correct and useful claims per summary (3.0(\rightarrow)8.4 internal; 3.6(\rightarrow)9.9 cancer multi-omics) and substantially improving downstream survival prediction (C-index 0.32(\rightarrow)0.63). These results demonstrate that uncertainty can serve as a control signal–enabling agents to abstain, communicate confidence, and become more reliable tools for complex structured-data environments.

[657] AppCopilot: Toward General, Accurate, Long-Horizon, and Efficient Mobile Agent

Jingru Fan, Yufan Dang, Jingyao Wu, Huatao Li, Runde Yang, Xiyuan Yang, Yuheng Wang, Zhong Zhang, Yaxi Lu, Yankai Lin, Zhiyuan Liu, Dahai Li, Chen Qian

Main category: cs.AI

TL;DR: AppCopilot is a multimodal, multi-agent on-device assistant that addresses four core mobile agent challenges: generalization, accuracy, long-horizon capability, and efficiency through an end-to-end system with multimodal foundation models, multi-agent collaboration, and hardware optimization.

DetailsMotivation: The rapid evolution of LLMs and multimodal models has created a fragmented mobile-agent landscape without addressing fundamental challenges. The paper aims to solve four core problems for practical mobile agents: generalization across tasks/modalities/apps/devices, precise interaction accuracy, long-horizon task capability, and efficient runtime on constrained devices.

Method: AppCopilot uses an end-to-end autonomous pipeline with multimodal foundation models (Chinese-English support), chain-of-thought reasoning, hierarchical task planning, multi-agent collaboration, and execution layer features including personalization, voice interaction, function calling, and cross-app orchestration. The system incorporates profiling-driven optimization for latency, memory, and energy across heterogeneous hardware.

Result: Empirical results show AppCopilot achieves significant improvements across all four dimensions: stronger generalization, higher-precision on-screen actions, more reliable long-horizon task completion, and faster, more resource-efficient runtime.

Conclusion: AppCopilot presents a comprehensive solution to the fundamental challenges in mobile agents, delivering a full-stack, closed-loop system that operationalizes practical mobile assistance through integrated multimodal models, multi-agent collaboration, and hardware-aware optimization.

Abstract: With the raid evolution of large language models and multimodal foundation models, the mobile-agent landscape has proliferated without converging on the fundamental challenges. This paper identifies four core problems that must be solved for mobile agents to deliver practical, scalable impact: (1) generalization across tasks, modalities, apps, and devices; (2) accuracy, specifically precise on-screen interaction and click targeting; (3) long-horizon capability for sustained, multi-step goals; and (4) efficiency, specifically high-performance runtime on resource-constrained devices. We present AppCopilot, a multimodal, multi-agent, general-purpose on-device assistant that operates across applications and constitutes a full-stack, closed-loop system from data to deployment. AppCopilot operationalizes this position through an end-to-end autonomous pipeline spanning data collection, training, deployment, high-quality and efficient inference, and mobile application development. At the model layer, it integrates multimodal foundation models with robust Chinese-English support. At the reasoning and control layer, it combines chain-of-thought reasoning, hierarchical task planning and decomposition, and multi-agent collaboration. At the execution layer, it enables user personalization and experiential adaptation, voice interaction, function calling, cross-app and cross-device orchestration, and comprehensive mobile app support. The system design incorporates profiling-driven optimization for latency, memory, and energy across heterogeneous hardware. Empirically, AppCopilot achieves significant improvements along all four dimensions: stronger generalization, higher-precision on-screen actions, more reliable long-horizon task completion, and faster, more resource-efficient runtime.

[658] GridMind: LLMs-Powered Agents for Power System Analysis and Operations

Hongwei Jin, Kibaek Kim, Jonghwan Kwon

Main category: cs.AI

TL;DR: GridMind is a multi-agent AI system that combines LLMs with engineering solvers to enable conversational power system analysis through natural language interfaces while maintaining numerical precision.

DetailsMotivation: Traditional power system analysis workflows are complex and create barriers to efficient decision-making in modern electric grids.

Method: Uses specialized AI agents coordinating AC Optimal Power Flow and N-1 contingency analysis through natural language interfaces with deterministic engineering solvers via function calls.

Result: Experimental evaluation on IEEE test cases shows the framework delivers correct solutions across all tested language models, with smaller LLMs achieving comparable accuracy with reduced latency.

Conclusion: Agentic AI is a viable paradigm for scientific computing, demonstrating conversational interfaces can enhance accessibility while preserving numerical rigor for critical engineering applications.

Abstract: The complexity of traditional power system analysis workflows presents significant barriers to efficient decision-making in modern electric grids. This paper presents GridMind, a multi-agent AI system that integrates Large Language Models (LLMs) with deterministic engineering solvers to enable conversational scientific computing for power system analysis. The system employs specialized agents coordinating AC Optimal Power Flow and N-1 contingency analysis through natural language interfaces while maintaining numerical precision via function calls. GridMind addresses workflow integration, knowledge accessibility, context preservation, and expert decision-support augmentation. Experimental evaluation on IEEE test cases demonstrates that the proposed agentic framework consistently delivers correct solutions across all tested language models, with smaller LLMs achieving comparable analytical accuracy with reduced computational latency. This work establishes agentic AI as a viable paradigm for scientific computing, demonstrating how conversational interfaces can enhance accessibility while preserving numerical rigor essential for critical engineering applications.

[659] UI-TARS-2 Technical Report: Advancing GUI Agent with Multi-Turn Reinforcement Learning

Haoming Wang, Haoyang Zou, Huatong Song, Jiazhan Feng, Junjie Fang, Junting Lu, Longxiang Liu, Qinyu Luo, Shihao Liang, Shijue Huang, Wanjun Zhong, Yining Ye, Yujia Qin, Yuwen Xiong, Yuxin Song, Zhiyong Wu, Bo Li, Chen Dun, Chong Liu, Fuxing Leng, Hanbin Wang, Hao Yu, Haobin Chen, Hongyi Guo, Jing Su, Jingjia Huang, Kai Shen, Kaiyu Shi, Lin Yan, Peiyao Zhao, Pengfei Liu, Qinghao Ye, Renjie Zheng, Wayne Xin Zhao, Wen Heng, Wenhao Huang, Wenqian Wang, Xiaobo Qin, Yi Lin, Youbin Wu, Zehui Chen, Zihao Wang, Baoquan Zhong, Xinchun Zhang, Xujing Li, Yuanfan Li, Zhongkai Zhao, Chengquan Jiang, Faming Wu, Haotian Zhou, Jinlin Pang, Li Han, Qianli Ma, Siyao Liu, Songhua Cai, Wenqi Fu, Xin Liu, Zhi Zhang, Bo Zhou, Guoliang Li, Jiajun Shi, Jiale Yang, Jie Tang, Li Li, Taoran Lu, Woyu Lin, Xiaokang Tong, Xinyao Li, Yichi Zhang, Yu Miao, Zhengxuan Jiang, Zili Li, Ziyuan Zhao, Chenxin Li, Dehua Ma, Feng Lin, Ge Zhang, Haihua Yang, Hangyu Guo, Hongda Zhu, Jiaheng Liu, Junda Du, Kai Cai, Kuanye Li, Lichen Yuan, Meilan Han, Minchao Wang, Shuyue Guo, Tianhao Cheng, Xiaobo Ma, Xiaojun Xiao, Xiaolong Huang, Xinjie Chen, Yidi Du, Yilin Chen, Yiwen Wang, Zhaojian Li, Zhenzhu Yang, Zhiyuan Zeng, Chaolin Jin, Chen Li, Hao Chen, Haoli Chen, Jian Chen, Qinghao Zhao, Guang Shi

Main category: cs.AI

TL;DR: UI-TARS-2 is a native GUI agent model that addresses key challenges in autonomous GUI agents through scalable data generation, multi-turn RL stabilization, hybrid environment integration, and unified sandbox platform, achieving state-of-the-art performance across multiple benchmarks.

DetailsMotivation: Address challenges in GUI autonomous agents including data scalability, multi-turn reinforcement learning limitations, GUI-only operation constraints, and environment stability issues.

Method: Systematic training methodology with data flywheel for scalable data generation, stabilized multi-turn RL framework, hybrid GUI environment integrating file systems and terminals, and unified sandbox platform for large-scale rollouts.

Result: Achieves 88.2 on Online-Mind2Web, 47.5 on OSWorld, 50.6 on WindowsAgentArena, 73.3 on AndroidWorld, outperforms Claude and OpenAI agents. Attains 59.8 mean normalized score across 15-game suite (~60% human performance), competitive with frontier models like OpenAI o3 on LMGame-Bench.

Conclusion: UI-TARS-2 demonstrates significant improvements over predecessor, strong generalization to diverse agent tasks, and potential to advance GUI agent state with robustness in real-world interactive scenarios.

Abstract: The development of autonomous agents for graphical user interfaces (GUIs) presents major challenges in artificial intelligence. While recent advances in native agent models have shown promise by unifying perception, reasoning, action, and memory through end-to-end learning, open problems remain in data scalability, multi-turn reinforcement learning (RL), the limitations of GUI-only operation, and environment stability. In this technical report, we present UI-TARS-2, a native GUI-centered agent model that addresses these challenges through a systematic training methodology: a data flywheel for scalable data generation, a stabilized multi-turn RL framework, a hybrid GUI environment that integrates file systems and terminals, and a unified sandbox platform for large-scale rollouts. Empirical evaluation demonstrates that UI-TARS-2 achieves significant improvements over its predecessor UI-TARS-1.5. On GUI benchmarks, it reaches 88.2 on Online-Mind2Web, 47.5 on OSWorld, 50.6 on WindowsAgentArena, and 73.3 on AndroidWorld, outperforming strong baselines such as Claude and OpenAI agents. In game environments, it attains a mean normalized score of 59.8 across a 15-game suite-roughly 60% of human-level performance-and remains competitive with frontier proprietary models (e.g., OpenAI o3) on LMGame-Bench. Additionally, the model can generalize to long-horizon information-seeking tasks and software engineering benchmarks, highlighting its robustness across diverse agent tasks. Detailed analyses of training dynamics further provide insights into achieving stability and efficiency in large-scale agent RL. These results underscore UI-TARS-2’s potential to advance the state of GUI agents and exhibit strong generalization to real-world interactive scenarios.

[660] The Landscape of Agentic Reinforcement Learning for LLMs: A Survey

Guibin Zhang, Hejia Geng, Xiaohang Yu, Zhenfei Yin, Zaibin Zhang, Zelin Tan, Heng Zhou, Zhongzhi Li, Xiangyuan Xue, Yijiang Li, Yifan Zhou, Yang Chen, Chen Zhang, Yutao Fan, Zihu Wang, Songtao Huang, Yue Liao, Hongru Wang, Mengyue Yang, Heng Ji, Michael Littman, Jun Wang, Shuicheng Yan, Philip Torr, Lei Bai

Main category: cs.AI

TL;DR: Agentic RL transforms LLMs from passive generators into autonomous decision-making agents using temporally extended POMDPs, with RL as the key mechanism for adaptive behavior.

DetailsMotivation: To formalize the paradigm shift from conventional LLM RL to agentic RL and provide a comprehensive taxonomy and practical resources for this emerging field.

Method: Proposes a twofold taxonomy around core agentic capabilities (planning, tool use, memory, reasoning, self-improvement, perception) and their applications, synthesizing over 500 recent works.

Result: A consolidated landscape of open-source environments, benchmarks, and frameworks to support future research in agentic RL.

Conclusion: Agentic RL represents a significant evolution in AI agents, with reinforcement learning enabling the development of scalable, general-purpose autonomous agents, though challenges remain in this rapidly evolving field.

Abstract: The emergence of agentic reinforcement learning (Agentic RL) marks a paradigm shift from conventional reinforcement learning applied to large language models (LLM RL), reframing LLMs from passive sequence generators into autonomous, decision-making agents embedded in complex, dynamic worlds. This survey formalizes this conceptual shift by contrasting the degenerate single-step Markov Decision Processes (MDPs) of LLM-RL with the temporally extended, partially observable Markov decision processes (POMDPs) that define Agentic RL. Building on this foundation, we propose a comprehensive twofold taxonomy: one organized around core agentic capabilities, including planning, tool use, memory, reasoning, self-improvement, and perception, and the other around their applications across diverse task domains. Central to our thesis is that reinforcement learning serves as the critical mechanism for transforming these capabilities from static, heuristic modules into adaptive, robust agentic behavior. To support and accelerate future research, we consolidate the landscape of open-source environments, benchmarks, and frameworks into a practical compendium. By synthesizing over five hundred recent works, this survey charts the contours of this rapidly evolving field and highlights the opportunities and challenges that will shape the development of scalable, general-purpose AI agents.

[661] Probability Bracket Notation: Multivariable Systems and Static Bayesian Networks

Xing M. Wang

Main category: cs.AI

TL;DR: Extension of Probability Bracket Notation (PBN) to multivariable systems and Bayesian networks, providing a unifying symbolic framework for probabilistic modeling with applications in predictions, inferences, and expectations.

DetailsMotivation: To create a symbolic framework inspired by quantum mechanics notation that can concisely express dependencies among multiple random variables and enable algebraic manipulation of probabilistic models, addressing limitations of traditional probability notation.

Method: Expanded PBN to handle joint, marginal, and conditional probability distributions, as well as marginal and conditional expectations. Applied to static Bayesian networks including both discrete and continuous variables, using Student BN and a customized Healthcare BN as examples.

Result: Developed a unifying operator-like framework that simplifies analysis of probabilistic models. Demonstrated applications in predictions, inferences (bottom-up and top-down approaches), and expectations. Extended to continuous variables and hybrid discrete-continuous networks.

Conclusion: PBN shows potential as both an educational tool and practical framework for probabilistic modeling, with applications in causal reasoning, data analytics, machine learning, and artificial intelligence.

Abstract: We expand the Probability Bracket Notation (PBN), a symbolic framework inspired by the Dirac notation in quantum mechanics, to multivariable probability systems and static Bayesian networks (BNs). By introducing PBN for joint, marginal, and conditional probability distributions (PDs), as well as marginal and conditional expectations, we demonstrate how to express dependencies among multiple random variables concisely and manipulate them algebraically. Using the well-known Student BN as an example of probabilistic graphical models (PGMs), we show how to apply PBN to analyze predictions, inferences (using both bottom-up and top-down approaches), and expectations. We also extend PBN to BNs with continuous variables. After reviewing linear Gaussian networks, we introduce a customized Healthcare BN that includes both continuous and discrete random variables, utilizes user-specific data, and provides tailored predictions through discrete-display (DD) nodes as proxies for their continuous variable parents. Compared to traditional probability notation, PBN offers a unifying operator-like framework that simplifies the analysis of probabilistic models. This work highlights the potential of PBN as both an educational tool and a practical framework for probabilistic modeling, paving the way for applications in causal reasoning, inferences, expectations, data analytics, machine learning, and artificial intelligence.

[662] A Novel Kuhnian Ontology for Epistemic Classification of STM Scholarly Articles

Khalid M. Saqr

Main category: cs.AI

TL;DR: KGX3 is a deterministic epistemic classification system that maps Kuhnian stages from research papers to provide early signals of scientific paradigm shifts, outperforming traditional citation metrics.

DetailsMotivation: Current research evaluation relies on opaque, lagging proxies like citations. The authors aim to create transparent, reproducible epistemic classification for better funding and policy decisions.

Method: Formalized KGX3 as a scenario-based model for mapping Kuhnian stages from papers, proved determinism of classification pipeline, defined epistemic manifold for paradigm maps, and implemented governance preserving interpretability while protecting IP.

Result: Validated across recent corpora, demonstrated operational complexity at global scale, and showed the system delivers early actionable signals of drift, crisis, and shift unavailable to citation metrics or citation-anchored NLP.

Conclusion: KGX3 represents the latest iteration of a deterministic epistemic engine developed since 2019, providing superior paradigm shift detection compared to traditional metrics.

Abstract: Despite rapid gains in scale, research evaluation still relies on opaque, lagging proxies. To serve the scientific community, we pursue transparency: reproducible, auditable epistemic classification useful for funding and policy. Here we formalize KGX3 as a scenario-based model for mapping Kuhnian stages from research papers, prove determinism of the classification pipeline, and define the epistemic manifold that yields paradigm maps. We report validation across recent corpora, operational complexity at global scale, and governance that preserves interpretability while protecting core IP. The system delivers early, actionable signals of drift, crisis, and shift unavailable to citation metrics or citations-anchored NLP. KGX3 is the latest iteration of a deterministic epistemic engine developed since 2019, originating as Soph.io (2020), advanced as iKuhn (2024), and field-tested through Preprint Watch in 2025.

[663] Hypothesis Network Planned Exploration for Rapid Meta-Reinforcement Learning Adaptation

Maxwell Joseph Jacobson, Rohan Menon, John Zeng, Yexiang Xue

Main category: cs.AI

TL;DR: HyPE is an active exploration method for Meta-RL that plans action sequences to efficiently identify the most similar previously learned task, outperforming passive exploration baselines.

DetailsMotivation: Passive exploration strategies in Meta-RL limit adaptation speed when informative transitions are rare, requiring a more active approach to task identification.

Method: Hypothesis-Planned Exploration (HyPE) actively plans sequences of actions in a joint latent space where state-action transitions from different tasks form distinct paths.

Result: HyPE achieves exponentially lower failure probability than passive strategies, identifies the closest task in 65-75% of trials (vs 18-28% baseline), and yields 4x more successful adaptations under the same sample budget.

Conclusion: HyPE serves as a drop-in improvement for most model-based Meta-RL algorithms, enabling faster and more efficient task adaptation through active planned exploration.

Abstract: Meta-Reinforcement Learning (Meta-RL) learns optimal policies across a series of related tasks. A central challenge in Meta-RL is rapidly identifying which previously learned task is most similar to a new one, in order to adapt to it quickly. Prior approaches, despite significant success, typically rely on passive exploration strategies such as periods of random action to characterize the new task in relation to the learned ones. While sufficient when tasks are clearly distinguishable, passive exploration limits adaptation speed when informative transitions are rare or revealed only by specific behaviors. We introduce Hypothesis-Planned Exploration (HyPE), a method that actively plans sequences of actions during adaptation to efficiently identify the most similar previously learned task. HyPE operates within a joint latent space, where state-action transitions from different tasks form distinct paths. This latent-space planning approach enables HyPE to serve as a drop-in improvement for most model-based Meta-RL algorithms. By using planned exploration, HyPE achieves exponentially lower failure probability compared to passive strategies when informative transitions are sparse. On a natural language Alchemy game, HyPE identified the closest task in 65-75% of trials, far outperforming the 18-28% passive exploration baseline, and yielding up to 4x more successful adaptations under the same sample budget.

[664] TRACE-CS: A Hybrid Logic-LLM System for Explainable Course Scheduling

Stylianos Loukas Vasileiou, William Yeoh

Main category: cs.AI

TL;DR: TRACE-CS is a hybrid system combining symbolic reasoning with LLMs to handle contrastive queries in course scheduling, providing provably correct explanations with natural language accessibility.

DetailsMotivation: Address the challenge of creating explainable AI agents for scheduling systems that balance logical correctness with user-friendly natural language explanations.

Method: Combines logic-based techniques to encode scheduling constraints and generate provably correct explanations with LLMs to process natural language queries and refine explanations into user-friendly responses.

Result: Developed a system that successfully integrates symbolic knowledge representation methods with large language models for course scheduling problems.

Conclusion: The hybrid approach demonstrates how combining symbolic reasoning with LLMs creates effective explainable AI agents that address fundamental challenges in deployed scheduling systems.

Abstract: We present TRACE-CS, a novel hybrid system that combines symbolic reasoning with large language models (LLMs)to address contrastive queries in course scheduling problems. TRACE-CS leverages logic-based techniques to encode scheduling constraints and generate provably correct explanations, while utilizing an LLM to process natural language queries and refine logical explanations into user friendly responses. This system showcases how combining symbolic KR methods with LLMs creates explainable AI agents that balance logical correctness with natural language accessibility, addressing a fundamental challenge in deployed scheduling systems.

[665] Learning to Coordinate without Communication under Incomplete Information

Shenghui Chen, Shufang Zhu, Giuseppe De Giacomo, Ufuk Topcu

Main category: cs.AI

TL;DR: Agents achieve coordination without communication by interpreting action sequences as intent signals using finite-state transducers.

DetailsMotivation: Achieving seamless coordination in cooperative games under incomplete information when communication is not feasible.

Method: Develop strategy by interpreting partner’s action sequences as intent signals, constructing finite-state transducers from deterministic finite automata for each possible action.

Result: Strategies significantly outperform uncoordinated ones and closely match performance of direct communication coordination.

Conclusion: Effective coordination can be achieved without verbal communication through action sequence interpretation and finite-state transducer construction.

Abstract: Achieving seamless coordination in cooperative games is a crucial challenge in artificial intelligence, particularly when players operate under incomplete information. While communication helps, it is not always feasible. In this paper, we explore how effective coordination can be achieved without verbal communication, relying solely on observing each other’s actions. Our method enables an agent to develop a strategy by interpreting its partner’s action sequences as intent signals, constructing a finite-state transducer built from deterministic finite automata, one for each possible action the agent can take. Experiments show that these strategies significantly outperform uncoordinated ones and closely match the performance of coordinating via direct communication.

[666] HAICOSYSTEM: An Ecosystem for Sandboxing Safety Risks in Human-AI Interactions

Xuhui Zhou, Hyunwoo Kim, Faeze Brahman, Liwei Jiang, Hao Zhu, Ximing Lu, Frank Xu, Bill Yuchen Lin, Yejin Choi, Niloofar Mireshghallah, Ronan Le Bras, Maarten Sap

Main category: cs.AI

TL;DR: HAICOSYSTEM is a framework for evaluating AI agent safety in complex social interactions through modular sandbox simulations and multi-dimensional risk assessment.

DetailsMotivation: As AI agents become more autonomous in human interactions and tool usage, there is an increasing need to address interactional safety risks in diverse social contexts.

Method: Developed a modular sandbox environment simulating multi-turn interactions between humans and AI agents with various tools. Created comprehensive evaluation framework covering operational, content-related, societal, and legal risks. Conducted 1840 simulations across 92 scenarios in 7 domains.

Result: State-of-the-art LLMs (both proprietary and open-source) exhibited safety risks in over 50% of cases, with higher risks when interacting with malicious users. The framework successfully emulated realistic user-AI interactions and complex tool usage.

Conclusion: Building safe AI agents for complex interactions remains challenging, especially against malicious users. The released code platform enables practitioners to create custom scenarios and evaluate agent safety to foster the AI safety ecosystem.

Abstract: AI agents are increasingly autonomous in their interactions with human users and tools, leading to increased interactional safety risks. We present HAICOSYSTEM, a framework examining AI agent safety within diverse and complex social interactions. HAICOSYSTEM features a modular sandbox environment that simulates multi-turn interactions between human users and AI agents, where the AI agents are equipped with a variety of tools (e.g., patient management platforms) to navigate diverse scenarios (e.g., a user attempting to access other patients’ profiles). To examine the safety of AI agents in these interactions, we develop a comprehensive multi-dimensional evaluation framework that uses metrics covering operational, content-related, societal, and legal risks. Through running 1840 simulations based on 92 scenarios across seven domains (e.g., healthcare, finance, education), we demonstrate that HAICOSYSTEM can emulate realistic user-AI interactions and complex tool use by AI agents. Our experiments show that state-of-the-art LLMs, both proprietary and open-sourced, exhibit safety risks in over 50% cases, with models generally showing higher risks when interacting with simulated malicious users. Our findings highlight the ongoing challenge of building agents that can safely navigate complex interactions, particularly when faced with malicious users. To foster the AI agent safety ecosystem, we release a code platform that allows practitioners to create custom scenarios, simulate interactions, and evaluate the safety and performance of their agents.

[667] From Anchors to Answers: A Novel Node Tokenizer for Integrating Graph Structure into Large Language Models

Yanbiao Ji, Chang Liu, Xin Chen, Dan Luo, Mei Li, Yue Ding, Wenqing Lin, Hongtao Lu

Main category: cs.AI

TL;DR: NT-LLM is a novel framework that uses anchor-based positional encoding to help LLMs process graph data efficiently without heavy computational overhead, addressing the misalignment between discrete graph distances and continuous embedding spaces.

DetailsMotivation: Current methods for enabling LLMs to process graph data either consume excessive computational resources by converting graphs to text or require complex graph neural networks with significant training overhead, creating a need for more efficient solutions.

Method: The approach uses strategically selected reference nodes as anchors and encodes each node’s position relative to these anchors, implementing a rank-preserving objective for positional encoding pretraining to capture topological information efficiently.

Result: NT-LLM achieves superior performance across diverse graph tasks from basic structural analysis to complex reasoning scenarios, demonstrating effective enhancement of LLMs’ graph understanding and reasoning capabilities.

Conclusion: This lightweight yet powerful framework offers an efficient solution for graph-based applications of language models, effectively bridging the gap between discrete graph structures and continuous embedding spaces without computational burden.

Abstract: Enabling large language models (LLMs) to effectively process and reason with graph-structured data remains a significant challenge despite their remarkable success in natural language tasks. Current approaches either convert graph structures into verbose textual descriptions, consuming substantial computational resources, or employ complex graph neural networks as tokenizers, which introduce significant training overhead. To bridge this gap, we present NT-LLM, a novel framework with an anchor-based positional encoding scheme for graph representation. Our approach strategically selects reference nodes as anchors and encodes each node’s position relative to these anchors, capturing essential topological information without the computational burden of existing methods. Notably, we identify and address a fundamental issue: the inherent misalignment between discrete hop-based distances in graphs and continuous distances in embedding spaces. By implementing a rank-preserving objective for positional encoding pretraining, NT-LLM achieves superior performance across diverse graph tasks ranging from basic structural analysis to complex reasoning scenarios. Our comprehensive evaluation demonstrates that this lightweight yet powerful approach effectively enhances LLMs’ ability to understand and reason with graph-structured information, offering an efficient solution for graph-based applications of language models.

[668] Understanding Impact of Human Feedback via Influence Functions

Taywon Min, Haeone Lee, Yongchan Kwon, Kimin Lee

Main category: cs.AI

TL;DR: Using influence functions to analyze and improve human feedback quality in RLHF by detecting biases and guiding labelers for better reward model alignment.

DetailsMotivation: Human feedback in RLHF is often noisy, inconsistent, and biased, especially for complex responses, leading to misaligned reward signals and unintended side effects during the alignment process.

Method: Proposes a compute-efficient approximation method to apply influence functions to LLM-based reward models and large-scale preference datasets, enabling measurement of human feedback impact.

Result: Experiments demonstrate two key applications: detecting common labeler biases in human feedback datasets and guiding labelers to refine their strategies for better alignment with expert feedback.

Conclusion: Influence functions can enhance feedback interpretability and contribute to scalable oversight in RLHF, helping labelers provide more accurate and consistent feedback for better model alignment.

Abstract: In Reinforcement Learning from Human Feedback (RLHF), it is crucial to learn suitable reward models from human feedback to align large language models (LLMs) with human intentions. However, human feedback can often be noisy, inconsistent, or biased, especially when evaluating complex responses. Such feedback can lead to misaligned reward signals, potentially causing unintended side effects during the RLHF process. To address these challenges, we explore the use of influence functions to measure the impact of human feedback on the performance of reward models. We propose a compute-efficient approximation method that enables the application of influence functions to LLM-based reward models and large-scale preference datasets. Our experiments showcase two key applications of influence functions: (1) detecting common labeler biases in human feedback datasets and (2) guiding labelers in refining their strategies to better align with expert feedback. By quantifying the impact of human feedback, we believe that influence functions can enhance feedback interpretability and contribute to scalable oversight in RLHF, helping labelers provide more accurate and consistent feedback. Source code is available at https://github.com/mintaywon/IF_RLHF

Tobias Rohe, Michael Kölle, Jan Matheis, Rüdiger Höpfl, Leo Sünkel, Claudia Linnhoff-Popien

Main category: cs.AI

TL;DR: Comparison of PPO reinforcement learning vs simulated annealing for satellite transponder link configuration optimization, showing simulated annealing performs better for this static problem.

DetailsMotivation: Satellite communication requires efficient link configuration to optimize limited bandwidth and power resources, but no previous studies had applied reinforcement learning to this specific problem.

Method: Developed a transponder environment and compared PPO reinforcement learning algorithm with simulated annealing metaheuristic in two experiments.

Result: Simulated annealing delivered better results than PPO for this static optimization problem.

Conclusion: While simulated annealing performed better, the research demonstrates the potential of reinforcement learning for optimization problems in satellite communications.

Abstract: Satellite communication is a key technology in our modern connected world. With increasingly complex hardware, one challenge is to efficiently configure links (connections) on a satellite transponder. Planning an optimal link configuration is extremely complex and depends on many parameters and metrics. The optimal use of the limited resources, bandwidth and power of the transponder is crucial. Such an optimization problem can be approximated using metaheuristic methods such as simulated annealing, but recent research results also show that reinforcement learning can achieve comparable or even better performance in optimization methods. However, there have not yet been any studies on link configuration on satellite transponders. In order to close this research gap, a transponder environment was developed as part of this work. For this environment, the performance of the reinforcement learning algorithm PPO was compared with the metaheuristic simulated annealing in two experiments. The results show that Simulated Annealing delivers better results for this static problem than the PPO algorithm, however, the research in turn also underlines the potential of reinforcement learning for optimization problems.

[670] SETS: Leveraging Self-Verification and Self-Correction for Improved Test-Time Scaling

Jiefeng Chen, Jie Ren, Xinyun Chen, Chengrun Yang, Ruoxi Sun, Jinsung Yoon, Sercan Ö Arık

Main category: cs.AI

TL;DR: SETS is a novel test-time computation method that combines parallel and sequential techniques to enhance LLM performance on complex reasoning tasks without requiring model training, leveraging self-verification and self-correction capabilities.

DetailsMotivation: Existing test-time computation methods have limitations - parallel methods are inefficient and saturate quickly, while sequential methods struggle to improve after few rounds. Current hybrid approaches require fine-tuned models, creating implementation barriers.

Method: SETS strategically combines parallel sampling with sequential refinement, unifying sampling, verification, and correction within a single framework that fully leverages LLMs’ inherent self-improvement abilities without any additional training.

Result: Comprehensive experiments on challenging benchmarks (planning, reasoning, math, coding) show SETS achieves significant performance improvements and more advantageous test-time scaling behavior compared to alternative methods.

Conclusion: SETS provides an effective and scalable test-time computation approach that overcomes limitations of existing methods by harnessing LLMs’ self-verification and self-correction capabilities, enabling enhanced performance on complex tasks without model training requirements.

Abstract: Recent advancements in Large Language Models (LLMs) have created new opportunities to enhance performance on complex reasoning tasks by leveraging test-time computation. However, existing scaling methods have key limitations: parallel methods like repeated sampling are often inefficient and quickly saturate, while sequential methods like SELF-REFINE struggle to improve after a few rounds. Although combining these approaches shows promise, current methods require fine-tuned reward and revision models. This paper proposes Self-Enhanced Test-Time Scaling (SETS), a simple yet effective approach that overcomes these limitations by strategically combining parallel and sequential techniques and fully leveraging LLMs’ self-improvement abilities. SETS exploits the inherent self-verification and self-correction capabilities of LLMs, unifying sampling, verification, and correction within a single framework. This facilitates efficient and scalable test-time computation for enhanced performance on complex tasks without any model training. Our comprehensive experimental results on challenging benchmarks spanning planning, reasoning, math, and coding demonstrate that SETS achieves significant performance improvements and more advantageous test-time scaling behavior than the alternatives.

[671] Competing LLM Agents in a Non-Cooperative Game of Opinion Polarisation

Amin Qasmi, Usman Naseem, Mehwish Nasim

Main category: cs.AI

TL;DR: Game-theoretic analysis of opinion formation using LLM agents shows confirmation bias increases group alignment but worsens polarization, while resource-intensive debunking strategies risk long-term ineffectiveness.

DetailsMotivation: To understand how social psychology principles like confirmation bias and resource constraints affect opinion formation and resistance to misinformation in competitive influence scenarios.

Method: Developed a non-cooperative game framework with LLM agents competing to influence a population, incorporating confirmation bias, resource optimization, and penalties for misinformation propagation/countering.

Result: Higher confirmation bias strengthens within-group opinion alignment but increases overall polarization; lower bias leads to fragmented opinions with limited belief shifts; high-resource debunking strategies initially align population but risk rapid resource depletion.

Conclusion: Optimal influence strategies require balancing confirmation bias levels and resource allocation, as both extreme bias and aggressive debunking approaches have significant trade-offs in long-term effectiveness.

Abstract: We introduce a novel non-cooperative game to analyse opinion formation and resistance, incorporating principles from social psychology such as confirmation bias, resource constraints, and influence penalties. Our simulation features Large Language Model (LLM) agents competing to influence a population, with penalties imposed for generating messages that propagate or counter misinformation. This framework integrates resource optimisation into the agents’ decision-making process. Our findings demonstrate that while higher confirmation bias strengthens opinion alignment within groups, it also exacerbates overall polarisation. Conversely, lower confirmation bias leads to fragmented opinions and limited shifts in individual beliefs. Investing heavily in a high-resource debunking strategy can initially align the population with the debunking agent, but risks rapid resource depletion and diminished long-term influence

[672] LapSum - One Method to Differentiate Them All: Ranking, Sorting and Top-k Selection

Łukasz Struski, Michał B. Bednarczyk, Igor T. Podolak, Jacek Tabor

Main category: cs.AI

TL;DR: Novel technique for differentiable order-type operations using efficient closed-form formula for LapSum inverse, achieving O(n log n) complexity and outperforming SOTA methods.

DetailsMotivation: Need for efficient differentiable ranking and ordering operations that can handle high-dimensional vectors and large k values with low computational complexity.

Method: Leverages closed-form formula for inverse of LapSum function (sum of Laplace distributions) to construct differentiable soft ranking, top-k selection, and permutations.

Result: Outperforms state-of-the-art techniques, provides efficient CPU and CUDA implementations, and enables O(n log n) computation of losses and gradients.

Conclusion: Practical and scalable method for large-scale differentiable ranking and ordering problems with superior performance and efficiency.

Abstract: We present a novel technique for constructing differentiable order-type operations, including soft ranking, soft top-k selection, and soft permutations. Our approach leverages an efficient closed-form formula for the inverse of the function LapSum, defined as the sum of Laplace distributions. This formulation ensures low computational and memory complexity in selecting the highest activations, enabling losses and gradients to be computed in $O(n\log{}n)$ time. Through extensive experiments, we demonstrate that our method outperforms state-of-the-art techniques for high-dimensional vectors and large $k$ values. Furthermore, we provide efficient implementations for both CPU and CUDA environments, underscoring the practicality and scalability of our method for large-scale ranking and differentiable ordering problems.

[673] Perspective-Shifted Neuro-Symbolic World Models: A Framework for Socially-Aware Robot Navigation

Kevin Alcedo, Pedro U. Lima, Rachid Alami

Main category: cs.AI

TL;DR: A neuro-symbolic model-based RL architecture for social navigation that addresses belief tracking in POMDPs using perspective-shift operators and influence-based abstractions.

DetailsMotivation: Social navigation requires reasoning about hidden beliefs and intentions of others, which traditional MDPs cannot handle effectively due to partial observability of mental states.

Method: Proposes a neuro-symbolic model-based reinforcement learning architecture with perspective-shift operators for belief estimation, leveraging influence-based abstractions in structured multi-agent settings.

Result: The approach enables agents to navigate alongside humans by accounting for others’ beliefs and intentions under uncertainty.

Conclusion: The proposed architecture effectively addresses the challenge of belief tracking in partially observable social navigation scenarios through theory of mind and epistemic planning principles.

Abstract: Navigating in environments alongside humans requires agents to reason under uncertainty and account for the beliefs and intentions of those around them. Under a sequential decision-making framework, egocentric navigation can naturally be represented as a Markov Decision Process (MDP). However, social navigation additionally requires reasoning about the hidden beliefs of others, inherently leading to a Partially Observable Markov Decision Process (POMDP), where agents lack direct access to others’ mental states. Inspired by Theory of Mind and Epistemic Planning, we propose (1) a neuro-symbolic model-based reinforcement learning architecture for social navigation, addressing the challenge of belief tracking in partially observable environments; and (2) a perspective-shift operator for belief estimation, leveraging recent work on Influence-based Abstractions (IBA) in structured multi-agent settings.

[674] Identifying Macro Causal Effects in a C-DMG over ADMGs

Simon Ferreira, Charles K. Assaad

Main category: cs.AI

TL;DR: Causal effect identification in partially specified causal graphs using cluster-directed mixed graphs (C-DMGs), with focus on macro causal effects and proving do-calculus completeness.

DetailsMotivation: Existing causal effect identification methods assume fully specified causal graphs, but complex domains like medicine often only have partial causal knowledge. C-DMGs provide a practical higher-level representation by grouping variables into clusters.

Method: The paper focuses on cluster-directed mixed graphs (C-DMGs) which can represent many ADMGs. It establishes that do-calculus is sound and complete for identifying macro causal effects in C-DMGs when cluster sizes are unknown or greater than one.

Result: The research proves that do-calculus is both sound and complete for macro causal effect identification in C-DMGs over ADMGs. It also provides graphical characterization of non-identifiability for macro causal effects.

Conclusion: This work advances causal effect identification in partially specified systems by providing formal guarantees for do-calculus in cluster-based representations, enabling practical causal inference in complex domains with incomplete knowledge.

Abstract: Causal effect identification using causal graphs is a fundamental challenge in causal inference. While extensive research has been conducted in this area, most existing methods assume the availability of fully specified directed acyclic graphs or acyclic directed mixed graphs. However, in complex domains such as medicine and epidemiology, complete causal knowledge is often unavailable, and only partial information about the system is accessible. This paper focuses on causal effect identification within partially specified causal graphs, with particular emphasis on cluster-directed mixed graphs (C-DMGs) which can represent many different acyclic directed mixed graphs (ADMGs). These graphs provide a higher-level representation of causal relationships by grouping variables into clusters, offering a more practical approach for handling complex systems. Unlike fully specified ADMGs, C-DMGs can contain cycles, which complicate their analysis and interpretation. Furthermore, their cluster-based nature introduces new challenges, as it gives rise to two distinct types of causal effects: macro causal effects and micro causal effects, each with different properties. In this work, we focus on macro causal effects, which describe the effects of entire clusters on other clusters. We establish that the do-calculus is both sound and complete for identifying these effects in C-DMGs over ADMGs when the cluster sizes are either unknown or of size greater than one. Additionally, we provide a graphical characterization of non-identifiability for macro causal effects in these graphs.

[675] Lemmanaid: Neuro-Symbolic Lemma Conjecturing

Yousef Alhessi, Sólrún Halla Einarsdóttir, George Granberry, Emily First, Moa Johansson, Sorin Lerner, Nicholas Smallbone

Main category: cs.AI

TL;DR: Lemmanaid combines LLMs and symbolic methods to automatically generate useful mathematical lemmas, outperforming both pure neural and symbolic approaches by discovering 29-39.5% of human-written lemmas.

DetailsMotivation: Automated lemma generation would significantly improve automated reasoning tools and lower barriers for formalizing mathematics in proof assistants, but current neural and symbolic approaches face challenges.

Method: Train LLM to generate lemma templates describing lemma shape, then use symbolic methods to fill in details. Combines neural template generation with symbolic completion.

Result: Outperforms both neural-only and symbolic methods, discovering 29-39.5% of human-written lemmas (8-15% improvement over neural-only approach). Tested on Isabelle’s HOL library and Archive of Formal Proofs.

Conclusion: Neuro-symbolic approach leveraging both LLMs and symbolic methods can generate useful lemmas across diverse domains, facilitating computer-assisted theory development and formalization.

Abstract: Automatically conjecturing useful, interesting and novel lemmas would greatly improve automated reasoning tools and lower the bar for formalizing mathematics in proof assistants. It is however a very challenging task for both neural and symbolic approaches. We present the first steps towards a practical neuro-symbolic lemma conjecturing tool, Lemmanaid, that combines Large Language Models (LLMs) and symbolic methods, and evaluate it on proof libraries for the Isabelle proof assistant. We train an LLM to generate lemma templates that describe the shape of a lemma, and use symbolic methods to fill in the details. We compare Lemmanaid against an LLM trained to generate complete lemma statements as well as previous fully symbolic conjecturing methods. Lemmanaid outperforms both neural and symbolic methods on test sets from Isabelle’s HOL library and from its Archive of Formal Proofs, discovering between 29-39.5% of the gold standard human written lemmas. This is 8-15% more lemmas than the neural-only method. By leveraging the best of both symbolic and neural methods we can generate useful lemmas for a wide range of input domains, facilitating computer-assisted theory development and formalization.

[676] Assessing AI-Generated Questions’ Alignment with Cognitive Frameworks in Educational Assessment

Antoun Yaacoub, Jérôme Da-Rugna, Zainab Assaghir

Main category: cs.AI

TL;DR: Study integrates Bloom’s Taxonomy into AI-driven MCQ generation tool (OneClickQuiz) to improve cognitive level alignment, showing DistilBERT achieves 91% accuracy in classifying questions across Bloom’s levels.

DetailsMotivation: To investigate whether incorporating Bloom's Taxonomy framework can improve the alignment of AI-generated multiple-choice questions with specific cognitive objectives in educational assessment tools.

Method: Developed dataset of 3691 questions categorized by Bloom’s levels; evaluated classification models including Multinomial Logistic Regression, Naive Bayes, Linear SVC, and Transformer-based DistilBERT model for question categorization effectiveness.

Result: Higher Bloom’s levels correlate with increased question length, complexity metrics (FKGL, LD); Multinomial Logistic Regression varied by level; merging higher categories improved accuracy; DistilBERT achieved highest performance with 91% overall validation accuracy.

Conclusion: Integration of Bloom’s Taxonomy into AI-driven assessment tools shows promise, with advanced models like DistilBERT significantly enhancing educational content generation and cognitive level classification.

Abstract: This study evaluates the integration of Bloom’s Taxonomy into OneClickQuiz, an Artificial Intelligence (AI) driven plugin for automating Multiple-Choice Question (MCQ) generation in Moodle. Bloom’s Taxonomy provides a structured framework for categorizing educational objectives into hierarchical cognitive levels. Our research investigates whether incorporating this taxonomy can improve the alignment of AI-generated questions with specific cognitive objectives. We developed a dataset of 3691 questions categorized according to Bloom’s levels and employed various classification models-Multinomial Logistic Regression, Naive Bayes, Linear Support Vector Classification (SVC), and a Transformer-based model (DistilBERT)-to evaluate their effectiveness in categorizing questions. Our results indicate that higher Bloom’s levels generally correlate with increased question length, Flesch-Kincaid Grade Level (FKGL), and Lexical Density (LD), reflecting the increased complexity of higher cognitive demands. Multinomial Logistic Regression showed varying accuracy across Bloom’s levels, performing best for “Knowledge” and less accurately for higher-order levels. Merging higher-level categories improved accuracy for complex cognitive tasks. Naive Bayes and Linear SVC also demonstrated effective classification for lower levels but struggled with higher-order tasks. DistilBERT achieved the highest performance, significantly improving classification of both lower and higher-order cognitive levels, achieving an overall validation accuracy of 91%. This study highlights the potential of integrating Bloom’s Taxonomy into AI-driven assessment tools and underscores the advantages of advanced models like DistilBERT for enhancing educational content generation.

[677] RECAST: Strengthening LLMs’ Complex Instruction Following with Constraint-Verifiable Data

Zhengkang Guo, Wenhao Liu, Mingchen Xie, Jingwen Xu, Zisu Huang, Muzhao Tian, Jianhan Xu, Muling Wu, Xiaohua Wang, Changze Lv, He-Da Wang, Hu Yao, Xiaoqing Zheng, Xuanjing Huang

Main category: cs.AI

TL;DR: RECAST framework creates large-scale datasets with complex constraints (30K+ instances) to improve LLMs’ ability to follow multi-constraint instructions, enabling both fine-tuning and reinforcement learning approaches.

DetailsMotivation: LLMs struggle with complex instructions containing many constraints (especially >10), limiting their practical application despite growing user sophistication in prompt crafting.

Method: Propose RECAST framework that synthesizes datasets with extracted real-world constraints, using rule-based validators for quantitative constraints and LLM-based validators for qualitative ones. Construct RECAST-30K dataset with 30k instances across 15 constraint types.

Result: Models fine-tuned on RECAST-30K show substantial improvements in following complex instructions. The verifiability enables reward functions for reinforcement learning, further boosting performance on complex tasks.

Conclusion: RECAST provides an effective framework for improving LLM performance on complex multi-constraint instructions through dataset synthesis and enables reinforcement learning approaches via constraint verifiability.

Abstract: Large language models (LLMs) are increasingly expected to tackle complex tasks, driven by their expanding applications and users’ growing proficiency in crafting sophisticated prompts. However, as the number of explicitly stated requirements increases (particularly more than 10 constraints), LLMs often struggle to accurately follow such complex instructions. To address this challenge, we propose RECAST, a novel framework for synthesizing datasets where each example incorporates far more constraints than those in existing benchmarks. These constraints are extracted from real-world prompt-response pairs to ensure practical relevance. RECAST enables automatic verification of constraint satisfaction via rule-based validators for quantitative constraints and LLM-based validators for qualitative ones. Using this framework, we construct RECAST-30K, a large-scale, high-quality dataset comprising 30k instances spanning 15 constraint types. Experimental results demonstrate that models fine-tuned on RECAST-30K show substantial improvements in following complex instructions. Moreover, the verifiability provided by RECAST enables the design of reward functions for reinforcement learning, which further boosts model performance on complex and challenging tasks.

[678] General agents contain world models

Jonathan Richens, David Abel, Alexis Bellot, Tom Everitt

Main category: cs.AI

TL;DR: The paper formally proves that world models are necessary for agents to achieve flexible, goal-directed behavior and generalize to multi-step tasks, showing that model-free learning alone is insufficient.

DetailsMotivation: To answer the fundamental question of whether world models are essential for flexible, goal-directed behavior or if model-free learning approaches are sufficient.

Method: The authors provide a formal mathematical proof showing that any agent capable of generalizing to multi-step goal-directed tasks must have learned a predictive model of its environment, and demonstrate how this model can be extracted from the agent’s policy.

Result: The research establishes that increasing agent performance or handling more complex goals requires learning increasingly accurate world models, and that world models can be extracted from successful agents’ policies.

Conclusion: This finding has significant implications for developing safe and general agents, bounding agent capabilities in complex environments, and provides new algorithmic approaches for eliciting world models from trained agents.

Abstract: Are world models a necessary ingredient for flexible, goal-directed behaviour, or is model-free learning sufficient? We provide a formal answer to this question, showing that any agent capable of generalizing to multi-step goal-directed tasks must have learned a predictive model of its environment. We show that this model can be extracted from the agent’s policy, and that increasing the agents performance or the complexity of the goals it can achieve requires learning increasingly accurate world models. This has a number of consequences: from developing safe and general agents, to bounding agent capabilities in complex environments, and providing new algorithms for eliciting world models from agents.

[679] OThink-R1: Intrinsic Fast/Slow Thinking Mode Switching for Over-Reasoning Mitigation

Shengjia Zhang, Junjie Wu, Jiawei Chen, Changwang Zhang, Xingyu Lou, Wangchunshu Zhou, Sheng Zhou, Can Wang, Jun Wang

Main category: cs.AI

TL;DR: OThink-R1 identifies and prunes redundant reasoning steps in large reasoning models, reducing token usage by 23% while maintaining accuracy by dynamically switching between fast and slow thinking modes.

DetailsMotivation: Large reasoning models often use excessive chain-of-thought reasoning for simple tasks that could be solved with fewer tokens, indicating unnecessary computational overhead.

Method: Systematic analysis of reasoning trajectories using identified paradigms and LLM-Judge classification, followed by pruning redundant steps while preserving logical validity through dynamic fast/slow thinking modes.

Result: 23% average reduction in reasoning redundancy across mathematical and question-answering tasks without compromising accuracy.

Conclusion: OThink-R1 provides an efficient approach to optimize reasoning models by eliminating unnecessary reasoning steps, offering practical guidelines for more efficient AI reasoning systems.

Abstract: Recent advanced large reasoning models (LRMs) leverage extended chain-of-thought (CoT) reasoning to solve complex tasks, achieving state-of-the-art performance. Despite their success, we identify a critical issue: a substantial portion of simple tasks solved by LRMs can also be addressed by non-reasoning LLMs using significantly fewer tokens, indicating the complex reasoning may not always be necessary. To address this, we systematically analyze the reasoning trajectories of LRMs and present a method utilizing identified paradigms and LLM-Judge to classify these trajectories as either Redundant Reasoning or Essential Reasoning. And we introduce OThink-R1, a method that prunes redundant reasoning steps while preserving logical validity. OThink-R1 dynamically employs the non-thinking mode (fast-thinking) for straightforward problems while engaging in deliberate thinking (slow-thinking) for complex problems. Experiments across mathematical and question-answering tasks demonstrate that OThink-R1 reduces reasoning redundancy by almost 23% on average without compromising accuracy, offering practical guidelines for efficient reasoning models. The code is available at https://github.com/AgenticIR-Lab/OThink-R1.

[680] Intelligent Assistants for the Semiconductor Failure Analysis with LLM-Based Planning Agents

Aline Dobrovsky, Konstantin Schekotihin, Christian Burmer

Main category: cs.AI

TL;DR: This paper presents an LLM-based Planning Agent system for automating semiconductor failure analysis workflows by orchestrating multiple AI components into cohesive processes.

DetailsMotivation: Semiconductor failure analysis is complex and knowledge-intensive, requiring integration of multiple AI models. As more AI components are deployed, there's a need to orchestrate them into efficient workflows that seamlessly integrate with the FA process.

Method: The authors design and implement an agentic AI system using a Large Language Model-based Planning Agent (LPA) that integrates LLMs with advanced planning capabilities and external tool utilization. The LPA autonomously processes complex queries, retrieves relevant data from external systems, and generates human-readable responses.

Result: Evaluation results demonstrate the agent’s operational effectiveness and reliability in supporting various FA tasks, including non-conformity detection, case retrieval from diverse data sources, and report generation from annotated images.

Conclusion: The LLM-based Planning Agent approach successfully addresses the challenge of orchestrating multiple AI components in semiconductor failure analysis, providing an effective solution for automating complex FA workflows through autonomous processing and intelligent integration of external tools and data sources.

Abstract: Failure Analysis (FA) is a highly intricate and knowledge-intensive process. The integration of AI components within the computational infrastructure of FA labs has the potential to automate a variety of tasks, including the detection of non-conformities in images, the retrieval of analogous cases from diverse data sources, and the generation of reports from annotated images. However, as the number of deployed AI models increases, the challenge lies in orchestrating these components into cohesive and efficient workflows that seamlessly integrate with the FA process. This paper investigates the design and implementation of an agentic AI system for semiconductor FA using a Large Language Model (LLM)-based Planning Agent (LPA). The LPA integrates LLMs with advanced planning capabilities and external tool utilization, allowing autonomous processing of complex queries, retrieval of relevant data from external systems, and generation of human-readable responses. The evaluation results demonstrate the agent’s operational effectiveness and reliability in supporting FA tasks.

Mingda Zhang, Na Zhao, Jianglong Qing, Qing xu, Kaiwen Pan, Ting luo

Main category: cs.AI

TL;DR: A framework combining prompt engineering with multidimensional knowledge graphs to significantly improve LLM performance in legal dispute analysis, achieving major gains in sensitivity, specificity, and citation accuracy.

DetailsMotivation: Current LLMs struggle with complex legal concepts, reasoning consistency, and accurate legal source citation in legal dispute analysis, limiting their effectiveness in intelligent legal assistance systems.

Method: Three-stage hierarchical prompt structure (task definition, knowledge background, reasoning guidance) combined with three-layer knowledge graph (legal ontology, representation, instance layers), plus four supporting methods for legal concept retrieval: direct code matching, semantic vector similarity, ontology path reasoning, and lexical segmentation.

Result: Significant performance improvements: sensitivity increased by 11.1%-11.3%, specificity by 5.4%-6.0%, and citation accuracy by 29.5%-39.7% through extensive testing.

Conclusion: The framework provides better legal analysis and judicial logic understanding, offering a new technical approach for intelligent legal assistance systems by effectively addressing LLM limitations in legal contexts.

Abstract: Legal dispute analysis is crucial for intelligent legal assistance systems. However, current LLMs face significant challenges in understanding complex legal concepts, maintaining reasoning consistency, and accurately citing legal sources. This research presents a framework combining prompt engineering with multidimensional knowledge graphs to improve LLMs’ legal dispute analysis. Specifically, the framework includes a three-stage hierarchical prompt structure (task definition, knowledge background, reasoning guidance) along with a three-layer knowledge graph (legal ontology, representation, instance layers). Additionally, four supporting methods enable precise legal concept retrieval: direct code matching, semantic vector similarity, ontology path reasoning, and lexical segmentation. Through extensive testing, results show major improvements: sensitivity increased by 11.1%-11.3%, specificity by 5.4%-6.0%, and citation accuracy by 29.5%-39.7%. As a result, the framework provides better legal analysis and understanding of judicial logic, thus offering a new technical method for intelligent legal assistance systems.

[682] A Multi-granularity Concept Sparse Activation and Hierarchical Knowledge Graph Fusion Framework for Rare Disease Diagnosis

Mingda Zhang, Na Zhao, Jianglong Qin, Guoyu Ye, Ruixiang Tang

Main category: cs.AI

TL;DR: A framework combining multi-granularity sparse activation of medical concepts with hierarchical knowledge graphs improves rare-disease diagnosis, achieving near-clinical accuracy thresholds and shortening diagnostic odysseys.

DetailsMotivation: Rare-disease diagnosis remains challenging due to insufficient knowledge representation depth, limited concept understanding, and constrained clinical reasoning in current medical LLMs.

Method: Proposes a framework with multi-granularity sparse activation of medical concepts coupled with hierarchical knowledge graph. Uses four complementary matching algorithms, diversity control, and five-level fallback strategy for precise concept activation, supported by a three-layer knowledge graph (taxonomy, clinical features, instances).

Result: Experiments on BioASQ rare-disease QA set show BLEU gains of 0.09, ROUGE gains of 0.05, accuracy gains of 0.12, with peak accuracy of 0.89 approaching the 0.90 clinical threshold. Expert evaluation confirms improvements in information quality, reasoning, and professional expression.

Conclusion: The approach effectively shortens the diagnostic odyssey for rare-disease patients by enhancing concept understanding and clinical reasoning through structured knowledge representation and precise activation mechanisms.

Abstract: Despite advances from medical large language models in healthcare, rare-disease diagnosis remains hampered by insufficient knowledge-representation depth, limited concept understanding, and constrained clinical reasoning. We propose a framework that couples multi-granularity sparse activation of medical concepts with a hierarchical knowledge graph. Four complementary matching algorithms, diversity control, and a five-level fallback strategy enable precise concept activation, while a three-layer knowledge graph (taxonomy, clinical features, instances) provides structured, up-to-date context. Experiments on the BioASQ rare-disease QA set show BLEU gains of 0.09, ROUGE gains of 0.05, and accuracy gains of 0.12, with peak accuracy of 0.89 approaching the 0.90 clinical threshold. Expert evaluation confirms improvements in information quality, reasoning, and professional expression, suggesting our approach shortens the “diagnostic odyssey” for rare-disease patients.

[683] Pareto-NRPA: A Novel Monte-Carlo Search Algorithm for Multi-Objective Optimization

Noé Lallouet, Tristan Cazenave, Cyrille Enderli

Main category: cs.AI

TL;DR: Pareto-NRPA extends single-objective NRPA to multi-objective optimization, using multiple policies to explore solution spaces and maintain non-dominated fronts, showing strong performance on constrained problems.

DetailsMotivation: To adapt the successful Nested Rollout Policy Adaptation (NRPA) algorithm from single-objective to multi-objective optimization problems, as no previous multi-objective adaptation of NRPA existed.

Method: Extends NRPA with multiple policies that concurrently explore different solution space regions, maintains non-dominated fronts at each search level, and adapts policies based on Pareto front diversity and sequence isolation.

Result: Achieves competitive performance against state-of-the-art multi-objective algorithms in convergence and diversity, particularly outperforming evolutionary algorithms on constrained search spaces in both MO-TSPTW and neural architecture search tasks.

Conclusion: Pareto-NRPA successfully generalizes NRPA to multi-objective optimization, demonstrating strong performance and constituting the first adaptation of NRPA to multi-objective settings.

Abstract: We introduce Pareto-NRPA, a new Monte-Carlo algorithm designed for multi-objective optimization problems over discrete search spaces. Extending the Nested Rollout Policy Adaptation (NRPA) algorithm originally formulated for single-objective problems, Pareto-NRPA generalizes the nested search and policy update mechanism to multi-objective optimization. The algorithm uses a set of policies to concurrently explore different regions of the solution space and maintains non-dominated fronts at each level of search. Policy adaptation is performed with respect to the diversity and isolation of sequences within the Pareto front. We benchmark Pareto-NRPA on two classes of problems: a novel bi-objective variant of the Traveling Salesman Problem with Time Windows problem (MO-TSPTW), and a neural architecture search task on well-known benchmarks. Results demonstrate that Pareto-NRPA achieves competitive performance against state-of-the-art multi-objective algorithms, both in terms of convergence and diversity of solutions. Particularly, Pareto-NRPA strongly outperforms state-of-the-art evolutionary multi-objective algorithms on constrained search spaces. To our knowledge, this work constitutes the first adaptation of NRPA to the multi-objective setting.

[684] Graph-Augmented Large Language Model Agents: Current Progress and Future Prospects

Yixin Liu, Guibin Zhang, Kun Wang, Shiyuan Li, Shirui Pan

Main category: cs.AI

TL;DR: Graph-augmented LLM Agents (GLA) use graphs to enhance LLM agent capabilities in planning, memory, tool usage, and multi-agent coordination, addressing limitations of standalone LLMs.

DetailsMotivation: LLM agents have impressive capabilities but are limited in key agentic procedures like reliable planning, long-term memory, tool management, and multi-agent coordination. Graphs can serve as powerful auxiliary structures to enhance these capabilities.

Method: Categorizing existing GLA methods by their primary functions in LLM agent systems (planning, memory, tool usage) and analyzing how graphs and graph learning algorithms contribute to each. Also discussing GLA solutions for multi-agent system orchestration, efficiency optimization, and trustworthiness.

Result: The paper provides a comprehensive overview of recent advances in Graph-augmented LLM Agents, categorizing methods and analyzing graph contributions across different agent functions.

Conclusion: Graphs play a crucial role in enhancing LLM agent systems. Future directions include improving structural adaptability and enabling unified, scalable, and multimodal GLA systems. This serves as a roadmap for future GLA research.

Abstract: Autonomous agents based on large language models (LLMs) have demonstrated impressive capabilities in a wide range of applications, including web navigation, software development, and embodied control. While most LLMs are limited in several key agentic procedures, such as reliable planning, long-term memory, tool management, and multi-agent coordination, graphs can serve as a powerful auxiliary structure to enhance structure, continuity, and coordination in complex agent workflows. Given the rapid growth and fragmentation of research on Graph-augmented LLM Agents (GLA), this paper offers a timely and comprehensive overview of recent advances and also highlights key directions for future work. Specifically, we categorize existing GLA methods by their primary functions in LLM agent systems, including planning, memory, and tool usage, and then analyze how graphs and graph learning algorithms contribute to each. For multi-agent systems, we further discuss how GLA solutions facilitate the orchestration, efficiency optimization, and trustworthiness of MAS. Finally, we highlight key future directions to advance this field, from improving structural adaptability to enabling unified, scalable, and multimodal GLA systems. We hope this paper can serve as a roadmap for future research on GLA and foster a deeper understanding of the role of graphs in LLM agent systems.

[685] LUMIR: an LLM-Driven Unified Agent Framework for Multi-task Infrared Spectroscopy Reasoning

Zujie Xie, Zixuan Chen, Jiheng Liang, Xiangyang Yu, Ziru Yu

Main category: cs.AI

TL;DR: LUMIR is an LLM-driven agent framework that achieves accurate infrared spectral analysis under low data conditions by integrating literature knowledge, automated preprocessing, and few-shot learning, outperforming traditional ML/DL models.

DetailsMotivation: Infrared spectroscopy faces challenges with high-dimensional signals and overlapping bands, while LLMs' generalization capabilities remain untapped for spectral interpretation. The study aims to leverage LLMs for automated spectral analysis with minimal labeled data.

Method: LUMIR framework integrates structured literature knowledge base, automated preprocessing, feature extraction, and predictive modeling. It mines peer-reviewed studies for validated strategies, transforms spectra into low-dimensional representations, and uses few-shot prompts for classification, regression, and anomaly detection.

Result: LUMIR achieved performance comparable to or surpassing established ML/DL models across diverse datasets (Milk NIR, Chinese herbs, CRP, industrial wastewater, Tecator, Corn), particularly in resource-limited settings.

Conclusion: Combining structured literature guidance with few-shot learning enables robust and scalable spectral interpretation. LUMIR establishes a new paradigm for applying LLMs to infrared spectroscopy with high accuracy using minimal labeled data across scientific and industrial domains.

Abstract: Infrared spectroscopy enables rapid, non destructive analysis of chemical and material properties, yet high dimensional signals and overlapping bands hinder conventional chemometric methods. Large language models (LLMs), with strong generalization and reasoning capabilities, offer new opportunities for automated spectral interpretation, but their potential in this domain remains largely untapped. This study introduces LUMIR (LLM-driven Unified agent framework for Multi-task Infrared spectroscopy Reasoning), an agent based framework designed to achieve accurate infrared spectral analysis under low data conditions. LUMIR integrates a structured literature knowledge base, automated preprocessing, feature extraction, and predictive modeling into a unified pipeline. By mining peer reviewed spectroscopy studies, it identifies validated preprocessing and feature derivation strategies, transforms spectra into low dimensional representations, and applies few-shot prompts for classification, regression, and anomaly detection. The framework was validated on diverse datasets, including the publicly available Milk near-infrared dataset, Chinese medicinal herbs, Citri Reticulatae Pericarpium(CRP) with different storage durations, an industrial wastewater COD dataset, and two additional public benchmarks, Tecator and Corn. Across these tasks, LUMIR achieved performance comparable to or surpassing established machine learning and deep learning models, particularly in resource limited settings. This work demonstrates that combining structured literature guidance with few-shot learning enables robust, scalable, and automated spectral interpretation. LUMIR establishes a new paradigm for applying LLMs to infrared spectroscopy, offering high accuracy with minimal labeled data and broad applicability across scientific and industrial domains.

[686] MetaAgent: Toward Self-Evolving Agent via Tool Meta-Learning

Hongjin Qian, Zheng Liu

Main category: cs.AI

TL;DR: MetaAgent is a self-evolving AI agent that learns through hands-on practice, dynamically improving its tool-use and reasoning capabilities without model retraining by generating help requests, self-reflection, and building persistent knowledge from experience.

DetailsMotivation: To create an agentic system that develops expertise through practical experience rather than static training, enabling continual self-improvement and robust knowledge discovery without requiring parameter updates.

Method: Starts with minimal workflow, generates natural language help requests routed to external tools, conducts self-reflection and answer verification, distills experience into contextual knowledge, and autonomously builds in-house tools and persistent knowledge base from tool-use history.

Result: Outperforms workflow-based baselines and matches/exceeds end-to-end trained agents on challenging benchmarks (GAIA, WebWalkerQA, BrowseCamp), demonstrating robust knowledge discovery capabilities.

Conclusion: MetaAgent shows promise for self-evolving agentic systems that can continually refine reasoning and tool-use strategies through meta tool learning, enabling general-purpose knowledge discovery without model retraining.

Abstract: In this work, we propose MetaAgent, an agentic paradigm inspired by the principle of learning-by-doing, where expertise is developed through hands-on practice and continual self-improvement. MetaAgent starts with a minimal workflow, equipped only with basic reasoning and adaptive help-seeking abilities. When a knowledge gap is encountered, MetaAgent generates natural language help requests, which are routed to the most suitable external tool by a dedicated tool router. As MetaAgent solves tasks, it continually conducts self-reflection and answer verification, distilling actionable experience into concise texts that are dynamically incorporated into future task contexts. Besides, MetaAgent autonomously builds in-house tools and a persistent knowledge base by organizing its tool-use history, further enhancing its ability to retrieve and integrate relevant information We term this continual, data-driven process as \textit{meta tool learning}, through which MetaAgent incrementally refines its reasoning and tool-use strategies, without changing model parameters or requiring further post-training. Evaluated on challenging knowledge discovery benchmarks, including GAIA, WebWalkerQA, and BrowseCamp, MetaAgent consistently outperforms workflow-based baselines and matches or exceeds end-to-end trained agents, demonstrating the promise of self-evolving agentic systems for robust, general-purpose knowledge discovery. We provide our source codes in https://github.com/qhjqhj00/MetaAgent.

[687] Profile-Aware Maneuvering: A Dynamic Multi-Agent System for Robust GAIA Problem Solving by AWorld

Zhitian Xie, Qintong Wu, Chengyue Yu, Chenyi Zhuang, Jinjie Gu

Main category: cs.AI

TL;DR: AWorld framework uses a multi-agent system with profile-aware supervision to improve LLM reliability when using external tools, achieving state-of-the-art performance on GAIA benchmark.

DetailsMotivation: Large language models relying on external tools face reliability challenges from extended contexts and noisy outputs, requiring robust supervision systems.

Method: Dynamic multi-agent system with Execution Agent supervised by Guard Agent using System Identification methodology to create performance fingerprints for targeted interventions.

Result: Significantly improved effectiveness and stability, outperforming single-agent systems and naive multi-agent counterparts, achieving first place on GAIA leaderboard among open-source projects.

Conclusion: Building trustworthy intelligent systems requires deep empirical understanding of each agent’s unique capabilities and limitations through profile-aware supervision.

Abstract: The rapid advancement of large language models (LLMs) has empowered intelligent agents to leverage diverse external tools for solving complex real-world problems. However, this reliance introduces new challenges, as extended contexts and noisy tool outputs can undermine system reliability. To address this, we propose a dynamic Multi-Agent System (MAS) in our AWorld framework, where an Execution Agent is supervised by a Guard Agent that provides on-demand dynamic maneuvering, verifying and correcting the reasoning process to improve robustness over single-agent systems. To move beyond this generic supervision, we enhance the architecture with a methodology inspired by System Identification from control theory. This method first profiles the Execution Agent offline on a benchmark dataset to create a “performance fingerprint” of its unique weaknesses. The Guard Agent then leverages this fingerprint online to deliver profile-aware supervision, making targeted interventions based on known failure patterns rather than merely reacting to immediate logical flaws. Extensive experiments on the GAIA dataset demonstrate that this profile-aware MAS significantly improves both effectiveness and stability, outperforming not only single-agent systems but also its naive counterpart. This superior performance led our system to achieve first place among open-source projects on the prestigious GAIA leaderboard. These findings highlight that building truly trustworthy intelligent systems requires not just collaboration, but a deep, empirically-grounded understanding of each agent’s unique capabilities and limitations.

[688] Inclusion Arena: An Open Platform for Evaluating Large Foundation Models with Real-World Apps

Kangyu Wang, Hongliang He, Lin Liu, Ruiqi Liang, Zhenzhong Lan, Jianguo Li

Main category: cs.AI

TL;DR: Inclusion Arena is a live leaderboard that ranks LLMs/MLLMs using real-world human feedback from AI applications, employing innovative Bradley-Terry model enhancements for reliable rankings.

DetailsMotivation: Existing benchmarks use static datasets or general crowdsourced prompts, failing to reflect real-world application performance. There's a need for evaluation that mirrors practical usage scenarios.

Method: Platform integrates pairwise model comparisons into natural user interactions. Uses Bradley-Terry model with two innovations: Placement Matches for cold-start rating estimation, and Proximity Sampling to prioritize battles between similarly capable models.

Result: Yields reliable and stable rankings, shows higher data transitivity than general crowdsourced datasets, and significantly mitigates malicious manipulation risk.

Conclusion: Inclusion Arena bridges the gap between AI model development and real-world applications, accelerating development of LLMs/MLLMs optimized for practical, user-centric deployments.

Abstract: Large Language Models (LLMs) and Multimodal Large Language Models (MLLMs) have ushered in a new era of AI capabilities, demonstrating near-human-level performance across diverse scenarios. While numerous benchmarks (e.g., MMLU) and leaderboards (e.g., Chatbot Arena) have been proposed to help evolve the development of LLMs and MLLMs, most rely on static datasets or crowdsourced general-domain prompts, often falling short of reflecting performance in real-world applications. To bridge this critical gap, we present Inclusion Arena, a live leaderboard that ranks models based on human feedback collected directly from AI-powered applications. Our platform integrates pairwise model comparisons into natural user interactions, ensuring evaluations reflect practical usage scenarios. For robust model ranking, we employ the Bradley-Terry model augmented with two key innovations: (1) Placement Matches, a cold-start mechanism to quickly estimate initial ratings for newly integrated models, and (2) Proximity Sampling, an intelligent comparison strategy that prioritizes battles between models of similar capabilities to maximize information gain and enhance rating stability. Extensive empirical analyses and simulations demonstrate that Inclusion Arena yields reliable and stable rankings, exhibits higher data transitivity compared to general crowdsourced datasets, and significantly mitigates the risk of malicious manipulation. By fostering an open alliance between foundation models and real-world applications, Inclusion Arena aims to accelerate the development of LLMs and MLLMs truly optimized for practical, user-centric deployments. The platform is publicly accessible at https://www.tbox.cn/about/model-ranking.

[689] Mobile-Agent-v3: Fundamental Agents for GUI Automation

Jiabo Ye, Xi Zhang, Haiyang Xu, Haowei Liu, Junyang Wang, Zhaoqing Zhu, Ziwei Zheng, Feiyu Gao, Junjie Cao, Zhengxi Lu, Jitong Liao, Qi Zheng, Fei Huang, Jingren Zhou, Ming Yan

Main category: cs.AI

TL;DR: GUI-Owl is a foundational GUI agent model that achieves SOTA performance on 10 GUI benchmarks. Mobile-Agent-v3 framework built on it further improves performance, setting new SOTA for open-source GUI agents.

DetailsMotivation: To develop a comprehensive GUI agent that can handle diverse tasks across desktop and mobile environments, addressing the need for end-to-end models capable of grounding, QA, planning, decision-making, and procedural knowledge.

Method: Three key innovations: 1) Cloud-based virtual environment infrastructure enabling self-evolving trajectory production, 2) Integration of UI grounding, planning, action semantics and reasoning patterns, 3) Scalable RL framework with asynchronous training and Trajectory-aware Relative Policy Optimization.

Result: GUI-Owl-7B achieves 66.4 on AndroidWorld and 29.4 on OSWorld. Mobile-Agent-v3 improves to 73.3 on AndroidWorld and 37.7 on OSWorld. TRPO achieves 34.9 on OSWorld.

Conclusion: GUI-Owl represents a significant advancement in GUI agent capabilities, providing both a strong foundational model and an improved framework that sets new state-of-the-art performance for open-source GUI agents across multiple benchmarks.

Abstract: This paper introduces GUI-Owl, a foundational GUI agent model that achieves state-of-the-art performance among open-source end-to-end models on ten GUI benchmarks across desktop and mobile environments, covering grounding, question answering, planning, decision-making, and procedural knowledge. GUI-Owl-7B achieves 66.4 on AndroidWorld and 29.4 on OSWorld. Building on this, we propose Mobile-Agent-v3, a general-purpose GUI agent framework that further improves performance to 73.3 on AndroidWorld and 37.7 on OSWorld, setting a new state-of-the-art for open-source GUI agent frameworks. GUI-Owl incorporates three key innovations: (1) Large-scale Environment Infrastructure: a cloud-based virtual environment spanning Android, Ubuntu, macOS, and Windows, enabling our Self-Evolving GUI Trajectory Production framework. This generates high-quality interaction data via automated query generation and correctness validation, leveraging GUI-Owl to refine trajectories iteratively, forming a self-improving loop. It supports diverse data pipelines and reduces manual annotation. (2) Diverse Foundational Agent Capabilities: by integrating UI grounding, planning, action semantics, and reasoning patterns, GUI-Owl supports end-to-end decision-making and can act as a modular component in multi-agent systems. (3) Scalable Environment RL: we develop a scalable reinforcement learning framework with fully asynchronous training for real-world alignment. We also introduce Trajectory-aware Relative Policy Optimization (TRPO) for online RL, achieving 34.9 on OSWorld. GUI-Owl and Mobile-Agent-v3 are open-sourced at https://github.com/X-PLUG/MobileAgent.

[690] AI Chaperones Are (Really) All You Need to Prevent Parasocial Relationships with Chatbots

Emma Rath, Stuart Armstrong, Rebecca Gorman

Main category: cs.AI

TL;DR: AI chaperone agent detects parasocial chatbot conversations early using state-of-the-art language models, achieving perfect detection on synthetic data with no false positives.

DetailsMotivation: Address urgent need for safeguards against AI sycophancy and parasocial relationships with chatbots that can harm children and adults, as current methods lack effective mitigation.

Method: Developed a response evaluation framework using repurposed state-of-the-art language model to analyze ongoing conversations for parasocial cues. Tested on 30 synthetic dialogues spanning parasocial, sycophantic, and neutral conversations with five-stage iterative evaluation.

Result: Successfully identified all parasocial conversations with no false positives under unanimity rule. Detection typically occurred within first few exchanges of conversation.

Conclusion: AI chaperones show promise as viable solution for reducing risks of parasocial relationships with chatbots, providing preliminary evidence for effective early detection.

Abstract: Emerging reports of the harms caused to children and adults by AI sycophancy and by parasocial ties with chatbots point to an urgent need for safeguards against such risks. Yet, preventing such dynamics is challenging: parasocial cues often emerge gradually in private conversations between chatbots and users, and we lack effective methods to mitigate these risks. We address this challenge by introducing a simple response evaluation framework (an AI chaperone agent) created by repurposing a state-of-the-art language model to evaluate ongoing conversations for parasocial cues. We constructed a small synthetic dataset of thirty dialogues spanning parasocial, sycophantic, and neutral conversations. Iterative evaluation with five-stage testing successfully identified all parasocial conversations while avoiding false positives under a unanimity rule, with detection typically occurring within the first few exchanges. These findings provide preliminary evidence that AI chaperones can be a viable solution for reducing the risk of parasocial relationships.

[691] Route-and-Execute: Auditable Model-Card Matching and Specialty-Level Deployment

Shayan Vassef, Soorya Ram Shimegekar, Abhay Goyal, Koustuv Saha, Pi Zonooz, Navin Kumar

Main category: cs.AI

TL;DR: A healthcare framework using a single vision-language model for both routing medical images to appropriate specialist models and performing multiple downstream tasks within specialties, reducing fragmentation in clinical workflows.

DetailsMotivation: Clinical workflows are fragmented with multiple scripts and task-specific networks, lacking efficiency, data-driven model identification, and standardized output delivery, leading to increased operational costs.

Method: Two complementary solutions: 1) VLM as model-card matcher with three-stage routing workflow (modality -> abnormality -> model-card ID) with early exit checks and answer selection; 2) Fine-tuning VLM on specialty-specific datasets for multiple downstream tasks.

Result: The single-model deployment matches or approaches specialized baselines across gastroenterology, hematology, ophthalmology, and pathology specialties.

Conclusion: One VLM can both decide (route) and do (perform tasks), reducing data scientist effort, shortening monitoring, increasing transparency, and lowering integration overhead compared to multi-agent pipelines.

Abstract: Clinical workflows are fragmented as a patchwork of scripts and task-specific networks that often handle triage, task selection, and model deployment. These pipelines are rarely streamlined for data science pipeline, reducing efficiency and raising operational costs. Workflows also lack data-driven model identification (from imaging/tabular inputs) and standardized delivery of model outputs. In response, we present a practical, healthcare-first framework that uses a single vision-language model (VLM) in two complementary roles. First (Solution 1), the VLM acts as an aware model-card matcher that routes an incoming image to the appropriate specialist model via a three-stage workflow (modality -> primary abnormality -> model-card id). Checks are provided by (i) stagewise prompts that allow early exit via None/Normal/Other and (ii) a stagewise answer selector that arbitrates between the top-2 candidates at each stage, reducing the chance of an incorrect selection and aligning the workflow with clinical risk tolerance. Second (Solution 2), we fine-tune the VLM on specialty-specific datasets ensuring a single model covers multiple downstream tasks within each specialty, maintaining performance while simplifying deployment. Across gastroenterology, hematology, ophthalmology, and pathology, our single-model deployment matches or approaches specialized baselines. Compared with pipelines composed of many task-specific agents, this approach shows that one VLM can both decide and do. It may reduce effort by data scientists, shorten monitoring, increase the transparency of model selection (with per-stage justifications), and lower integration overhead.

[692] ST-Raptor: LLM-Powered Semi-Structured Table Question Answering

Zirui Tang, Boyu Niu, Xuanhe Zhou, Boxiu Li, Wei Zhou, Jiannan Wang, Guoliang Li, Xinyi Zhang, Fan Wu

Main category: cs.AI

TL;DR: ST-Raptor is a tree-based framework using LLMs for semi-structured table QA, outperforming baselines by up to 20% accuracy with hierarchical tree modeling and verification mechanisms.

DetailsMotivation: Existing methods struggle with semi-structured tables (financial reports, medical records) due to information loss during conversion or inability to handle complex layouts, requiring costly human interpretation.

Method: Proposes Hierarchical Orthogonal Tree (HO-Tree) to capture complex layouts, defines tree operations for LLMs, decomposes questions into sub-questions with operation pipelines, and uses two-stage verification (forward and backward validation).

Result: Outperforms nine baselines by up to 20% in answer accuracy on SSTQA dataset containing 764 questions over 102 real-world semi-structured tables.

Conclusion: ST-Raptor effectively automates semi-structured table QA by modeling complex layouts with tree structures and guiding LLMs through operation pipelines with verification, achieving significant accuracy improvements.

Abstract: Semi-structured tables, widely used in real-world applications (e.g., financial reports, medical records, transactional orders), often involve flexible and complex layouts (e.g., hierarchical headers and merged cells). These tables generally rely on human analysts to interpret table layouts and answer relevant natural language questions, which is costly and inefficient. To automate the procedure, existing methods face significant challenges. First, methods like NL2SQL require converting semi-structured tables into structured ones, which often causes substantial information loss. Second, methods like NL2Code and multi-modal LLM QA struggle to understand the complex layouts of semi-structured tables and cannot accurately answer corresponding questions. To this end, we propose ST-Raptor, a tree-based framework for semi-structured table question answering using large language models. First, we introduce the Hierarchical Orthogonal Tree (HO-Tree), a structural model that captures complex semi-structured table layouts, along with an effective algorithm for constructing the tree. Second, we define a set of basic tree operations to guide LLMs in executing common QA tasks. Given a user question, ST-Raptor decomposes it into simpler sub-questions, generates corresponding tree operation pipelines, and conducts operation-table alignment for accurate pipeline execution. Third, we incorporate a two-stage verification mechanism: forward validation checks the correctness of execution steps, while backward validation evaluates answer reliability by reconstructing queries from predicted answers. To benchmark the performance, we present SSTQA, a dataset of 764 questions over 102 real-world semi-structured tables. Experiments show that ST-Raptor outperforms nine baselines by up to 20% in answer accuracy. The code is available at https://github.com/weAIDB/ST-Raptor.

[693] Hermes 4 Technical Report

Ryan Teknium, Roger Jin, Jai Suphavadeeprasit, Dakota Mahan, Jeffrey Quesnelle, Joe Li, Chen Guang, Shannon Sands, Karan Malhotra

Main category: cs.AI

TL;DR: Hermes 4 is a family of hybrid reasoning models that combines structured multi-turn reasoning with broad instruction-following capabilities, addressing challenges in data curation, synthesis, training, and evaluation at scale.

DetailsMotivation: To develop models capable of both structured reasoning across multiple turns and general instruction following, addressing the challenges of scaling such hybrid approaches.

Method: Combines structured multi-turn reasoning with broad instruction-following through careful data curation, synthesis, training methodologies, and comprehensive evaluation across multiple domains.

Result: The models are comprehensively evaluated across mathematical reasoning, coding, knowledge, comprehension, and alignment benchmarks, with both quantitative performance metrics and qualitative behavioral analysis reported.

Conclusion: Hermes 4 represents a successful hybrid reasoning approach, with all model weights published publicly to support open research and community advancement.

Abstract: We present Hermes 4, a family of hybrid reasoning models that combine structured, multi-turn reasoning with broad instruction-following ability. We describe the challenges encountered during data curation, synthesis, training, and evaluation, and outline the solutions employed to address these challenges at scale. We comprehensively evaluate across mathematical reasoning, coding, knowledge, comprehension, and alignment benchmarks, and we report both quantitative performance and qualitative behavioral analysis. To support open research, all model weights are published publicly at https://huggingface.co/collections/NousResearch/hermes-4-collection-68a731bfd452e20816725728

[694] VistaWise: Building Cost-Effective Agent with Cross-Modal Knowledge Graph for Minecraft

Honghao Fu, Junlong Ren, Qi Chai, Deheng Ye, Yujun Cai, Hao Wang

Main category: cs.AI

TL;DR: VistaWise is a cost-effective agent framework that integrates cross-modal knowledge and finetunes object detection to reduce domain-specific training needs from millions to hundreds of samples, achieving SOTA performance in open-world tasks.

DetailsMotivation: LLMs show promise in embodied decision-making but are hindered by lack of domain-specific knowledge, and existing methods require prohibitive development costs for large-scale domain-specific data finetuning.

Method: Integrates visual information and textual dependencies into cross-modal knowledge graph, uses retrieval-based pooling strategy for task-related information extraction, and employs desktop-level skill library for direct Minecraft client operation via mouse/keyboard.

Result: Achieves state-of-the-art performance across various open-world tasks, demonstrating effectiveness in reducing development costs while enhancing agent performance.

Conclusion: VistaWise provides an effective framework that significantly reduces domain-specific training requirements while improving embodied agent performance in open-world environments.

Abstract: Large language models (LLMs) have shown significant promise in embodied decision-making tasks within virtual open-world environments. Nonetheless, their performance is hindered by the absence of domain-specific knowledge. Methods that finetune on large-scale domain-specific data entail prohibitive development costs. This paper introduces VistaWise, a cost-effective agent framework that integrates cross-modal domain knowledge and finetunes a dedicated object detection model for visual analysis. It reduces the requirement for domain-specific training data from millions of samples to a few hundred. VistaWise integrates visual information and textual dependencies into a cross-modal knowledge graph (KG), enabling a comprehensive and accurate understanding of multimodal environments. We also equip the agent with a retrieval-based pooling strategy to extract task-related information from the KG, and a desktop-level skill library to support direct operation of the Minecraft desktop client via mouse and keyboard inputs. Experimental results demonstrate that VistaWise achieves state-of-the-art performance across various open-world tasks, highlighting its effectiveness in reducing development costs while enhancing agent performance.

[695] Building Self-Evolving Agents via Experience-Driven Lifelong Learning: A Framework and Benchmark

Yuxuan Cai, Yipeng Hao, Jie Zhou, Hang Yan, Zhikai Lei, Rui Zhen, Zhenhua Han, Yutao Yang, Junsong Li, Qianjun Pan, Tianyu Huai, Qin Chen, Xin Li, Kai Chen, Bo Zhang, Xipeng Qiu, Liang He

Main category: cs.AI

TL;DR: The paper introduces Experience-driven Lifelong Learning (ELL), a framework for creating self-evolving AI agents that learn continuously through real-world interaction, built on four core principles: experience exploration, long-term memory, skill learning, and knowledge internalization.

DetailsMotivation: As AI advances toward general intelligence, there's a need to shift from systems optimized for static tasks to creating open-ended agents that can learn continuously through real-world interaction.

Method: The ELL framework is built on four core principles: (1) Experience Exploration - continuous self-motivated interaction with dynamic environments, (2) Long-term Memory - preserving and structuring historical knowledge, (3) Skill Learning - abstracting patterns into reusable skills, and (4) Knowledge Internalization - converting experiences into intuitive capabilities. The authors also introduce StuLife, a benchmark dataset simulating a student’s college journey.

Result: The paper presents a comprehensive framework for lifelong learning agents but does not provide specific experimental results in the abstract. The StuLife benchmark is introduced as an evaluation tool.

Conclusion: The ELL framework provides a structured approach to building self-evolving AI agents capable of continuous growth, addressing the shift from static task optimization to open-ended learning through the proposed four principles and benchmark dataset.

Abstract: As AI advances toward general intelligence, the focus is shifting from systems optimized for static tasks to creating open-ended agents that learn continuously. In this paper, we introduce Experience-driven Lifelong Learning (ELL), a framework for building self-evolving agents capable of continuous growth through real-world interaction. The framework is built on four core principles: (1) Experience Exploration: Agents learn through continuous, self-motivated interaction with dynamic environments, navigating interdependent tasks and generating rich experiential trajectories. (2) Long-term Memory: Agents preserve and structure historical knowledge, including personal experiences, domain expertise, and commonsense reasoning, into a persistent memory system. (3) Skill Learning: Agents autonomously improve by abstracting recurring patterns from experience into reusable skills, which are actively refined and validated for application in new tasks. (4) Knowledge Internalization: Agents internalize explicit and discrete experiences into implicit and intuitive capabilities as “second nature”. We also introduce StuLife, a benchmark dataset for ELL that simulates a student’s holistic college journey, from enrollment to academic and personal development, across three core phases and ten detailed sub-scenarios. StuLife is designed around three key paradigm

[696] MAB Optimizer for Estimating Math Question Difficulty via Inverse CV without NLP

Surajit Das, Gourav Roy, Aleksei Eliseev, Ram Kumar Rajendran

Main category: cs.AI

TL;DR: APME framework uses reinforcement learning to estimate question difficulty from solver performance data (marks and time) without linguistic features or expert labels, achieving high accuracy across diverse educational contexts.

DetailsMotivation: Traditional human labeling is subjective and existing NLP approaches fail in symbolic domains like algebra, creating need for objective, domain-agnostic methods for determining question difficulty in Intelligent Tutoring Systems.

Method: Reinforcement learning-based Multi-Armed Bandit framework using solver performance data (marks obtained and time taken) with inverse coefficient of variation as risk-adjusted metric for adaptive assessment.

Result: Achieved average R2 of 0.9213 and average RMSE of 0.0584 across three heterogeneous datasets, consistently outperforming regression-based, NLP-driven, and IRT baseline models.

Conclusion: Domain-agnostic, self-supervised approach effectively estimates question difficulty, aligns with pedagogical principles, and can be extended to any domain with solver interaction data.

Abstract: The evolution of technology and education is driving the emergence of Intelligent & Autonomous Tutoring Systems (IATS), where objective and domain-agnostic methods for determining question difficulty are essential. Traditional human labeling is subjective, and existing NLP-based approaches fail in symbolic domains like algebra. This study introduces the Approach of Passive Measures among Educands (APME), a reinforcement learning-based Multi-Armed Bandit (MAB) framework that estimates difficulty solely from solver performance data – marks obtained and time taken – without requiring linguistic features or expert labels. By leveraging the inverse coefficient of variation as a risk-adjusted metric, the model provides an explainable and scalable mechanism for adaptive assessment. Empirical validation was conducted on three heterogeneous datasets. Across these diverse contexts, the model achieved an average R2 of 0.9213 and an average RMSE of 0.0584, confirming its robustness, accuracy, and adaptability to different educational levels and assessment formats. Compared with baseline approaches-such as regression-based, NLP-driven, and IRT models-the proposed framework consistently outperformed alternatives, particularly in purely symbolic domains. The findings highlight that (i) item heterogeneity strongly influences perceived difficulty, and (ii) variance in solver outcomes is as critical as mean performance for adaptive allocation. Pedagogically, the model aligns with Vygotskys Zone of Proximal Development by identifying tasks that balance challenge and attainability, supporting motivation while minimizing disengagement. This domain-agnostic, self-supervised approach advances difficulty tagging in IATS and can be extended beyond algebra wherever solver interaction data is available

[697] Instructional Agents: LLM Agents on Automated Course Material Generation for Teaching Faculties

Huaiyuan Yao, Wanpeng Xu, Justin Turnau, Nadia Kellam, Hua Wei

Main category: cs.AI

TL;DR: Instructional Agents is a multi-agent LLM framework that automates end-to-end course material generation through role-based collaboration, significantly reducing development time while maintaining quality.

DetailsMotivation: High-quality instructional material preparation is labor-intensive and requires extensive coordination among faculty, designers, and TAs. There's a need to democratize access to quality education in resource-constrained settings.

Method: Multi-agent LLM framework with four operational modes (Autonomous, Catalog-Guided, Feedback-Guided, Full Co-Pilot) that simulates role-based educational collaboration to generate cohesive course materials including syllabus, lectures, slides, and assessments.

Result: Evaluated across five university-level computer science courses, the system produces high-quality instructional materials while significantly reducing development time and human workload.

Conclusion: Instructional Agents provides a scalable, cost-effective framework to support institutions with limited instructional design capacity, democratizing access to high-quality education particularly in underserved settings.

Abstract: Preparing high-quality instructional materials remains a labor-intensive process that often requires extensive coordination among teaching faculty, instructional designers, and teaching assistants. In this work, we present Instructional Agents, a multi-agent large language model (LLM) framework designed to automate end-to-end course material generation, including syllabus creation, lecture scripts, LaTeX-based slides, and assessments. Unlike existing AI-assisted educational tools that focus on isolated tasks, Instructional Agents simulates role-based collaboration among educational agents to produce cohesive and pedagogically aligned content. The system operates in four modes: Autonomous, Catalog-Guided, Feedback-Guided, and Full Co-Pilot mode, enabling flexible control over the degree of human involvement. We evaluate Instructional Agents across five university-level computer science courses and show that it produces high-quality instructional materials while significantly reducing development time and human workload. By supporting institutions with limited instructional design capacity, Instructional Agents provides a scalable and cost-effective framework to democratize access to high-quality education, particularly in underserved or resource-constrained settings.

[698] AWorld: Orchestrating the Training Recipe for Agentic AI

Chengyue Yu, Siyuan Lu, Chenyi Zhuang, Dong Wang, Qintong Wu, Zongyue Li, Runsheng Gan, Chunfeng Wang, Siqi Hou, Gaochi Huang, Wenlong Yan, Lifeng Hong, Aohui Xue, Yanfeng Wang, Jinjie Gu, David Tsai, Tao Lin

Main category: cs.AI

TL;DR: AWorld is an open-source distributed system that accelerates agent-environment interaction by 14.6x, enabling efficient reinforcement learning and producing a Qwen3-32B agent that outperforms GPT-4o on GAIA benchmark.

DetailsMotivation: The learning from practice paradigm is crucial for Agentic AI but suffers from inefficient experience generation, especially in complex benchmarks like GAIA, creating a significant bottleneck.

Method: Developed AWorld, an open-source distributed system that distributes tasks across clusters to accelerate experience collection compared to standard single-node sequential execution.

Result: Achieved 14.6x speedup in experience collection and trained a Qwen3-32B agent that reaches 32.23% pass@1 accuracy on GAIA test set, surpassing GPT-4o (27.91%) and rivaling DeepSeek-V3 (31.89%).

Conclusion: AWorld provides a practical blueprint for complete agentic AI training pipeline, making extensive reinforcement learning practical and scalable from efficient interaction to demonstrable model improvement.

Abstract: The learning from practice paradigm is crucial for developing capable Agentic AI systems, yet it is severely hampered by inefficient experience generation, a bottleneck especially pronounced in complex benchmarks like GAIA. To address this, we introduce AWorld, an open-source system engineered for large-scale agent-environment interaction. By distributing tasks across a cluster, AWorld accelerates experience collection by 14.6x compared to standard single-node, sequential execution. This critical speedup makes extensive reinforcement learning practical and scalable. Leveraging this capability, we trained a Qwen3-32B-based agent that achieves pass@1 accuracy of 32.23% on the GAIA test set, which surpasses GPT-4o (27.91%) and rivals DeepSeek-V3 (31.89%). Our open-source system and the resulting agent provide a practical blueprint for a complete agentic AI training pipeline, from efficient interaction to demonstrable model improvement.

cs.SD

[699] From Sound to Sight: Towards AI-authored Music Videos

Leo Vitasovic, Stella Graßhof, Agnes Mercedes Kloft, Ville V. Lehtola, Martin Cunneen, Justyna Starostka, Glenn McGarry, Kun Li, Sami S. Brandt

Main category: cs.SD

TL;DR: Novel AI pipelines for automatic music video generation using deep learning models to analyze audio features and generate corresponding video content with emotional alignment.

DetailsMotivation: Traditional music visualization systems are limited by handcrafted transformations, lacking expressiveness and automation. The paper aims to create more sophisticated, automated music video generation using modern AI techniques.

Method: Two novel pipelines using off-the-shelf deep learning models: 1) audio analysis to detect musical qualities and emotional cues, 2) text-to-video generation using language models to create scene descriptions and generative models to produce corresponding video clips.

Result: Preliminary user evaluation showed promising results in storytelling potential, visual coherency, and emotional alignment with the music, demonstrating the effectiveness of latent feature techniques.

Conclusion: Latent feature techniques and deep generative models have significant potential to expand music visualization beyond traditional approaches, offering more expressive and automated music video generation.

Abstract: Conventional music visualisation systems rely on handcrafted ad hoc transformations of shapes and colours that offer only limited expressiveness. We propose two novel pipelines for automatically generating music videos from any user-specified, vocal or instrumental song using off-the-shelf deep learning models. Inspired by the manual workflows of music video producers, we experiment on how well latent feature-based techniques can analyse audio to detect musical qualities, such as emotional cues and instrumental patterns, and distil them into textual scene descriptions using a language model. Next, we employ a generative model to produce the corresponding video clips. To assess the generated videos, we identify several critical aspects and design and conduct a preliminary user evaluation that demonstrates storytelling potential, visual coherency and emotional alignment with the music. Our findings underscore the potential of latent feature techniques and deep generative models to expand music visualisation beyond traditional approaches.

[700] A Survey on Evaluation Metrics for Music Generation

Faria Binte Kader, Santu Karmaker

Main category: cs.SD

TL;DR: This paper addresses the research gap in music generation evaluation by proposing a taxonomy for evaluation metrics and identifying limitations in current methodologies, while suggesting future directions for comprehensive evaluation frameworks.

DetailsMotivation: Despite advancements in music generation systems, evaluation methodologies have not kept pace due to music's complex nature including structure, coherence, creativity, and emotional expressiveness.

Method: The paper introduces a detailed taxonomy for evaluation metrics for both audio and symbolic music representations, and provides a critical review of current evaluation methodologies.

Result: The review identifies major limitations including poor correlation between objective metrics and human perception, cross-cultural bias, and lack of standardization hindering cross-model comparisons.

Conclusion: The paper proposes future research directions towards building a comprehensive evaluation framework for music generation evaluation to address the identified gaps.

Abstract: Despite significant advancements in music generation systems, the methodologies for evaluating generated music have not progressed as expected due to the complex nature of music, with aspects such as structure, coherence, creativity, and emotional expressiveness. In this paper, we shed light on this research gap, introducing a detailed taxonomy for evaluation metrics for both audio and symbolic music representations. We include a critical review identifying major limitations in current evaluation methodologies which includes poor correlation between objective metrics and human perception, cross-cultural bias, and lack of standardization that hinders cross-model comparisons. Addressing these gaps, we further propose future research directions towards building a comprehensive evaluation framework for music generation evaluation.

[701] Algorithms for Collaborative Harmonization

Eyal Briman, Eyal Leizerovich, Nimrod Talmon

Main category: cs.SD

TL;DR: This paper presents aggregation algorithms for musical harmonization that balance collective representation and musical coherence, finding Kemeny and plurality-based methods most effective.

DetailsMotivation: Musical harmonization shares similarities with text aggregation but operates in a more structured language. The research aims to develop aggregation methods that can combine multiple harmonization suggestions while maintaining both collective representation and musical quality.

Method: The authors present different algorithms for aggregating harmonies from multiple agents, analyzing their computational complexities. The methods include Kemeny and plurality-based aggregation approaches specifically designed for musical harmonization scenarios.

Result: The results show that Kemeny and plurality-based algorithms are most effective in achieving both representation of collective suggestions and maintaining musical coherence in the aggregated harmonization sequences.

Conclusion: For musical harmonization aggregation, Kemeny and plurality-based algorithms provide the best balance between representing collective input and ensuring musical coherence, making them suitable for structured aggregation tasks in musical domains.

Abstract: We consider a specific scenario of text aggregation, in the realm of musical harmonization. Musical harmonization shares similarities with text aggregation, however the language of harmony is more structured than general text. Concretely, given a set of harmonization suggestions for a given musical melody, our interest lies in devising aggregation algorithms that yield an harmonization sequence that satisfies the following two key criteria: (1) an effective representation of the collective suggestions; and (2) an harmonization that is musically coherent. We present different algorithms for the aggregation of harmonies given by a group of agents and analyze their complexities. The results indicate that the Kemeny and plurality-based algorithms are most effective in assessing representation and maintaining musical coherence.

[702] CoComposer: LLM Multi-agent Collaborative Music Composition

Peiwen Xing, Aske Plaat, Niki van Stein

Main category: cs.SD

TL;DR: CoComposer is a multi-agent system that improves music composition quality and controllability compared to existing LLM-based systems, though still lags behind specialized MusicLM in pure music quality.

DetailsMotivation: Existing AI music composition tools have limitations in generation duration, musical quality, and controllability, creating a need for better systems.

Method: A multi-agent system with five collaborating agents based on traditional music composition workflow, using AudioBox-Aesthetics system and tested with three LLMs (GPT-4o, DeepSeek-V3-0324, Gemini-2.5-Flash).

Result: CoComposer outperforms existing multi-agent LLM-based systems in music quality and beats single-agent systems in production complexity. It has better interpretability and editability than MusicLM, though MusicLM produces better music.

Conclusion: Multi-agent approach improves music composition quality and controllability over existing LLM systems, but specialized music models like MusicLM still excel in pure audio quality.

Abstract: Existing AI Music composition tools are limited in generation duration, musical quality, and controllability. We introduce CoComposer, a multi-agent system that consists of five collaborating agents, each with a task based on the traditional music composition workflow. Using the AudioBox-Aesthetics system, we experimentally evaluate CoComposer on four compositional criteria. We test with three LLMs (GPT-4o, DeepSeek-V3-0324, Gemini-2.5-Flash), and find (1) that CoComposer outperforms existing multi-agent LLM-based systems in music quality, and (2) compared to a single-agent system, in production complexity. Compared to non- LLM MusicLM, CoComposer has better interpretability and editability, although MusicLM still produces better music.

[703] The Name-Free Gap: Policy-Aware Stylistic Control in Music Generation

Ashwin Nagarajan, Hao-Wen Dong

Main category: cs.SD

TL;DR: Lightweight LLM-generated descriptors can provide effective stylistic control for music generation without using artist names, though artist names remain the strongest signal. This reveals limitations in current artist name restriction policies.

DetailsMotivation: Existing music stylization methods require retraining or specialized conditioning, complicating reproducibility and limiting policy compliance when artist names are restricted. The paper explores whether lightweight, human-readable modifiers can provide policy-robust stylistic control.

Method: Used MusicGen-small to evaluate two artists (Billie Eilish and Ludovico Einaudi) with 15 reference excerpts each. Compared baseline prompts, artist-name prompts, and five descriptor sets generated by LLM. Evaluation used VGGish and CLAP embeddings with distributional similarity and a new min-distance attribution metric.

Result: Artist names were the strongest control signal, but name-free descriptors recovered much of this effect. Cross-artist transfers reduced alignment, showing descriptors encode targeted stylistic cues. The study presents a descriptor table across ten contemporary artists.

Conclusion: Existing safeguards restricting artist names may not fully prevent style imitation. The research defines the ’name-free gap’ and provides a reproducible evaluation protocol for prompt-level controllability in music generation.

Abstract: Text-to-music models capture broad attributes such as instrumentation or mood, but fine-grained stylistic control remains an open challenge. Existing stylization methods typically require retraining or specialized conditioning, which complicates reproducibility and limits policy compliance when artist names are restricted. We study whether lightweight, human-readable modifiers sampled from a large language model can provide a policy-robust alternative for stylistic control. Using MusicGen-small, we evaluate two artists: Billie Eilish (vocal pop) and Ludovico Einaudi (instrumental piano). For each artist, we use fifteen reference excerpts and evaluate matched seeds under three conditions: baseline prompts, artist-name prompts, and five descriptor sets. All prompts are generated using a large language model. Evaluation uses both VGGish and CLAP embeddings with distributional and per-clip similarity measures, including a new min-distance attribution metric. Results show that artist names are the strongest control signal across both artists, while name-free descriptors recover much of this effect. This highlights that existing safeguards such as the restriction of artist names in music generation prompts may not fully prevent style imitation. Cross-artist transfers reduce alignment, showing that descriptors encode targeted stylistic cues. We also present a descriptor table across ten contemporary artists to illustrate the breadth of the tokens. Together these findings define the name-free gap, the controllability difference between artist-name prompts and policy-compliant descriptors, shown through a reproducible evaluation protocol for prompt-level controllability.

[704] Generalizable Audio Spoofing Detection using Non-Semantic Representations

Arnab Das, Yassine El Kheir, Carlos Franzreb, Tim Herzig, Tim Polzehl, Sebastian Möller

Main category: cs.SD

TL;DR: Proposes using non-semantic universal audio representations for deepfake detection, achieving superior generalization on out-of-domain data compared to existing methods.

DetailsMotivation: Existing deepfake detection solutions lack generalizability and fail on real-world data, creating urgent need for robust countermeasures against synthetic audio spoofing attacks.

Method: Leverages non-semantic universal audio representations using TRILL and TRILLsson models for spoofing detection.

Result: Achieves comparable in-domain performance while significantly outperforming state-of-the-art approaches on out-of-domain test sets, showing superior generalization on public-domain data.

Conclusion: Non-semantic audio representations provide a highly generalizable solution for deepfake detection that works better across diverse real-world scenarios than hand-crafted features, semantic embeddings, or end-to-end architectures.

Abstract: Rapid advancements in generative modeling have made synthetic audio generation easy, making speech-based services vulnerable to spoofing attacks. Consequently, there is a dire need for robust countermeasures more than ever. Existing solutions for deepfake detection are often criticized for lacking generalizability and fail drastically when applied to real-world data. This study proposes a novel method for generalizable spoofing detection leveraging non-semantic universal audio representations. Extensive experiments have been performed to find suitable non-semantic features using TRILL and TRILLsson models. The results indicate that the proposed method achieves comparable performance on the in-domain test set while significantly outperforming state-of-the-art approaches on out-of-domain test sets. Notably, it demonstrates superior generalization on public-domain data, surpassing methods based on hand-crafted features, semantic embeddings, and end-to-end architectures.

[705] Evaluating the Effectiveness of Transformer Layers in Wav2Vec 2.0, XLS-R, and Whisper for Speaker Identification Tasks

Linus Stuhlmann, Michael Alexander Saxer

Main category: cs.SD

TL;DR: Evaluation of Wav2Vec 2.0, XLS-R, and Whisper speech encoders for speaker identification, analyzing layer-wise representations and optimal transformer layer configurations.

DetailsMotivation: To assess and compare the performance of advanced speech encoder models in speaker identification tasks and understand how different layers capture speaker-specific features.

Method: Fine-tuned Wav2Vec 2.0, XLS-R, and Whisper models, then analyzed their layer-wise representations using SVCCA, k-means clustering, and t-SNE visualizations to determine optimal transformer layer configurations.

Result: Wav2Vec 2.0 and XLS-R effectively capture speaker-specific features in early layers with improved stability after fine-tuning. Whisper performs better in deeper layers. Optimal number of transformer layers for each model was determined.

Conclusion: Different speech encoder models have varying layer-wise effectiveness for speaker identification, with Wav2Vec 2.0 and XLS-R excelling in early layers and Whisper in deeper layers, providing guidance for optimal model configuration in speaker ID tasks.

Abstract: This study evaluates the performance of three advanced speech encoder models, Wav2Vec 2.0, XLS-R, and Whisper, in speaker identification tasks. By fine-tuning these models and analyzing their layer-wise representations using SVCCA, k-means clustering, and t-SNE visualizations, we found that Wav2Vec 2.0 and XLS-R capture speaker-specific features effectively in their early layers, with fine-tuning improving stability and performance. Whisper showed better performance in deeper layers. Additionally, we determined the optimal number of transformer layers for each model when fine-tuned for speaker identification tasks.

[706] Towards High-Fidelity and Controllable Bioacoustic Generation via Enhanced Diffusion Learning

Tianyu Song, Ton Viet Ta

Main category: cs.SD

TL;DR: BirdDiff is a generative framework that synthesizes high-fidelity bird calls from noisy field recordings using multi-scale enhancement and diffusion-based generation with multi-modal conditioning.

DetailsMotivation: To support biomonitoring and supplement scarce data for endangered species by generating realistic animal vocalizations from noisy field recordings, which remains a major challenge in bioacoustics.

Method: A two-stage framework: 1) ‘zeroth layer’ multi-scale adaptive bird-call enhancement for SNR improvement, 2) diffusion-based generator conditioned on MFCCs, species labels, and textual descriptions.

Result: Achieved +10.45 dB SNR gain, lowest spectral distortion (0.54 ISD), improved FAD (0.213), JSD (0.226), and classification accuracy of 70.1% (vs 35.9% baseline) with 8/12 species exceeding 70% accuracy.

Conclusion: BirdDiff enables high-fidelity, controllable bird call generation directly from noisy field recordings, demonstrating significant improvements over baseline methods across multiple quality metrics.

Abstract: Generative modeling offers new opportunities for bioacoustics, enabling the synthesis of realistic animal vocalizations that could support biomonitoring efforts and supplement scarce data for endangered species. However, directly generating bird call waveforms from noisy field recordings remains a major challenge. We propose BirdDiff, a generative framework designed to synthesize bird calls from a noisy dataset of 12 wild bird species. The model incorporates a “zeroth layer” stage for multi-scale adaptive bird-call enhancement, followed by a diffusion-based generator conditioned on three modalities: Mel-frequency cepstral coefficients, species labels, and textual descriptions. The enhancement stage improves signal-to-noise ratio (SNR) while minimizing spectral distortion, achieving the highest SNR gain (+10.45 dB) and lowest Itakura-Saito Distance (0.54) compared to three widely used non-training enhancement methods. We evaluate BirdDiff against a baseline generative model, DiffWave. Our method yields substantial improvements in generative quality metrics: Fr'echet Audio Distance (0.590 to 0.213), Jensen-Shannon Divergence (0.259 to 0.226), and Number of Statistically-Different Bins (7.33 to 5.58). To assess species-specific detail preservation, we use a ResNet50 classifier trained on the original dataset to identify generated samples. Classification accuracy improves from 35.9% (DiffWave) to 70.1% (BirdDiff), with 8 of 12 species exceeding 70% accuracy. These results demonstrate that BirdDiff enables high-fidelity, controllable bird call generation directly from noisy field recordings.

[707] SaD: A Scenario-Aware Discriminator for Speech Enhancement

Xihao Yuan, Siqi Liu, Yan Chen, Hang Zhou, Chang Liu, Hanting Chen, Jie Hu

Main category: cs.SD

TL;DR: Proposes a scenario-aware discriminator for GAN-based speech enhancement that captures scene-specific features and performs frequency-domain division to improve quality assessment without changing generator architectures.

DetailsMotivation: Current GAN optimization strategies focus mainly on generator architecture refinement or discriminator quality metrics, overlooking rich contextual information from diverse scenarios.

Method: Developed a scenario-aware discriminator that captures scene-specific features and performs frequency-domain division for more accurate quality assessment of enhanced speech.

Result: Comprehensive experiments on three representative models using two public datasets show the method effectively adapts to various generator architectures without structural changes, unlocking further performance gains.

Conclusion: The proposed scenario-aware discriminator enables better speech enhancement performance across different scenarios by leveraging contextual information and frequency-domain analysis.

Abstract: Generative adversarial network-based models have shown remarkable performance in the field of speech enhancement. However, the current optimization strategies for these models predominantly focus on refining the architecture of the generator or enhancing the quality evaluation metrics of the discriminator. This approach often overlooks the rich contextual information inherent in diverse scenarios. In this paper, we propose a scenario-aware discriminator that captures scene-specific features and performs frequency-domain division, thereby enabling a more accurate quality assessment of the enhanced speech generated by the generator. We conducted comprehensive experiments on three representative models using two publicly available datasets. The results demonstrate that our method can effectively adapt to various generator architectures without altering their structure, thereby unlocking further performance gains in speech enhancement across different scenarios.

[708] PicoAudio2: Temporal Controllable Text-to-Audio Generation with Natural Language Description

Zihao Zheng, Zeyu Xie, Xuenan Xu, Wen Wu, Chao Zhang, Mengyue Wu

Main category: cs.SD

TL;DR: PicoAudio2 improves text-to-audio generation with better temporal control and audio quality by combining real annotated data with simulated data, using a timestamp matrix for fine-grained control.

DetailsMotivation: Existing text-to-audio generation methods have limited sound event categories and rely only on simulated data, which restricts audio quality and generalization to real data.

Method: Uses a grounding model to annotate event timestamps in real audio-text datasets, combines real and simulated data for training, and encodes timestamp information into a timestamp matrix for fine-grained control.

Result: PicoAudio2 shows superior performance in temporal controllability and audio quality compared to existing methods.

Conclusion: The proposed approach of combining real annotated data with simulated data and using timestamp matrices significantly improves text-to-audio generation performance.

Abstract: Controllable text-to-audio generation (TTA) has attracted much attention recently. Although existing works can achieve fine-grained controllability based on timestamp information, sound event categories are limited to a fixed set. Moreover, since only simulated data is used for training, the generated audio quality and generalization performance on real data are limited. To tackle this issue, we propose PicoAudio2, improving temporal-controllable TTA via a new data processing pipeline and model architecture. Specifically, we use a grounding model to annotate event timestamps of real audio-text datasets to curate temporally-strong real data, in addition to simulation data from existing works. The model is trained on the combination of real and simulation data. Moreover, following PicoAudio, we encode timestamp information into a timestamp matrix to provide extra fine-grained time-aligned information to the model, on top of the coarse-grained textual description. Experiments show that PicoAudio2 exhibits superior performance in terms of temporal controllability and audio quality.

[709] AImoclips: A Benchmark for Evaluating Emotion Conveyance in Text-to-Music Generation

Gyehun Go, Satbyul Han, Ahyeon Choi, Eunjin Choi, Juhan Nam, Jeong Mi Park

Main category: cs.SD

TL;DR: AImoclips benchmark evaluates text-to-music systems’ emotional fidelity, revealing commercial models produce overly pleasant music while open-source models under-deliver on emotion, with all systems showing bias toward emotional neutrality.

DetailsMotivation: Current text-to-music generation systems focus on text alignment and human preference but lack evaluation of emotional fidelity - how well they convey intended emotions to human listeners.

Method: Created benchmark with 12 emotion intents across valence-arousal space, generated over 1,000 music clips using 6 state-of-the-art TTM systems, and had 111 participants rate perceived valence/arousal on 9-point Likert scale.

Result: Commercial systems produce music perceived as more pleasant than intended; open-source systems produce less pleasant music. High-arousal emotions conveyed more accurately. All systems show bias toward emotional neutrality.

Conclusion: The benchmark reveals key limitations in affective controllability of TTM systems and provides insights for developing emotionally aligned music generation models.

Abstract: Recent advances in text-to-music (TTM) generation have enabled controllable and expressive music creation using natural language prompts. However, the emotional fidelity of TTM systems remains largely underexplored compared to human preference or text alignment. In this study, we introduce AImoclips, a benchmark for evaluating how well TTM systems convey intended emotions to human listeners, covering both open-source and commercial models. We selected 12 emotion intents spanning four quadrants of the valence-arousal space, and used six state-of-the-art TTM systems to generate over 1,000 music clips. A total of 111 participants rated the perceived valence and arousal of each clip on a 9-point Likert scale. Our results show that commercial systems tend to produce music perceived as more pleasant than intended, while open-source systems tend to perform the opposite. Emotions are more accurately conveyed under high-arousal conditions across all models. Additionally, all systems exhibit a bias toward emotional neutrality, highlighting a key limitation in affective controllability. This benchmark offers valuable insights into model-specific emotion rendering characteristics and supports future development of emotionally aligned TTM systems.

[710] Adaptive Vehicle Speed Classification via BMCNN with Reinforcement Learning-Enhanced Acoustic Processing

Yuli Zhang, Pengfei Fan, Ruiyuan Jiang, Hankang Gu, Dongyao Jia, Xinheng Wang

Main category: cs.SD

TL;DR: Hybrid deep learning + reinforcement learning framework for acoustic vehicle speed classification using dual-branch CNN with MFCC/wavelet features and attention-enhanced DQN for early decision making

DetailsMotivation: Traffic congestion requires intelligent transportation systems for real-time management, needing efficient and accurate vehicle speed classification methods

Method: Dual-branch BMCNN processes MFCC and wavelet features to capture complementary frequency patterns, combined with attention-enhanced DQN that adaptively selects minimal audio frames and triggers early decisions when confidence thresholds are reached

Result: 95.99% accuracy on IDMT-Traffic dataset and 92.3% on SZUR-Acoustic dataset, with up to 1.63x faster average processing via early termination, outperforming A3C, DDDQN, SA2C, PPO, and TD3 methods

Conclusion: The method provides superior accuracy-efficiency trade-off and is suitable for real-time ITS deployment in heterogeneous urban environments

Abstract: Traffic congestion remains a pressing urban challenge, requiring intelligent transportation systems for real-time management. We present a hybrid framework that combines deep learning and reinforcement learning for acoustic vehicle speed classification. A dual-branch BMCNN processes MFCC and wavelet features to capture complementary frequency patterns. An attention-enhanced DQN adaptively selects the minimal number of audio frames and triggers early decisions once confidence thresholds are reached. Evaluations on IDMT-Traffic and our SZUR-Acoustic (Suzhou) datasets show 95.99% and 92.3% accuracy, with up to 1.63x faster average processing via early termination. Compared with A3C, DDDQN, SA2C, PPO, and TD3, the method provides a superior accuracy-efficiency trade-off and is suitable for real-time ITS deployment in heterogeneous urban environments.

[711] Speech Command Recognition Using LogNNet Reservoir Computing for Embedded Systems

Yuriy Izotov, Andrei Velichko

Main category: cs.SD

TL;DR: Low-resource speech command recognizer using energy-based VAD, optimized MFCC pipeline, and LogNNet classifier achieves 92% accuracy with minimal hardware requirements suitable for IoT devices.

DetailsMotivation: To develop a speech recognition system that can operate under strict memory and compute constraints for battery-powered IoT devices and wireless sensor networks.

Method: Combines energy-based voice activity detection, optimized MFCC pipeline with adaptive binning (64-dimensional features), and LogNNet reservoir-computing classifier with 64:33:9:4 architecture.

Result: Achieves 92.04% accuracy in speaker-independent evaluation, requires significantly fewer parameters than deep learning models, and hardware implementation on Arduino Nano uses only 18KB RAM (55% utilization) with ~90% real-time accuracy.

Conclusion: The complete pipeline enables reliable on-device speech-command recognition under strict memory and compute limits, making it suitable for battery-powered IoT applications.

Abstract: This paper presents a low-resource speech-command recognizer combining energy-based voice activity detection (VAD), an optimized Mel-Frequency Cepstral Coefficients (MFCC) pipeline, and the LogNNet reservoir-computing classifier. Using four commands from the Speech Commands da-taset downsampled to 8 kHz, we evaluate four MFCC aggregation schemes and find that adaptive binning (64-dimensional feature vector) offers the best accuracy-to-compactness trade-off. The LogNNet classifier with architecture 64:33:9:4 reaches 92.04% accuracy under speaker-independent evaluation, while requiring significantly fewer parameters than conventional deep learn-ing models. Hardware implementation on Arduino Nano 33 IoT (ARM Cor-tex-M0+, 48 MHz, 32 KB RAM) validates the practical feasibility, achieving ~90% real-time recognition accuracy while consuming only 18 KB RAM (55% utilization). The complete pipeline (VAD -> MFCC -> LogNNet) thus enables reliable on-device speech-command recognition under strict memory and compute limits, making it suitable for battery-powered IoT nodes, wire-less sensor networks, and hands-free control interfaces.

[712] From Discord to Harmony: Decomposed Consonance-based Training for Improved Audio Chord Estimation

Andrea Poltronieri, Xavier Serra, Martín Rocamora

Main category: cs.SD

TL;DR: This paper addresses challenges in Audio Chord Estimation (ACE) by proposing consonance-based evaluation metrics and a novel conformer-based model that integrates consonance concepts and handles class imbalance through decomposed chord estimation.

DetailsMotivation: ACE faces challenges including annotator subjectivity causing inconsistent labels and class imbalance in datasets, leading to performance plateaus in existing systems.

Method: Proposed consonance-informed distance metric for perceptual similarity evaluation, and a conformer-based ACE model with consonance-based label smoothing that decomposes chord estimation into root, bass, and note activations.

Result: Consonance-based metrics better capture musically meaningful agreement between annotations, and the proposed model addresses class imbalance issues.

Conclusion: Incorporating consonance concepts into both evaluation metrics and model architecture provides more effective solutions for ACE challenges, particularly in handling annotator subjectivity and class imbalance.

Abstract: Audio Chord Estimation (ACE) holds a pivotal role in music information research, having garnered attention for over two decades due to its relevance for music transcription and analysis. Despite notable advancements, challenges persist in the task, particularly concerning unique characteristics of harmonic content, which have resulted in existing systems’ performances reaching a glass ceiling. These challenges include annotator subjectivity, where varying interpretations among annotators lead to inconsistencies, and class imbalance within chord datasets, where certain chord classes are over-represented compared to others, posing difficulties in model training and evaluation. As a first contribution, this paper presents an evaluation of inter-annotator agreement in chord annotations, using metrics that extend beyond traditional binary measures. In addition, we propose a consonance-informed distance metric that reflects the perceptual similarity between harmonic annotations. Our analysis suggests that consonance-based distance metrics more effectively capture musically meaningful agreement between annotations. Expanding on these findings, we introduce a novel ACE conformer-based model that integrates consonance concepts into the model through consonance-based label smoothing. The proposed model also addresses class imbalance by separately estimating root, bass, and all note activations, enabling the reconstruction of chord labels from decomposed outputs.

[713] TinyMusician: On-Device Music Generation with Knowledge Distillation and Mixed Precision Quantization

Hainan Wang, Mehdi Hosseinzadeh, Reza Rawassizadeh

Main category: cs.SD

TL;DR: TinyMusician is a lightweight music generation model distilled from MusicGen that achieves 93% of MusicGen-Small’s performance with 55% smaller model size, making it deployable on mobile devices without cloud dependency.

DetailsMotivation: Transformer-based music generation models require massive computational resources and have slow inference times due to large parameter counts, making them impractical for deployment on edge devices like smartphones and wearables with limited resources.

Method: The model integrates two key innovations: (1) Stage-mixed Bidirectional and Skewed KL-Divergence for knowledge distillation, and (2) Adaptive Mixed-Precision Quantization to reduce model size while maintaining performance.

Result: TinyMusician retains 93% of MusicGen-Small’s performance while reducing model size by 55%, making it the first mobile-deployable music generation model that maintains high audio fidelity without cloud dependency.

Conclusion: TinyMusician successfully addresses the computational challenges of transformer-based music generation models, enabling efficient deployment on resource-constrained edge devices while preserving audio quality and eliminating the need for cloud services.

Abstract: The success of the generative model has gained unprecedented attention in the music generation area. Transformer-based architectures have set new benchmarks for model performance. However, their practical adoption is hindered by some critical challenges: the demand for massive computational resources and inference time, due to their large number of parameters. These obstacles make them infeasible to deploy on edge devices, such as smartphones and wearables, with limited computational resources. In this work, we present TinyMusician, a lightweight music generation model distilled from MusicGen (a State-of-the-art music generation model). TinyMusician integrates two innovations: (i) Stage-mixed Bidirectional and Skewed KL-Divergence and (ii) Adaptive Mixed-Precision Quantization. The experimental results demonstrate that TinyMusician retains 93% of the MusicGen-Small performance with 55% less model size. TinyMusician is the first mobile-deployable music generation model that eliminates cloud dependency while maintaining high audio fidelity and efficient resource usage

[714] A Unified Denoising and Adaptation Framework for Self-Supervised Bengali Dialectal ASR

Swadhin Biswas, Imran, Tuhin Sheikh

Main category: cs.SD

TL;DR: A novel unified framework for Bengali ASR that addresses dialect diversity and environmental noise using WavLM with masked speech denoising and multi-stage fine-tuning, achieving state-of-the-art performance.

DetailsMotivation: Bengali ASR faces challenges due to vast dialectal diversity and environmental noise, hindering technological accessibility for 270+ million speakers. Existing SSL models lack explicit noise handling and dialect adaptation mechanisms.

Method: Uses WavLM model pre-trained with masked speech denoising for inherent noise robustness. Implements multi-stage fine-tuning: first adapts to standard Bengali for linguistic foundation, then specializes for noise-robust dialect recognition with targeted data augmentation.

Result: Significantly outperforms strong baselines including fine-tuned wav2vec 2.0 and Whisper model. Achieves state-of-the-art performance on comprehensive Bengali dialect benchmark under various noisy conditions from clean audio to low SNR levels.

Conclusion: Establishes new state-of-the-art for Bengali ASR and provides scalable blueprint for developing practical ASR systems for other low-resource, high-variation languages globally.

Abstract: Automatic Speech Recognition (ASR) for Bengali, the world’s fifth most spoken language, remains a significant challenge, critically hindering technological accessibility for its over 270 million speakers. This challenge is compounded by two persistent and intertwined factors: the language’s vast dialectal diversity and the prevalence of acoustic noise in real-world environments. While state-of-the-art self-supervised learning (SSL) models have advanced ASR for low-resource languages, they often lack explicit mechanisms to handle environmental noise during pre-training or specialized adaptation strategies for the complex phonetic and lexical variations across Bengali dialects. This paper introduces a novel, unified framework designed to address these dual challenges simultaneously. Our approach is founded on the WavLM model, which is uniquely pre-trained with a masked speech denoising objective, making it inherently robust to acoustic distortions. We propose a specialized multi-stage fine-tuning strategy that first adapts the model to general-domain standard Bengali to establish a strong linguistic foundation and subsequently specializes it for noise-robust dialectal recognition through targeted data augmentation. The framework is rigorously evaluated on a comprehensive benchmark comprising multiple Bengali dialects under a wide range of simulated noisy conditions, from clean audio to low Signal-to-Noise Ratio (SNR) levels. Experimental results demonstrate that the proposed framework significantly outperforms strong baselines, including standard fine-tuned wav2vec 2.0 and the large-scale multilingual Whisper model. This work establishes a new state-of-the-art for this task and provides a scalable, effective blueprint for developing practical ASR systems for other low-resource, high-variation languages globally.

[715] EZhouNet:A framework based on graph neural network and anchor interval for the respiratory sound event detection

Yun Chu, Qiuhao Wang, Enze Zhou, Qian Liu, Gang Zheng

Main category: cs.SD

TL;DR: Proposes a graph neural network framework with anchor intervals for respiratory sound event detection that handles variable-length audio and provides precise temporal localization, outperforming existing methods.

DetailsMotivation: Existing respiratory sound event detection methods are limited by frame-level predictions, fixed-length audio constraints, and lack of exploration on location information impact. The subjective nature of auscultation and variability between experts calls for more robust automated detection systems.

Method: Graph neural network-based framework with anchor intervals that can process variable-length audio. Incorporates respiratory position information to enhance discrimination between abnormal sounds.

Result: Experiments on SPRSound 2024 and HF Lung V1 datasets demonstrate effectiveness. The approach shows improved flexibility and applicability for respiratory sound detection, with position information enhancing abnormal sound discrimination.

Conclusion: The proposed method addresses limitations of existing respiratory sound event detection approaches by providing better temporal localization, handling variable audio lengths, and leveraging location information for improved performance.

Abstract: Auscultation is a key method for early diagnosis of respiratory and pulmonary diseases, relying on skilled healthcare professionals. However, the process is often subjective, with variability between experts. As a result, numerous deep learning-based automatic classification methods have emerged, most of which focus on respiratory sound classification. In contrast, research on respiratory sound event detection remains limited. Existing sound event detection methods typically rely on frame-level predictions followed by post-processing to generate event-level outputs, making interval boundaries challenging to learn directly. Furthermore, many approaches can only handle fixed-length audio, lim- iting their applicability to variable-length respiratory sounds. Additionally, the impact of respiratory sound location information on detection performance has not been extensively explored. To address these issues, we propose a graph neural network-based framework with anchor intervals, capable of handling variable-length audio and providing more precise temporal localization for abnormal respi- ratory sound events. Our method improves both the flexibility and applicability of respiratory sound detection. Experiments on the SPRSound 2024 and HF Lung V1 datasets demonstrate the effec- tiveness of the proposed approach, and incorporating respiratory position information enhances the discrimination between abnormal sounds.

[716] The AudioMOS Challenge 2025

Wen-Chin Huang, Hui Wang, Cheng Liu, Yi-Chiao Wu, Andros Tjandra, Wei-Ning Hsu, Erica Cooper, Yong Qin, Tomoki Toda

Main category: cs.SD

TL;DR: AudioMOS Challenge 2025 summary - first challenge for automatic subjective quality prediction of synthetic audio with 3 tracks covering text-to-music, Meta Audiobox Aesthetics, and speech quality assessment.

DetailsMotivation: To establish the first standardized challenge for automatic evaluation of synthetic audio quality, addressing the growing need for objective assessment methods as audio generation systems advance.

Method: Organized three challenge tracks: 1) text-to-music quality and alignment assessment, 2) Meta Audiobox Aesthetics evaluation across multiple audio types, 3) synthetic speech quality at different sampling rates. Involved 24 teams from academia and industry.

Result: The challenge successfully attracted participation and confirmed improvements over baseline methods, demonstrating progress in automatic audio quality assessment.

Conclusion: The AudioMOS Challenge 2025 establishes a foundation for standardized evaluation and is expected to accelerate development in automatic quality assessment for audio generation systems.

Abstract: This is the summary paper for the AudioMOS Challenge 2025, the very first challenge for automatic subjective quality prediction for synthetic audio. The challenge consists of three tracks. The first track aims to assess text-to-music samples in terms of overall quality and textual alignment. The second track is based on the four evaluation dimensions of Meta Audiobox Aesthetics, and the test set consists of text-to-speech, text-to-audio, and text-to-music samples. The third track focuses on synthetic speech quality assessment in different sampling rates. The challenge attracted 24 unique teams from both academia and industry, and improvements over the baselines were confirmed. The outcome of this challenge is expected to facilitate development and progress in the field of automatic evaluation for audio generation systems.

[717] CabinSep: IR-Augmented Mask-Based MVDR for Real-Time In-Car Speech Separation with Distributed Heterogeneous Arrays

Runduo Han, Yanxin Hu, Yihui Fu, Zihan Zhang, Yukai Jv, Li Chen, Lei Xie

Main category: cs.SD

TL;DR: CabinSep is a lightweight neural MVDR speech separation approach that reduces ASR errors by 17.5% using spatial features, MVDR processing, and data augmentation with real and simulated impulse responses.

DetailsMotivation: To improve speech recognition accuracy in human-vehicle interaction by effectively separating overlapping speech from multiple speakers in cabin environments.

Method: Uses channel information for spatial features, MVDR processing during inference to reduce speech distortion, and combines simulated and real-recorded impulse responses for data augmentation to improve speaker localization.

Result: Achieves 17.5% relative reduction in speech recognition error rate compared to state-of-the-art DualSep model, with low computational complexity of only 0.4 GMACs.

Conclusion: CabinSep provides an effective and lightweight solution for speech separation in vehicle cabins, significantly improving ASR performance through spatial feature utilization, MVDR processing, and enhanced data augmentation techniques.

Abstract: Separating overlapping speech from multiple speakers is crucial for effective human-vehicle interaction. This paper proposes CabinSep, a lightweight neural mask-based minimum variance distortionless response (MVDR) speech separation approach, to reduce speech recognition errors in back-end automatic speech recognition (ASR) models. Our contributions are threefold: First, we utilize channel information to extract spatial features, which improves the estimation of speech and noise masks. Second, we employ MVDR during inference, reducing speech distortion to make it more ASR-friendly. Third, we introduce a data augmentation method combining simulated and real-recorded impulse responses (IRs), improving speaker localization at zone boundaries and further reducing speech recognition errors. With a computational complexity of only 0.4 GMACs, CabinSep achieves a 17.5% relative reduction in speech recognition error rate in a real-recorded dataset compared to the state-of-the-art DualSep model. Demos are available at: https://cabinsep.github.io/cabinsep/.

[718] ArabEmoNet: A Lightweight Hybrid 2D CNN-BiLSTM Model with Attention for Robust Arabic Speech Emotion Recognition

Ali Abouzeid, Bilal Elbouardi, Mohamed Maged, Shady Shehata

Main category: cs.SD

TL;DR: ArabEmoNet is a lightweight 2D CNN architecture for Arabic speech emotion recognition that achieves state-of-the-art performance with only 1M parameters, making it 74-90x smaller than existing models while preserving nuanced emotional cues from Mel spectrograms.

DetailsMotivation: Arabic is a low-resource language for speech emotion recognition with limited data and research. Existing methods using discrete MFCC features and 1D convolutions miss important spectro-temporal patterns critical for emotion detection.

Method: Uses Mel spectrograms processed through 2D convolutional neural networks instead of traditional MFCC features with 1D convolutions, preserving nuanced emotional information in spectro-temporal patterns.

Result: Achieves superior performance with only 1 million parameters (90x smaller than HuBERT base and 74x smaller than Whisper), making it ideal for resource-constrained environments while delivering state-of-the-art results.

Conclusion: ArabEmoNet advances Arabic speech emotion recognition by offering exceptional performance and accessibility for real-world applications through its lightweight yet effective architecture.

Abstract: Speech emotion recognition is vital for human-computer interaction, particularly for low-resource languages like Arabic, which face challenges due to limited data and research. We introduce ArabEmoNet, a lightweight architecture designed to overcome these limitations and deliver state-of-the-art performance. Unlike previous systems relying on discrete MFCC features and 1D convolutions, which miss nuanced spectro-temporal patterns, ArabEmoNet uses Mel spectrograms processed through 2D convolutions, preserving critical emotional cues often lost in traditional methods. While recent models favor large-scale architectures with millions of parameters, ArabEmoNet achieves superior results with just 1 million parameters, 90 times smaller than HuBERT base and 74 times smaller than Whisper. This efficiency makes it ideal for resource-constrained environments. ArabEmoNet advances Arabic speech emotion recognition, offering exceptional performance and accessibility for real-world applications.

[719] Music Genre Classification Using Machine Learning Techniques

Alokit Mishra, Ryyan Akhtar

Main category: cs.SD

TL;DR: SVM with hand-crafted features outperforms CNN on Mel spectrograms for music genre classification on GTZAN dataset, showing traditional feature engineering remains effective for moderately sized datasets.

DetailsMotivation: To compare classical machine learning methods with deep learning approaches for automatic music genre classification and understand their performance differences on constrained datasets.

Method: Comparative analysis using SVM and ensemble methods with hand-crafted audio features versus CNN operating on Mel spectrograms, evaluated on the GTZAN dataset.

Result: SVM with domain-specific feature engineering achieved superior classification accuracy compared to the end-to-end CNN model.

Conclusion: Traditional feature extraction maintains relevance in audio processing, and deep learning may not be universally superior, especially for moderately sized datasets where engineered features provide regularization against overfitting.

Abstract: This paper presents a comparative analysis of machine learning methodologies for automatic music genre classification. We evaluate the performance of classical classifiers, including Support Vector Machines (SVM) and ensemble methods, trained on a comprehensive set of hand-crafted audio features, against a Convolutional Neural Network (CNN) operating on Mel spectrograms. The study is conducted on the widely-used GTZAN dataset. Our findings demonstrate a noteworthy result: the SVM, leveraging domain-specific feature engineering, achieves superior classification accuracy compared to the end-to-end CNN model. We attribute this outcome to the data-constrained nature of the benchmark dataset, where the strong inductive bias of engineered features provides a regularization effect that mitigates the risk of overfitting inherent in high-capacity deep learning models. This work underscores the enduring relevance of traditional feature extraction in practical audio processing tasks and provides a critical perspective on the universal applicability of deep learning, especially for moderately sized datasets.

[720] FireRedTTS-2: Towards Long Conversational Speech Generation for Podcast and Chatbot

Kun Xie, Feiyu Shen, Junjie Li, Fenglong Xie, Xu Tang, Yao Hu

Main category: cs.SD

TL;DR: FireRedTTS-2 is a streaming TTS system for multi-speaker dialogues that enables real-time interactive chat with stable synthesis, accurate speaker switching, and context-aware prosody.

DetailsMotivation: Current dialogue generation approaches require complete dialogue text before synthesis, produce inseparable speech with all voices, and suffer from unstable synthesis, inaccurate speaker transitions, and incoherent prosody, making them unsuitable for interactive chat.

Method: Uses a new 12.5Hz streaming speech tokenizer for faster training/inference and richer semantics. Adopts text-speech interleaved format with speaker-labeled text and aligned speech tokens. Employs a dual-transformer architecture: large decoder-only transformer predicts tokens at first layer, smaller one completes subsequent layers.

Result: Seamlessly integrates with chat frameworks, produces emotionally expressive speech with minimal fine-tuning. Surpasses MoonCast, Zipvoice-Dialogue, and MOSS-TTSD in podcast generation for intelligibility, speaker-turn reliability, and perceived naturalness with context-consistent prosody.

Conclusion: FireRedTTS-2 enables real-time streaming multi-speaker dialogue generation with improved stability, naturalness, and speaker switching capabilities, making it suitable for interactive applications.

Abstract: Current dialogue generation approaches typically require the complete dialogue text before synthesis and produce a single, inseparable speech containing all voices, making them unsuitable for interactive chat; moreover, they suffer from unstable synthesis, inaccurate speaker transitions, and incoherent prosody. In this work, we present FireRedTTS-2, a long-form streaming TTS system for multi-speaker dialogue generation, delivering stable, natural speech with reliable speaker switching and context-aware prosody. A new 12.5Hz streaming speech tokenizer accelerates training and inference, extends maximum dialogue length, encodes richer semantics to stabilize text-to-token modeling and supports high-fidelity streaming generation for real-time applications. We adopt a text-speech interleaved format, concatenating speaker-labeled text with aligned speech tokens in chronological order, and model it with a dual-transformer: a large decoder-only transformer predicts tokens at the first layer, and a smaller one completes subsequent layers. Experimental results show that FireRedTTS-2 integrates seamlessly with chat frameworks and, with minimal fine-tuning, produces emotionally expressive speech guided by implicit contextual cues. In podcast generation, it surpasses existing systems including MoonCast, Zipvoice-Dialogue, and MOSS-TTSD in objective intelligibility, speaker-turn reliability, and perceived naturalness with context-consistent prosody. Our demos are available at https://fireredteam.github.io/demos/firered_tts_2.

[721] AudioRWKV: Efficient and Stable Bidirectional RWKV for Audio Pattern Recognition

Jiayu Xiong, Jun Xue, Jianlong Kwan, Jing Wang

Main category: cs.SD

TL;DR: AudioRWKV (A-RWKV) is a new efficient and stable architecture for audio modeling that combines RWKV7’s recurrent formulation with 2D convolutions for spectro-temporal patterns and bidirectional context modeling with linear complexity.

DetailsMotivation: Transformers have O(L^2) complexity that hinders long-sequence processing, while Mamba architectures become unstable when scaling. Need for efficient and stable audio modeling architecture.

Method: Inherits RWKV7’s stable recurrent formulation, replaces 1D token-shift with 2D depthwise separable convolution, adapts causal WKV kernel to bidirectional WKV (Bi-WKV) for global context modeling with linear complexity.

Result: A-RWKV-S (22M) achieves performance parity with AuM-B (92M), exhibits more stable throughput than AST, and achieves up to 13.3X speedup for long-form audio (~5 minutes 28 seconds).

Conclusion: A-RWKV provides an efficient and stable alternative to Transformers and Mamba models for audio processing, enabling seamless scaling to larger models with linear computational complexity.

Abstract: Recently, Transformers (e.g., Audio Spectrogram Transformers, AST) and state-space models (e.g., Audio Mamba, AuM) have achieved remarkable progress in audio modeling. However, the O(L^2) computational complexity of the Transformer architecture hinders efficient long-sequence processing, while the Mamba architecture tends to become unstable when scaling parameters and data. To address these challenges, this paper proposes AudioRWKV (A-RWKV), a highly efficient and stable architecture for audio modeling. Specifically, we inherit the stable and efficient recurrent formulation of RWKV7 and replace its 1D token-shift operation with a 2D depthwise separable convolution to better capture local spectro-temporal patterns. Furthermore, we adapt the original causal WKV kernel into a bidirectional WKV kernel (Bi-WKV), enabling global context modeling over the entire audio sequence while maintaining linear computational complexity. Benefiting from the inherent stability of the RWKV7 foundation, A-RWKV scales seamlessly to larger model sizes. Experimental results demonstrate that, under the same linear-model regime, A-RWKV-S (22M) achieves performance parity with AuM-B (92M) while exhibiting more stable throughput than AST; for long-form audio (~5 minutes 28 seconds), WKV7 achieves up to a 13.3X speedup in processing.

[722] Spectrogram Patch Codec: A 2D Block-Quantized VQ-VAE and HiFi-GAN for Neural Speech Coding

Luis Felipe Chary, Miguel Arjona Ramirez

Main category: cs.SD

TL;DR: A neural speech codec using single-stage quantization on mel-spectrogram patches instead of complex RVQ stacks, achieving competitive quality at 7.5 kbits/s with low latency.

DetailsMotivation: To challenge the need for complex residual vector quantization (RVQ) stacks in neural speech codecs by introducing a simpler, single-stage quantization approach that enables low-latency streaming.

Method: Operates directly on mel-spectrogram treated as 2D data, quantizing non-overlapping 4x4 patches into a shared codebook. Uses late-stage adversarial fine-tuning for VQ-VAE and trains HiFi-GAN vocoder from scratch on reconstructed spectrograms.

Result: Achieves competitive perceptual quality and intelligibility at approximately 7.5 kbits/s for 16 kHz speech, as measured by STOI, PESQ, MCD, and ViSQOL metrics against state-of-the-art neural codecs.

Conclusion: The simplified, non-residual architecture provides an effective and open foundation for future low-latency codec designs, demonstrating that complex RVQ stacks are not necessary for high-quality neural speech coding.

Abstract: We present a neural speech codec that challenges the need for complex residual vector quantization (RVQ) stacks by introducing a simpler, single-stage quantization approach. Our method operates directly on the mel-spectrogram, treating it as a 2D data and quantizing non-overlapping 4x4 patches into a single, shared codebook. This patchwise design simplifies the architecture, enables low-latency streaming, and yields a discrete latent grid. To ensure high-fidelity synthesis, we employ a late-stage adversarial fine-tuning for the VQ-VAE and train a HiFi-GAN vocoder from scratch on the codec’s reconstructed spectrograms. Operating at approximately 7.5 kbits/s for 16 kHz speech, our system was evaluated against several state-of-the-art neural codecs using objective metrics such as STOI, PESQ, MCD, and ViSQOL. The results demonstrate that our simplified, non-residual architecture achieves competitive perceptual quality and intelligibility, validating it as an effective and open foundation for future low-latency codec designs.

[723] Speech transformer models for extracting information from baby cries

Guillem Bonafos, Jéremy Rouch, Lény Lego, David Reby, Hugues Patural, Nicolas Mathevon, Rémy Emonet

Main category: cs.SD

TL;DR: Pre-trained speech models’ latent representations effectively classify baby cries and encode vocal instability and identity information, showing promise for non-speech audio tasks.

DetailsMotivation: To explore the applicability of pre-trained speech models to non-speech data (baby cries) and understand what acoustic properties are encoded in their latent representations.

Method: Evaluated five pre-trained speech models on eight baby cries datasets (115 hours, 960 babies), assessing latent representations across all classification tasks for each dataset.

Result: Latent representations effectively classified baby cries and encoded key information about vocal source instability and baby identity.

Conclusion: The study provides valuable insights for designing future models for similar tasks like emotion detection, demonstrating the transferability of speech models to non-speech audio domains.

Abstract: Transfer learning using latent representations from pre-trained speech models achieves outstanding performance in tasks where labeled data is scarce. However, their applicability to non-speech data and the specific acoustic properties encoded in these representations remain largely unexplored. In this study, we investigate both aspects. We evaluate five pre-trained speech models on eight baby cries datasets, encompassing 115 hours of audio from 960 babies. For each dataset, we assess the latent representations of each model across all available classification tasks. Our results demonstrate that the latent representations of these models can effectively classify human baby cries and encode key information related to vocal source instability and identity of the crying baby. In addition, a comparison of the architectures and training strategies of these models offers valuable insights for the design of future models tailored to similar tasks, such as emotion detection.

[724] TTA-Bench: A Comprehensive Benchmark for Evaluating Text-to-Audio Models

Hui Wang, Cheng Liu, Junyang Chen, Haoze Liu, Yuhang Jia, Shiwan Zhao, Jiaming Zhou, Haoqin Sun, Hui Bu, Yong Qin

Main category: cs.SD

TL;DR: TTA-Bench is a comprehensive benchmark for Text-to-Audio generation models that evaluates across 7 dimensions including accuracy, robustness, fairness, and toxicity using 2,999 diverse prompts and over 118,000 human annotations.

DetailsMotivation: Current TTA evaluation methods focus narrowly on perceptual quality while overlooking robustness, generalization, and ethical concerns, creating a need for more holistic evaluation standards.

Method: Developed a benchmark with 2,999 diverse prompts generated through automated and manual methods, using a unified evaluation protocol combining objective metrics with human annotations from experts and general users.

Result: Benchmarked 10 state-of-the-art TTA models, providing detailed insights into their strengths and limitations across functional performance, reliability, and social responsibility dimensions.

Conclusion: TTA-Bench establishes a new standard for holistic and responsible evaluation of TTA systems, with open-sourced dataset and evaluation tools to advance the field.

Abstract: Text-to-Audio (TTA) generation has made rapid progress, but current evaluation methods remain narrow, focusing mainly on perceptual quality while overlooking robustness, generalization, and ethical concerns. We present TTA-Bench, a comprehensive benchmark for evaluating TTA models across functional performance, reliability, and social responsibility. It covers seven dimensions including accuracy, robustness, fairness, and toxicity, and includes 2,999 diverse prompts generated through automated and manual methods. We introduce a unified evaluation protocol that combines objective metrics with over 118,000 human annotations from both experts and general users. Ten state-of-the-art models are benchmarked under this framework, offering detailed insights into their strengths and limitations. TTA-Bench establishes a new standard for holistic and responsible evaluation of TTA systems. The dataset and evaluation tools are open-sourced at https://nku-hlt.github.io/tta-bench/.

[725] AudioCodecBench: A Comprehensive Benchmark for Audio Codec Evaluation

Lu Wang, Hao Chen, Siyu Wu, Zhiyue Wu, Hao Zhou, Chengfeng Zhang, Ting Wang, Haodi Zhang

Main category: cs.SD

TL;DR: This paper addresses challenges in audio tokenization for multimodal LLMs by providing suitable definitions for semantic and acoustic tokens and introducing a comprehensive evaluation framework across multiple dimensions.

DetailsMotivation: Existing research lacks suitable definitions for semantic and acoustic tokens in audio tokenization, and current evaluations are limited to specific domains/tasks, preventing fair comparisons of different codecs.

Method: The paper provides appropriate definitions for semantic and acoustic tokens and introduces a systematic evaluation framework that assesses codecs across four dimensions: audio reconstruction metrics, codebook index stability, decoder-only transformer perplexity, and downstream probe task performance.

Result: The results demonstrate the correctness of the provided definitions and reveal correlations among reconstruction metrics, codebook ID stability, downstream task performance, and perplexity.

Conclusion: The proposed framework enables comprehensive and fair evaluation of audio codecs for MLLMs, addressing previous limitations in token definitions and evaluation methodologies.

Abstract: Multimodal Large Language Models (MLLMs) have been widely applied in speech and music. This tendency has led to a focus on audio tokenization for Large Models (LMs). Unlike semantic-only text tokens, audio tokens must both capture global semantic content and preserve fine-grained acoustic details. Moreover, they provide a discrete method for speech and music that can be effectively integrated into MLLMs. However, existing research is unsuitable in the definitions of semantic tokens and acoustic tokens. In addition, the evaluation of different codecs typically concentrates on specific domains or tasks, such as reconstruction or Automatic Speech Recognition (ASR) task, which prevents fair and comprehensive comparisons. To address these problems, this paper provides suitable definitions for semantic and acoustic tokens and introduces a systematic evaluation framework. This framework allows for a comprehensive assessment of codecs’ capabilities which evaluate across four dimensions: audio reconstruction metric, codebook index (ID) stability, decoder-only transformer perplexity, and performance on downstream probe tasks. Our results show the correctness of the provided suitable definitions and the correlation among reconstruction metrics, codebook ID stability, downstream probe tasks and perplexity.

[726] ESTM: An Enhanced Dual-Branch Spectral-Temporal Mamba for Anomalous Sound Detection

Chengyuan Ma, Peng Jia, Hongyue Guo, Wenming Yang

Main category: cs.SD

TL;DR: Proposes ESTM framework using dual-path Mamba architecture with time-frequency decoupled modeling for industrial equipment anomalous sound detection, achieving improved performance on DCASE 2020 dataset.

DetailsMotivation: Existing methods struggle to capture long-range temporal patterns and cross-band dynamic coupling effects in machine acoustic features due to limited local receptive fields.

Method: Uses dual-path Mamba architecture with Selective State-Space Models for long-range sequence modeling, fuses enhanced Mel spectrograms and raw audio features, and incorporates TriStat-Gating module for improved anomaly sensitivity.

Result: Demonstrates improved anomalous detection performance on the DCASE 2020 Task 2 dataset.

Conclusion: The proposed ESTM framework effectively addresses time-frequency coupling challenges in industrial sound anomaly detection, validating the effectiveness of the dual-path Mamba approach with time-frequency decoupled modeling.

Abstract: The core challenge in industrial equipment anoma lous sound detection (ASD) lies in modeling the time-frequency coupling characteristics of acoustic features. Existing modeling methods are limited by local receptive fields, making it difficult to capture long-range temporal patterns and cross-band dynamic coupling effects in machine acoustic features. In this paper, we propose a novel framework, ESTM, which is based on a dual-path Mamba architecture with time-frequency decoupled modeling and utilizes Selective State-Space Models (SSM) for long-range sequence modeling. ESTM extracts rich feature representations from different time segments and frequency bands by fusing enhanced Mel spectrograms and raw audio features, while further improving sensitivity to anomalous patterns through the TriStat-Gating (TSG) module. Our experiments demonstrate that ESTM improves anomalous detection performance on the DCASE 2020 Task 2 dataset, further validating the effectiveness of the proposed method.

[727] FLM-Audio: Natural Monologues Improves Native Full-Duplex Chatbots via Dual Training

Yiqun Yao, Xiang Li, Xin Jiang, Xuezhi Fang, Naitong Yu, Wenjia Ma, Aixin Sun, Yequan Wang

Main category: cs.SD

TL;DR: Proposes natural monologues with dual training paradigm to align text and audio at different bitrates for full-duplex dialog models, overcoming word-level alignment limitations.

DetailsMotivation: Existing full-duplex models face challenges aligning textual monologues with audio streams at different bitrates. Word-level alignment degrades language ability of pre-trained models and requires accurate timestamps, causing cascading errors and high pre-processing costs.

Method: Uses textual monologues in continuous tokens sequence (“natural” monologues) that mimic human cognitive behavior. Implements dual training paradigm alternating monologue position (leading or trailing audio) across different training stages.

Result: Developed FLM-Audio, a 7B spoken dialog model that demonstrates superior responsiveness, duplexity, and chatting experiences as confirmed by experimental results.

Conclusion: The proposed natural monologues with dual training paradigm effectively solves the alignment challenge between text and audio streams, enabling high-performance full-duplex dialog models without the limitations of word-level alignment.

Abstract: Full-duplex dialog models are designed to listen and speak simultaneously with rapid responses to fast-changing user input. Among existing approaches, native full-duplex models merges different channels (e.g. listen and speak) in a single time step, overcoming the high response latency inherent to time-division multiplexing time-division multiplexing (TDM) alternatives. Yet, a key challenge remains: aligning textual monologues with audio streams that operate at different bitrates. The prevailing solution relies on word-level alignment, but this can degrade the language ability of large pre-trained models. Moreover, it requires highly accurate timestamps for every token, which introduces cascading errors and increases pre-processing costs. In this paper, we propose textual monologues in continuous tokens sequence, namely “natural” monologues, which mimics humanoid cognitive behavior in dialogs. For temporal alignment, we alternate the position of the natural monologue - leading or trailing the audio - across different training stages. This “dual” training paradigm proves highly effective in building FLM-Audio, our 7B spoken dialog model that demonstrates superior responsiveness, duplexity, and chatting experiences, as confirmed by experimental results.

[728] Multi-stream Convolutional Neural Network with Frequency Selection for Robust Speaker Verification

Wei Yao, Shen Chen, Jiamin Cui, Yaolin Lou

Main category: cs.SD

TL;DR: Proposes multi-stream CNN framework with frequency selection technique for speaker verification, achieving 20.53% relative improvement in minDCF over single-stream baseline.

DetailsMotivation: To enhance speaker verification by enabling machines to learn from partial frequency ranges rather than full frequency spectrum, improving acoustic modeling robustness.

Method: Multi-stream CNN framework that segments full frequency band into sub-bands, with each stream processing different frequency ranges in parallel, followed by mean normalization and pooling for fused embeddings.

Result: Significant performance improvement with 20.53% relative improvement in minimum Decision Cost Function (minDCF) on VoxCeleb dataset compared to single-stream baseline.

Conclusion: Frequency selection technique in multi-stream CNN framework effectively enhances speaker verification performance by leveraging diverse temporal embeddings from different frequency sub-bands.

Abstract: Speaker verification aims to verify whether an input speech corresponds to the claimed speaker, and conventionally, this kind of system is deployed based on single-stream scenario, wherein the feature extractor operates in full frequency range. In this paper, we hypothesize that machine can learn enough knowledge to do classification task when listening to partial frequency range instead of full frequency range, which is so called frequency selection technique, and further propose a novel framework of multi-stream Convolutional Neural Network (CNN) with this technique for speaker verification tasks. The proposed framework accommodates diverse temporal embeddings generated from multiple streams to enhance the robustness of acoustic modeling. For the diversity of temporal embeddings, we consider feature augmentation with frequency selection, which is to manually segment the full-band of frequency into several sub-bands, and the feature extractor of each stream can select which sub-bands to use as target frequency domain. Different from conventional single-stream solution wherein each utterance would only be processed for one time, in this framework, there are multiple streams processing it in parallel. The input utterance for each stream is pre-processed by a frequency selector within specified frequency range, and post-processed by mean normalization. The normalized temporal embeddings of each stream will flow into a pooling layer to generate fused embeddings. We conduct extensive experiments on VoxCeleb dataset, and the experimental results demonstrate that multi-stream CNN significantly outperforms single-stream baseline with 20.53 % of relative improvement in minimum Decision Cost Function (minDCF).

[729] A Neural Speech Codec for Noise Robust Speech Coding

Jiayi Huang, Zeyu Yan, Wenbin Jiang, He Wang, Fei Wen

Main category: cs.SD

TL;DR: The paper proposes a two-stage training framework for joint compression and enhancement of noisy speech signals, with theoretical foundation showing it achieves optimal distortion-perception tradeoff.

DetailsMotivation: To address the joint compression and enhancement problem for noisy speech signals, providing a theoretically grounded alternative to heuristic training methods like SoundStream.

Method: A two-stage optimization procedure: first optimize encoder-decoder pair using distortion loss only, then fix encoder and optimize perceptual decoder using perception loss. This is implemented as a training framework.

Result: The proposed codec outperforms SoundStream and other representative codecs in both objective and subjective evaluation metrics across various noise and bit-rate conditions.

Conclusion: The two-stage training method provides a theoretically founded approach for joint speech compression and enhancement that achieves superior performance compared to existing heuristic methods.

Abstract: This paper considers the joint compression and enhancement problem for speech signal in the presence of noise. Recently, the SoundStream codec, which relies on end-to-end joint training of an encoder-decoder pair and a residual vector quantizer by a combination of adversarial and reconstruction losses,has shown very promising performance, especially in subjective perception quality. In this work, we provide a theoretical result to show that, to simultaneously achieve low distortion and high perception in the presence of noise, there exist an optimal two-stage optimization procedure for the joint compression and enhancement problem. This procedure firstly optimizes an encoder-decoder pair using only distortion loss and then fixes the encoder to optimize a perceptual decoder using perception loss. Based on this result, we construct a two-stage training framework for joint compression and enhancement of noisy speech signal. Unlike existing training methods which are heuristic, the proposed two-stage training method has a theoretical foundation. Finally, experimental results for various noise and bit-rate conditions are provided. The results demonstrate that a codec trained by the proposed framework can outperform SoundStream and other representative codecs in terms of both objective and subjective evaluation metrics. Code is available at \textit{https://github.com/jscscloris/SEStream}.

[730] I2TTS: Image-indicated Immersive Text-to-speech Synthesis with Spatial Perception

Jiawei Zhang, Tian-Hao Zhang, Jun Wang, Jiaran Gao, Xinyuan Qian, Xu-Cheng Yin

Main category: cs.SD

TL;DR: A novel multi-modal TTS approach called I2TTS that integrates visual scene prompts to control spatial perception and immersive experience in speech synthesis.

DetailsMotivation: Existing TTS systems focus on natural-sounding speech but overlook spatial perception needed for immersive experiences in gaming and virtual reality applications.

Method: Introduces a scene prompt encoder that integrates visual scene prompts into the synthesis pipeline, plus a reverberation classification and refinement technique to adjust mel-spectrograms for accurate spatial matching.

Result: The model achieves high-quality scene and spatial matching without compromising speech naturalness, demonstrating significant advancement in context-aware speech synthesis.

Conclusion: I2TTS successfully bridges the gap between traditional TTS and spatial perception requirements, enabling immersive speech synthesis for gaming and VR applications.

Abstract: Controlling the style and characteristics of speech synthesis is crucial for adapting the output to specific contexts and user requirements. Previous Text-to-speech (TTS) works have focused primarily on the technical aspects of producing natural-sounding speech, such as intonation, rhythm, and clarity. However, they overlook the fact that there is a growing emphasis on spatial perception of synthesized speech, which may provide immersive experience in gaming and virtual reality. To solve this issue, in this paper, we present a novel multi-modal TTS approach, namely Image-indicated Immersive Text-to-speech Synthesis (I2TTS). Specifically, we introduce a scene prompt encoder that integrates visual scene prompts directly into the synthesis pipeline to control the speech generation process. Additionally, we propose a reverberation classification and refinement technique that adjusts the synthesized mel-spectrogram to enhance the immersive experience, ensuring that the involved reverberation condition matches the scene accurately. Experimental results demonstrate that our model achieves high-quality scene and spatial matching without compromising speech naturalness, marking a significant advancement in the field of context-aware speech synthesis. Project demo page: https://spatialTTS.github.io/ Index Terms-Speech synthesis, scene prompt, spatial perception

[731] Dynamic Fusion Multimodal Network for SpeechWellness Detection

Wenqiang Sun, Han Yin, Jisheng Bai, Jianfeng Chen

Main category: cs.SD

TL;DR: Lightweight multimodal system combining speech and text with dynamic fusion for suicide risk detection, achieving 78% parameter reduction and 5% accuracy improvement.

DetailsMotivation: Suicide is a leading cause of adolescent death, and previous approaches focused on single modalities (text or audio) in isolation. Multimodal integration provides more comprehensive mental state understanding.

Method: Multi-branch multimodal system with time-domain and time-frequency acoustic features plus semantic representations. Uses dynamic fusion with learnable weights to adaptively integrate modalities. Lightweight structure designed for efficiency.

Result: Superior performance compared to baseline with 78% reduction in model parameters and 5% improvement in accuracy.

Conclusion: The proposed lightweight multimodal system with dynamic fusion effectively integrates speech and text modalities for improved suicide risk detection while maintaining computational efficiency.

Abstract: Suicide is one of the leading causes of death among adolescents. Previous suicide risk prediction studies have primarily focused on either textual or acoustic information in isolation, the integration of multimodal signals, such as speech and text, offers a more comprehensive understanding of an individual’s mental state. Motivated by this, and in the context of the 1st SpeechWellness detection challenge, we explore a lightweight multi-branch multimodal system based on a dynamic fusion mechanism for speechwellness detection. To address the limitation of prior approaches that rely on time-domain waveforms for acoustic analysis, our system incorporates both time-domain and time-frequency (TF) domain acoustic features, as well as semantic representations. In addition, we introduce a dynamic fusion block to adaptively integrate information from different modalities. Specifically, it applies learnable weights to each modality during the fusion process, enabling the model to adjust the contribution of each modality. To enhance computational efficiency, we design a lightweight structure by simplifying the original baseline model. Experimental results demonstrate that the proposed system exhibits superior performance compared to the challenge baseline, achieving a 78% reduction in model parameters and a 5% improvement in accuracy.

cs.LG

[732] Diagnosing Psychiatric Patients: Can Large Language and Machine Learning Models Perform Effectively in Emergency Cases?

Abu Shad Ahammed, Sayeri Mukherjee, Roman Obermaisser

Main category: cs.LG

TL;DR: Research on using machine learning and LLMs to diagnose psychiatric disorders from behavioral patterns in emergency rescue data

DetailsMotivation: Mental disorders are often misdiagnosed due to lack of visible symptoms, making emergency identification challenging but crucial for patient care

Method: Collected data from emergency psychiatric patients at a German rescue station, used various ML models including Llama 3.1 to analyze behavioral patterns for diagnostic assessment

Result: Evaluated predictive capabilities of models for identifying patients with mental disorders in rescue scenarios

Conclusion: ML and LLMs show potential as efficient tools for psychiatric assessment in emergency situations

Abstract: Mental disorders are clinically significant patterns of behavior that are associated with stress and/or impairment in social, occupational, or family activities. People suffering from such disorders are often misjudged and poorly diagnosed due to a lack of visible symptoms compared to other health complications. During emergency situations, identifying psychiatric issues is that’s why challenging but highly required to save patients. In this paper, we have conducted research on how traditional machine learning and large language models (LLM) can assess these psychiatric patients based on their behavioral patterns to provide a diagnostic assessment. Data from emergency psychiatric patients were collected from a rescue station in Germany. Various machine learning models, including Llama 3.1, were used with rescue patient data to assess if the predictive capabilities of the models can serve as an efficient tool for identifying patients with unhealthy mental disorders, especially in rescue cases.

[733] Mitigating Data Exfiltration Attacks through Layer-Wise Learning Rate Decay Fine-Tuning

Elie Thellier, Huiyu Li, Nicholas Ayache, Hervé Delingette

Main category: cs.LG

TL;DR: A defense method that perturbs model parameters through fine-tuning with decaying layer-wise learning rate to prevent data exfiltration attacks while maintaining utility performance.

DetailsMotivation: Data lakes enable training on sensitive medical data but introduce serious privacy risks from attacks that can exfiltrate training data through model parameters or multi-task learning memorization.

Method: Proposes fine-tuning with decaying layer-wise learning rate to perturb model parameters at export time, corrupting embedded data without degrading task performance.

Result: Evaluations on medical datasets (DermaMNIST, ChestMNIST, MIMIC-CXR) show maintained utility performance, effective disruption of state-of-the-art exfiltration attacks, outperforms prior defenses, and renders exfiltrated data unusable.

Conclusion: Provides a practical defense against data leakage in data lake-trained models and centralized federated learning, with discussions on adaptive attacks highlighting future challenges.

Abstract: Data lakes enable the training of powerful machine learning models on sensitive, high-value medical datasets, but also introduce serious privacy risks due to potential leakage of protected health information. Recent studies show adversaries can exfiltrate training data by embedding latent representations into model parameters or inducing memorization via multi-task learning. These attacks disguise themselves as benign utility models while enabling reconstruction of high-fidelity medical images, posing severe privacy threats with legal and ethical implications. In this work, we propose a simple yet effective mitigation strategy that perturbs model parameters at export time through fine-tuning with a decaying layer-wise learning rate to corrupt embedded data without degrading task performance. Evaluations on DermaMNIST, ChestMNIST, and MIMIC-CXR show that our approach maintains utility task performance, effectively disrupts state-of-the-art exfiltration attacks, outperforms prior defenses, and renders exfiltrated data unusable for training. Ablations and discussions on adaptive attacks highlight challenges and future directions. Our findings offer a practical defense against data leakage in data lake-trained models and centralized federated learning.

[734] ZeroQAT: Your Quantization-aware Training but Efficient

Qitao Tan, Xiaoying Song, Jin Lu, Guoming Li, Jun Liu, Lingzi Hong, Caiwen Ding, Jundong Li, Xiaoming Zhai, Shaoyi Huang, Wei Niu, Geng Yuan

Main category: cs.LG

TL;DR: ZeroQAT is a zeroth-order optimization-based quantization-aware training framework that eliminates backpropagation overhead while achieving PTQ efficiency and QAT accuracy for LLM quantization.

DetailsMotivation: Existing low-bit post-training quantization methods suffer from accuracy degradation due to cumulative error propagation and misalignment issues, while quantization-aware training is too computationally expensive due to backpropagation requirements.

Method: ZeroQAT uses forward-only gradient estimation through zeroth-order optimization to eliminate backpropagation, jointly learning quantized weights, weight clipping thresholds, and equivalent transformations to handle quantization error and activation outliers.

Result: Experiments show ZeroQAT achieves the efficiency of post-training quantization while maintaining the accuracy of quantization-aware training.

Conclusion: ZeroQAT provides a practical solution for high-quality low-bit quantization of large language models by combining efficiency with accuracy through zeroth-order optimization.

Abstract: Quantization is an effective technique to reduce the deployment cost of large language models (LLMs), and post-training quantization (PTQ) has been widely studied due to its efficiency. However, existing low-bit PTQ methods suffer from accuracy degradation because their layer-wise optimization introduces cumulative error propagation and misalignment between local reconstruction objectives and downstream performance. While quantization-aware training (QAT) provides a principled solution, its reliance on backpropagation incurs prohibitive data, time, and memory costs, limiting its practicality. To address these challenges, we propose ZeroQAT, a zeroth-order optimization-based QAT framework. ZeroQAT leverages forward-only gradient estimation to eliminate the need for backpropagation, significantly reducing computational and memory overhead while retaining the benefits of end-to-end optimization. Moreover, ZeroQAT jointly learns quantized weights, weight clipping thresholds, and equivalent transformations to mitigate quantization error and handle activation outliers. Experiments demonstrate that ZeroQAT achieves the efficiency of PTQ while retaining the accuracy of QAT, offering a practical solution for high-quality low-bit quantization of LLMs.

[735] Industrial Steel Slag Flow Data Loading Method for Deep Learning Applications

Mert Sehri, Ana Cardoso, Francisco de Assis Boldt, Patrick Dumond

Main category: cs.LG

TL;DR: Hybrid CNN-LSTM model with RMS preprocessing achieves 99.10% accuracy for slag flow detection using vibration data in steel casting, outperforming traditional methods.

DetailsMotivation: Steel casting processes suffer financial losses from slag flow contamination, requiring accurate detection to prevent quality issues and operational inefficiencies.

Method: Hybrid deep learning model combining 1D CNN and LSTM layers, using raw vibration signals from accelerometers with RMS preprocessing and selective embedding data loading strategy across 16 domains.

Result: Achieved 99.10% +/- 0.30 test accuracy, demonstrating robust classification performance and superior generalization compared to standard 1D CNN models.

Conclusion: The method provides a practical, scalable solution for real-time slag flow monitoring, improving reliability and operational efficiency in steel manufacturing.

Abstract: Steel casting processes are vulnerable to financial losses due to slag flow contamination, making accurate slag flow condition detection essential. This study introduces a novel cross-domain diagnostic method using vibration data collected from an industrial steel foundry to identify various stages of slag flow. A hybrid deep learning model combining one-dimensional convolutional neural networks and long short-term memory layers is implemented, tested, and benchmarked against a standard one-dimensional convolutional neural network. The proposed method processes raw time-domain vibration signals from accelerometers and evaluates performance across 16 distinct domains using a realistic cross-domain dataset split. Results show that the hybrid convolutional neural network and long short-term memory architecture, when combined with root mean square preprocessing and a selective embedding data loading strategy, achieves robust classification accuracy, outperforming traditional models and loading techniques. The highest test accuracy of 99.10 +/- 0.30 demonstrates the method’s capability for generalization and industrial relevance. This work presents a practical and scalable solution for real-time slag flow monitoring, contributing to improved reliability and operational efficiency in steel manufacturing.

[736] Transfer Learning for Minimum Operating Voltage Prediction in Advanced Technology Nodes: Leveraging Legacy Data and Silicon Odometer Sensing

Yuxuan Yin, Rebecca Chen, Boxun Xu, Chen He, Peng Li

Main category: cs.LG

TL;DR: Transfer learning framework using 16nm legacy data and on-chip sensor features for accurate V_min prediction at 5nm node

DetailsMotivation: Accurate chip performance prediction is critical for energy efficiency and reliability, but challenging at advanced nodes due to limited training data and complex process variation relationships

Method: Novel transfer learning framework that leverages abundant 16nm legacy data combined with on-chip silicon odometer sensor data to characterize localized process variations

Result: Significantly improved prediction accuracy for minimum operating voltage (V_min) at the 5nm technology node

Conclusion: The proposed approach effectively addresses data scarcity and process variation challenges in advanced semiconductor manufacturing through transfer learning and sensor-based feature integration

Abstract: Accurate prediction of chip performance is critical for ensuring energy efficiency and reliability in semiconductor manufacturing. However, developing minimum operating voltage ($V_{min}$) prediction models at advanced technology nodes is challenging due to limited training data and the complex relationship between process variations and $V_{min}$. To address these issues, we propose a novel transfer learning framework that leverages abundant legacy data from the 16nm technology node to enable accurate $V_{min}$ prediction at the advanced 5nm node. A key innovation of our approach is the integration of input features derived from on-chip silicon odometer sensor data, which provide fine-grained characterization of localized process variations – an essential factor at the 5nm node – resulting in significantly improved prediction accuracy.

[737] A-FloPS: Accelerating Diffusion Sampling with Adaptive Flow Path Sampler

Cheng Jin, Zhenyu Xiao, Yuantao Gu

Main category: cs.LG

TL;DR: A-FloPS is a training-free acceleration framework that reparameterizes diffusion model sampling trajectories into flow-matching form with adaptive velocity decomposition, achieving state-of-the-art performance with as few as 5 function evaluations.

DetailsMotivation: Diffusion models provide state-of-the-art generative performance but suffer from computational inefficiency due to their iterative sampling process. Existing training-free acceleration methods are fundamentally constrained by inefficient sampling trajectories.

Method: The method reparameterizes sampling trajectories of pre-trained diffusion models into flow-matching form and augments with adaptive velocity decomposition. It maps diffusion scores to flow-compatible velocities and factorizes velocity field into linear drift and residual components with suppressed temporal variation.

Result: Extensive experiments show A-FloPS consistently outperforms state-of-the-art training-free samplers in both sample quality and efficiency. With only 5 function evaluations, it achieves substantially lower FID and generates sharper, more coherent images in conditional image generation and text-to-image synthesis.

Conclusion: A-FloPS is a versatile and effective solution for high-quality, low-latency generative modeling that also improves native flow-based generative models, demonstrating its generality across different generative modeling approaches.

Abstract: Diffusion models deliver state-of-the-art generative performance across diverse modalities but remain computationally expensive due to their inherently iterative sampling process. Existing training-free acceleration methods typically improve numerical solvers for the reverse-time ODE, yet their effectiveness is fundamentally constrained by the inefficiency of the underlying sampling trajectories. We propose A-FloPS (Adaptive Flow Path Sampler), a principled, training-free framework that reparameterizes the sampling trajectory of any pre-trained diffusion model into a flow-matching form and augments it with an adaptive velocity decomposition. The reparameterization analytically maps diffusion scores to flow-compatible velocities, yielding integration-friendly trajectories without retraining. The adaptive mechanism further factorizes the velocity field into a linear drift term and a residual component whose temporal variation is actively suppressed, restoring the accuracy benefits of high-order integration even in extremely low-NFE regimes. Extensive experiments on conditional image generation and text-to-image synthesis show that A-FloPS consistently outperforms state-of-the-art training-free samplers in both sample quality and efficiency. Notably, with as few as $5$ function evaluations, A-FloPS achieves substantially lower FID and generates sharper, more coherent images. The adaptive mechanism also improves native flow-based generative models, underscoring its generality. These results position A-FloPS as a versatile and effective solution for high-quality, low-latency generative modeling.

[738] Exploring and Reshaping the Weight Distribution in LLM

Chunming Ye, Songzhou Li, Xu Xu

Main category: cs.LG

TL;DR: This paper analyzes weight distribution correlations across LLM layers, discovers power-law distributions in cosine distances, develops a data generator to create distribution-aligned weights, and improves LoRA training performance through weight reshaping.

DetailsMotivation: To understand how weight distribution correlations between different layers in large language models affect LoRA training effectiveness, and leverage these insights to improve training performance.

Method: Analyzed cosine distances between weights of different layers using singular value decomposition, discovered power-law distribution characteristics, designed a data generator using Gaussian process and Pareto distribution functions, and reshaped LoRA initialization weights based on distribution patterns.

Result: Experimental results show that reshaping LoRA initialization weights according to the discovered distribution characteristics improves LoRA training performance without changing model structure or training process.

Conclusion: The power-law distribution characteristics of weight correlations across LLM layers can be leveraged to optimize LoRA initialization, leading to improved training effectiveness and model performance.

Abstract: The performance of Large Language Models is influenced by their characteristics such as architecture, model sizes, decoding methods and so on. Due to differences in structure or function, the weights in different layers of large models have varying distributions. This paper explores the correlations between different types of layers in terms of weights distribution and studies the potential impact of these correlations on LoRA training effectiveness. Firstly, the study reveals that in the model the cosine distances between weights of different layers manifest power-law distribution. We extract Query-projection, down-projection and other weight matrices from the self-attention layers and MLP layers, calculate the singular values of the matrices using singular value decomposition, and organize a certain number of singular values into matrices according to projection’s type. By analyzing the probability distribution of the cosine distances between these matrices, it is found that the cosine distances values between them have distinct power-law distribution characteristics. Secondly, based on the results of distance calculations and analysis across different layers of model, a qualitative method is proposed to describe the distribution characteristics of different models. Next, to construct weights that align with the distribution characteristics, a data generator is designed using a combination of Gaussian process and Pareto distribution functions. The generator is used to simulate the generation of data that aligns with specific distribution characteristics. Finally, based on the aforementioned distribution characteristics and data generation method, the weights in LoRA initialization are reshaped for training. Experimental results indicate that, without altering the model structure or training process, this method achieves a certain improvement in the performance of LoRA training.

[739] Teaching AI to Remember: Insights from Brain-Inspired Replay in Continual Learning

Jina Kim

Main category: cs.LG

TL;DR: Internal replay mechanism inspired by brain memory consolidation helps mitigate catastrophic forgetting in continual learning, but reduces initial task accuracy and increases representational overlap, creating a trade-off between memory stability and learning plasticity.

DetailsMotivation: Address catastrophic forgetting in artificial neural networks during continual learning by drawing inspiration from human brain's memory consolidation processes, specifically focusing on internal replay mechanisms.

Method: Evaluated internal replay mechanism using CIFAR-100 dataset in class-incremental setting, both in isolation and combined with Synaptic Intelligence (SI). Used various analysis techniques including log-likelihood distributions, reconstruction errors, silhouette scores, and UMAP projections.

Result: Internal replay significantly reduces forgetting, especially when paired with SI, but at the cost of reduced initial task accuracy. Analysis revealed increased representational overlap in latent space, potentially limiting task-specific differentiation.

Conclusion: Current brain-inspired methods have limitations in balancing retention and adaptability. The findings highlight a trade-off between memory stability and learning plasticity, suggesting need for future research to better balance these competing objectives in continual learning systems.

Abstract: Artificial neural networks (ANNs) continue to face challenges in continual learning, particularly due to catastrophic forgetting, the loss of previously learned knowledge when acquiring new tasks. Inspired by memory consolidation in the human brain, we investigate the internal replay mechanism proposed by~\citep{brain_inspired_replay1}, which reactivates latent representations of prior experiences during learning. As internal replay was identified as the most influential component among the brain-inspired mechanisms in their framework, it serves as the central focus of our in-depth investigation. Using the CIFAR-100 dataset in a class-incremental setting, we evaluate the effectiveness of internal replay, both in isolation and in combination with Synaptic Intelligence (SI). Our experiments show that internal replay significantly mitigates forgetting, especially when paired with SI, but at the cost of reduced initial task accuracy, highlighting a trade-off between memory stability and learning plasticity. Further analyses using log-likelihood distributions, reconstruction errors, silhouette scores, and UMAP projections reveal that internal replay increases representational overlap in latent space, potentially limiting task-specific differentiation. These results underscore the limitations of current brain-inspired methods and suggest future directions for balancing retention and adaptability in continual learning systems.

[740] Adaptive Physics-Informed Neural Networks with Multi-Category Feature Engineering for Hydrogen Sorption Prediction in Clays, Shales, and Coals

Mohammad Nooraiepour, Mohammad Masoudi, Zezhang Song, Helge Hellevang

Main category: cs.LG

TL;DR: Adaptive physics-informed neural network with multi-category feature engineering achieves highly accurate hydrogen sorption prediction across clays, shales, and coals with R2=0.979 and robust uncertainty quantification.

DetailsMotivation: Traditional experimental methods for hydrogen sorption prediction are time-consuming, error-prone, and limited in capturing geological heterogeneity, which hinders underground hydrogen storage, natural hydrogen exploration, and radioactive waste containment applications.

Method: Developed an adaptive physics-informed neural network (PINN) framework integrating classical isotherm models with thermodynamic constraints. Used deep residual networks with multi-head attention, adaptive loss functions, and Monte Carlo dropout for uncertainty quantification. Employed multi-category feature engineering across seven categories and comprehensive dataset of 155 samples (50 clays, 60 shales, 45 coals).

Result: Achieved significant accuracy with R2=0.979 and RMSE=0.045 mol/kg, 67% faster convergence despite 15-fold increased complexity. Robust lithology-specific performance: clays (R2=0.981), shales (R2=0.971), coals (R2=0.978) with 85-91% reliability scores. Interpretability analysis showed hydrogen adsorption capacity dominates predictions with 86.7% feature pairs exhibiting strong interactions.

Conclusion: The adaptive physics-informed framework provides accurate, efficient hydrogen sorption prediction, accelerates site screening, and enables risk-informed decision-making through robust uncertainty quantification, validating the necessity of non-linear modeling approaches for complex geological systems.

Abstract: Accurate prediction of hydrogen sorption in clays, shales, and coals is vital for advancing underground hydrogen storage, natural hydrogen exploration, and radioactive waste containment. Traditional experimental methods, while foundational, are time-consuming, error-prone, and limited in capturing geological heterogeneity. This study introduces an adaptive physics-informed neural network (PINN) framework with multi-category feature engineering to enhance hydrogen sorption prediction. The framework integrates classical isotherm models with thermodynamic constraints to ensure physical consistency while leveraging deep learning flexibility. A comprehensive dataset consisting of 155 samples, which includes 50 clays, 60 shales, and 45 coals, was employed, incorporating diverse compositional properties and experimental conditions. Multi-category feature engineering across seven categories captured complex sorption dynamics. The PINN employs deep residual networks with multi-head attention, optimized via adaptive loss functions and Monte Carlo dropout for uncertainty quantification. K-fold cross-validation and hyperparameter optimization achieve significant accuracy (R2 = 0.979, RMSE = 0.045 mol per kg) with 67% faster convergence despite 15-fold increased complexity. The framework demonstrates robust lithology-specific performance across clay minerals (R2 = 0.981), shales (R2 = 0.971), and coals (R2 = 0.978), maintaining 85-91% reliability scores. Interpretability analysis via SHAP, accumulated local effects, and Friedman’s H-statistics reveal that hydrogen adsorption capacity dominates predictions, while 86.7% of feature pairs exhibit strong interactions, validating the necessity of non-linear modeling approaches. This adaptive physics-informed framework accelerates site screening and enables risk-informed decision-making through robust uncertainty quantification.

[741] Applying Deep Learning to Anomaly Detection of Russian Satellite Activity for Indications Prior to Military Activity

David Kurtenbach, Megan Manly, Zach Metzinger

Main category: cs.LG

TL;DR: Deep learning anomaly detection applied to Russian space objects reveals statistically significant behavioral changes before Ukraine invasion, providing potential early warning indicators for future conflicts.

DetailsMotivation: To detect early warning signs of aggressive military behavior by analyzing anomalous activity patterns of Russian-owned space objects prior to the Ukraine invasion using publicly available orbital data.

Method: Used multiple deep learning models (Isolation Forest, traditional AE, VAE, KAN, and novel Anchor AE) on 5-year TLE data, training individual models for each space object to detect reconstruction errors exceeding threshold sigma in six orbital elements.

Result: Identified statistically significant anomalies in Russian RSO activity patterns, with detailed anomalous findings at the individual orbital element level during the 6-month pre-invasion period.

Conclusion: Deep learning anomaly detection can effectively identify early warning indicators of military aggression through space object behavior analysis, providing valuable insights for future conflict prediction and monitoring.

Abstract: We apply deep learning techniques for anomaly detection to analyze activity of Russian-owned resident space objects (RSO) prior to the Ukraine invasion and assess the results for any findings that can be used as indications and warnings (I&W) of aggressive military behavior for future conflicts. Through analysis of anomalous activity, an understanding of possible tactics and procedures can be established to assess the existence of statistically significant changes in Russian RSO pattern of life/pattern of behavior (PoL/PoB) using publicly available two-line element (TLE) data. This research looks at statistical and deep learning approaches to assess anomalous activity. The deep learning methods assessed are isolation forest (IF), traditional autoencoder (AE), variational autoencoder (VAE), Kolmogorov Arnold Network (KAN), and a novel anchor-loss based autoencoder (Anchor AE). Each model is used to establish a baseline of on-orbit activity based on a five-year data sample. The primary investigation period focuses on the six months leading up to the invasion date of February 24, 2022. Additional analysis looks at RSO activity during an active combat period by sampling TLE data after the invasion date. The deep learning autoencoder models identify anomalies based on reconstruction errors that surpass a threshold sigma. To capture the nuance and unique characteristics of each RSO an individual model was trained for each observed space object. The research made an effort to prioritize explainability and interpretability of the model results thus each observation was assessed for anomalous behavior of the individual six orbital elements versus analyzing the input data as a single monolithic observation. The results demonstrate not only statistically significant anomalies of Russian RSO activity but also details anomalous findings to the individual orbital element.

[742] From Data to Decision: A Multi-Stage Framework for Class Imbalance Mitigation in Optical Network Failure Analysis

Yousuf Moiz Ali, Jaroslaw E. Prilepsky, Nicola Sambo, Joao Pedro, Mohammad M. Hosseini, Antonio Napoli, Sergei K. Turitsyn, Pedro Freire

Main category: cs.LG

TL;DR: Comparison of pre-, in-, and post-processing methods for class imbalance in optical network failure management, showing post-processing excels in detection while GenAI methods lead in identification.

DetailsMotivation: Severe class imbalance in optical network failure management where normal instances vastly outnumber failure cases, with post-processing methods being largely unexplored compared to pre- and in-processing techniques.

Method: Direct comparison of pre-, in-, and post-processing approaches for class imbalance mitigation using experimental dataset, evaluating methods like Threshold Adjustment, Random Under-Sampling, SMOTE, Meta-Learning, and Generative AI approaches.

Result: For failure detection: post-processing (Threshold Adjustment) achieves highest F1 improvement (15.3%), Random Under-Sampling provides fastest inference. For failure identification: GenAI methods deliver best performance gains (24.2%), post-processing shows limited impact in multi-class settings.

Conclusion: Method effectiveness depends on scenario: over-sampling (SMOTE) best when class overlap present and latency critical; Meta-Learning best without latency constraints; Generative AI most effective in low-overlap scenarios with minimal inference time.

Abstract: Machine learning-based failure management in optical networks has gained significant attention in recent years. However, severe class imbalance, where normal instances vastly outnumber failure cases, remains a considerable challenge. While pre- and in-processing techniques have been widely studied, post-processing methods are largely unexplored. In this work, we present a direct comparison of pre-, in-, and post-processing approaches for class imbalance mitigation in failure detection and identification using an experimental dataset. For failure detection, post-processing methods-particularly Threshold Adjustment-achieve the highest F1 score improvement (up to 15.3%), while Random Under-Sampling provides the fastest inference. In failure identification, GenAI methods deliver the most substantial performance gains (up to 24.2%), whereas post-processing shows limited impact in multi-class settings. When class overlap is present and latency is critical, over-sampling methods such as the SMOTE are most effective; without latency constraints, Meta-Learning yields the best results. In low-overlap scenarios, Generative AI approaches provide the highest performance with minimal inference time.

[743] T-MLP: Tailed Multi-Layer Perceptron for Level-of-Detail Signal Representation

Chuanxiang Yang, Yuanfeng Zhou, Guangshun Wei, Siyu Ren, Yuan Liu, Junhui Hou, Wenping Wang

Main category: cs.LG

TL;DR: T-MLP extends standard MLP with multiple output branches (tails) to enable multi-scale Level-of-Detail signal representation, outperforming existing neural LoD methods.

DetailsMotivation: Standard MLPs operate at a single scale and lack native support for Level-of-Detail representation, which is critical for efficient modeling and transmission of signals like images and 3D shapes.

Method: Introduces Tailed MLP (T-MLP) that attaches multiple output branches to hidden layers, enabling direct supervision at multiple depths with a specialized loss formulation and training strategy.

Result: Extensive experiments show T-MLP outperforms other neural LoD baselines across various signal representation tasks.

Conclusion: The proposed T-MLP architecture successfully enables multi-scale Level-of-Detail representation by extending standard MLPs with multiple output branches and appropriate training strategies.

Abstract: Level-of-detail (LoD) representation is critical for efficiently modeling and transmitting various types of signals, such as images and 3D shapes. In this work, we present a novel neural architecture that supports LoD signal representation. Our architecture is based on an elaborate modification of the widely used Multi-Layer Perceptron (MLP), which inherently operates at a single scale and therefore lacks native support for LoD. Specifically, we introduce the Tailed Multi-Layer Perceptron (T-MLP) that extends the MLP by attaching multiple output branches, also called tails, to its hidden layers, enabling direct supervision at multiple depths. Our loss formulation and training strategy allow each hidden layer to effectively learn a target signal at a specific LoD, thus enabling multi-scale modeling. Extensive experimental results show that our T-MLP outperforms other neural LoD baselines across a variety of signal representation tasks.

[744] AnomalyExplainer Explainable AI for LLM-based anomaly detection using BERTViz and Captum

Prasasthy Balasubramanian, Dumindu Kankanamge, Ekaterina Gilman, Mourad Oussalah

Main category: cs.LG

TL;DR: A framework combining anomaly detection with visual XAI tools (BERTViz, Captum) and natural language explanations to improve cybersecurity threat analysis, using RoBERTa which achieved 99.6% accuracy on HDFS logs.

DetailsMotivation: Address limitations in conversational AI for cybersecurity including false positives, complex model management, and lack of trust in AI decisions despite existing XAI approaches.

Method: Developed a framework that detects anomalies and provides explanations through visual tools (BERTViz and Captum) combined with natural language reports based on attention outputs from transformer models.

Result: RoBERTa achieved 99.6% accuracy and strong anomaly detection, outperforming Falcon-7B and DeBERTa, with better flexibility than Mistral-7B on HDFS dataset. User feedback confirmed ease of use and improved anomaly understanding.

Conclusion: The framework successfully reduces manual effort, speeds up remediation, and strengthens cybersecurity workflows by providing high-quality explanations that build trust in AI-driven security systems.

Abstract: Conversational AI and Large Language Models (LLMs) have become powerful tools across domains, including cybersecurity, where they help detect threats early and improve response times. However, challenges such as false positives and complex model management still limit trust. Although Explainable AI (XAI) aims to make AI decisions more transparent, many security analysts remain uncertain about its usefulness. This study presents a framework that detects anomalies and provides high-quality explanations through visual tools BERTViz and Captum, combined with natural language reports based on attention outputs. This reduces manual effort and speeds up remediation. Our comparative analysis showed that RoBERTa offers high accuracy (99.6 %) and strong anomaly detection, outperforming Falcon-7B and DeBERTa, as well as exhibiting better flexibility than large-scale Mistral-7B on the HDFS dataset from LogHub. User feedback confirms the chatbot’s ease of use and improved understanding of anomalies, demonstrating the ability of the developed framework to strengthen cybersecurity workflows.

[745] SynCircuit: Automated Generation of New Synthetic RTL Circuits Can Enable Big Data in Circuits

Shang Liu, Jing Wang, Wenji Fang, Zhiyao Xie

Main category: cs.LG

TL;DR: SynCircuit is a novel framework that generates synthetic circuit data in HDL format using diffusion models and MCTS optimization to address the data scarcity problem in AI-assisted IC design.

DetailsMotivation: The lack of publicly available circuit design data is the primary bottleneck for developing AI-assisted IC design methods, as current datasets are extremely limited.

Method: Three-step framework: 1) Customized diffusion-based generative model for Directed Cyclic Graph generation, 2) Circuit constraint enforcement through graph refinement, 3) Monte Carlo tree search for logic redundancy optimization.

Result: Experimental results show SynCircuit generates more realistic synthetic circuits and enhances ML model performance in downstream circuit design tasks.

Conclusion: SynCircuit successfully addresses the circuit data scarcity problem and demonstrates the potential of synthetic data generation for advancing AI-assisted IC design methodologies.

Abstract: In recent years, AI-assisted IC design methods have demonstrated great potential, but the availability of circuit design data is extremely limited, especially in the public domain. The lack of circuit data has become the primary bottleneck in developing AI-assisted IC design methods. In this work, we make the first attempt, SynCircuit, to generate new synthetic circuits with valid functionalities in the HDL format. SynCircuit automatically generates synthetic data using a framework with three innovative steps: 1) We propose a customized diffusion-based generative model to resolve the Directed Cyclic Graph (DCG) generation task, which has not been well explored in the AI community. 2) To ensure our circuit is valid, we enforce the circuit constraints by refining the initial graph generation outputs. 3) The Monte Carlo tree search (MCTS) method further optimizes the logic redundancy in the generated graph. Experimental results demonstrate that our proposed SynCircuit can generate more realistic synthetic circuits and enhance ML model performance in downstream circuit design tasks.

[746] Balanced Multimodal Learning: An Unidirectional Dynamic Interaction Perspective

Shijie Wang, Li Zhang, Xinyan Liang, Yuhua Qian, Shen Hu

Main category: cs.LG

TL;DR: UDI introduces a sequential training approach that abandons joint loss to address modality imbalance in multimodal learning, using anchor modality guidance and dynamic interaction adjustment.

DetailsMotivation: Traditional multimodal joint loss causes modality imbalance where strong modalities dominate weaker ones, limiting individual modality information utilization and inter-modality interactions.

Method: Unidirectional Dynamic Interaction (UDI) - sequential training where anchor modality is trained first to convergence, then guides other modalities via unsupervised loss with dynamic interaction adjustment.

Result: UDI outperforms existing methods in handling modality imbalance and achieves performance improvement in multimodal learning tasks.

Conclusion: UDI’s proactive sequential approach effectively prevents modality domination and enables better cross-modal feature learning compared to reactive joint loss methods.

Abstract: Multimodal learning typically utilizes multimodal joint loss to integrate different modalities and enhance model performance. However, this joint learning strategy can induce modality imbalance, where strong modalities overwhelm weaker ones and limit exploitation of individual information from each modality and the inter-modality interaction information.Existing strategies such as dynamic loss weighting, auxiliary objectives and gradient modulation mitigate modality imbalance based on joint loss. These methods remain fundamentally reactive, detecting and correcting imbalance after it arises, while leaving the competitive nature of the joint loss untouched. This limitation drives us to explore a new strategy for multimodal imbalance learning that does not rely on the joint loss, enabling more effective interactions between modalities and better utilization of information from individual modalities and their interactions. In this paper, we introduce Unidirectional Dynamic Interaction (UDI), a novel strategy that abandons the conventional joint loss in favor of a proactive, sequential training scheme. UDI first trains the anchor modality to convergence, then uses its learned representations to guide the other modality via unsupervised loss. Furthermore, the dynamic adjustment of modality interactions allows the model to adapt to the task at hand, ensuring that each modality contributes optimally. By decoupling modality optimization and enabling directed information flow, UDI prevents domination by any single modality and fosters effective cross-modal feature learning. Our experimental results demonstrate that UDI outperforms existing methods in handling modality imbalance, leading to performance improvement in multimodal learning tasks.

[747] A Multimodal Deep Learning Framework for Early Diagnosis of Liver Cancer via Optimized BiLSTM-AM-VMD Architecture

Cheng Cheng, Zeping Chen, Xavier Wang

Main category: cs.LG

TL;DR: A multimodal deep learning framework (BiLSTM-AM-VMD) combining bidirectional LSTM, multi-head attention, and variational mode decomposition for early liver cancer diagnosis using clinical, biochemical, and imaging data.

DetailsMotivation: To improve early liver cancer diagnosis by integrating heterogeneous medical data sources and enhancing both prediction accuracy and model interpretability.

Method: Proposes BiLSTM-AM-VMD framework that integrates bidirectional LSTM for sequence modeling, multi-head attention mechanism for feature importance, and variational mode decomposition for signal processing of multimodal medical data.

Result: Experimental results on real-world datasets show superior performance compared to traditional machine learning and baseline deep learning models.

Conclusion: The proposed multimodal framework effectively improves early liver cancer diagnosis accuracy while maintaining interpretability through attention mechanisms.

Abstract: This paper proposes a novel multimodal deep learning framework integrating bidirectional LSTM, multi-head attention mechanism, and variational mode decomposition (BiLSTM-AM-VMD) for early liver cancer diagnosis. Using heterogeneous data that include clinical characteristics, biochemical markers, and imaging-derived variables, our approach improves both prediction accuracy and interpretability. Experimental results on real-world datasets demonstrate superior performance over traditional machine learning and baseline deep learning models.

[748] Mitigating Clinician Information Overload: Generative AI for Integrated EHR and RPM Data Analysis

Ankit Shetgaonkar, Dipen Pradhan, Lakshit Arora, Sanjay Surendranath Girija, Shashank Kapoor, Aman Raj

Main category: cs.LG

TL;DR: GenAI and LLMs can help manage healthcare data overload from combined RPM and EHR sources by providing clinical insights and improving efficiency through natural language applications.

DetailsMotivation: The sheer volume and heterogeneity of patient data from real-time Remote Patient Monitoring (RPM) and traditional Electronic Health Records (EHRs) create significant information overload challenges for clinicians.

Method: The paper provides a comprehensive overview of GenAI capabilities and explores LLM-powered applications for navigating longitudinal patient data and providing clinical decision support through natural language dialogue.

Result: GenAI techniques show potential for streamlining clinician workflows, personalizing care, and enhancing clinical efficiency, though challenges around data integration, quality, privacy, safety validation, bias mitigation, and clinical acceptance remain.

Conclusion: This work represents the first summarization of GenAI techniques specifically addressing clinician data overload from combined RPM/EHR data complexities, highlighting both opportunities and critical challenges for implementation.

Abstract: Generative Artificial Intelligence (GenAI), particularly Large Language Models (LLMs), offer powerful capabilities for interpreting the complex data landscape in healthcare. In this paper, we present a comprehensive overview of the capabilities, requirements and applications of GenAI for deriving clinical insights and improving clinical efficiency. We first provide some background on the forms and sources of patient data, namely real-time Remote Patient Monitoring (RPM) streams and traditional Electronic Health Records (EHRs). The sheer volume and heterogeneity of this combined data present significant challenges to clinicians and contribute to information overload. In addition, we explore the potential of LLM-powered applications for improving clinical efficiency. These applications can enhance navigation of longitudinal patient data and provide actionable clinical decision support through natural language dialogue. We discuss the opportunities this presents for streamlining clinician workflows and personalizing care, alongside critical challenges such as data integration complexity, ensuring data quality and RPM data reliability, maintaining patient privacy, validating AI outputs for clinical safety, mitigating bias, and ensuring clinical acceptance. We believe this work represents the first summarization of GenAI techniques for managing clinician data overload due to combined RPM / EHR data complexities.

[749] Experimental Assessment of a Multi-Class AI/ML Architecture for Real-Time Characterization of Cyber Events in a Live Research Reactor

Zachery Dahm, Konstantinos Vasili, Vasileios Theos, Konstantinos Gkouliaras, William Richards, True Miller, Brian Jowers, Stylianos Chatzidakis

Main category: cs.LG

TL;DR: AI/ML multi-layered architecture successfully distinguishes cybersecurity events from operational anomalies in nuclear reactors using combined IT/OT data streams.

DetailsMotivation: There's growing interest in applying AI/ML in nuclear industry for anomaly detection and cybersecurity, but limited research exists on real reactor implementations.

Method: Developed multi-layered AI/ML architecture integrating IT and OT data streams, tested on Purdue University’s PUR-1 research reactor with 14 system states and 13.8+ million data points including false data injections and DoS attacks.

Result: AI/ML successfully distinguished normal, abnormal, and cybersecurity events even during DoS attacks. Combined IT/OT data improved accuracy but presented synchronization challenges.

Conclusion: AI/ML shows significant promise for nuclear cybersecurity but requires refinement for complex event differentiation and multi-class architectures.

Abstract: There is increased interest in applying Artificial Intelligence and Machine Learning (AI/ML) within the nuclear industry and nuclear engineering community. Effective implementation of AI/ML could offer benefits to the nuclear domain, including enhanced identification of anomalies, anticipation of system failures, and operational schedule optimization. However, limited work has been done to investigate the feasibility and applicability of AI/ML tools in a functioning nuclear reactor. Here, we go beyond the development of a single model and introduce a multi-layered AI/ML architecture that integrates both information technology and operational technology data streams to identify, characterize, and differentiate (i) among diverse cybersecurity events and (ii) between cyber events and other operational anomalies. Leveraging Purdue Universitys research reactor, PUR-1, we demonstrate this architecture through a representative use case that includes multiple concurrent false data injections and denial-of-service attacks of increasing complexity under realistic reactor conditions. The use case includes 14 system states (1 normal, 13 abnormal) and over 13.8 million multi-variate operational and information technology data points. The study demonstrated the capability of AI/ML to distinguish between normal, abnormal, and cybersecurity-related events, even under challenging conditions such as denial-of-service attacks. Combining operational and information technology data improved classification accuracy but posed challenges related to synchronization and collection during certain cyber events. While results indicate significant promise for AI/ML in nuclear cybersecurity, the findings also highlight the need for further refinement in handling complex event differentiation and multi-class architectures.

[750] Speech Foundation Models Generalize to Time Series Tasks from Wearable Sensor Data

Jaya Narain, Zakaria Aldeneh, Shirley Ren

Main category: cs.LG

TL;DR: Speech foundation models like HuBERT and wav2vec 2.0 can be effectively applied to wearable sensor time series data, achieving state-of-the-art performance on tasks like mood classification, arrhythmia detection, and activity classification through simple probing methods.

DetailsMotivation: Both speech and sensor time series data share similar time- and frequency-domain characteristics, suggesting that speech foundation models might learn domain-independent representations that could benefit wearable sensor applications.

Method: Using pre-trained speech foundation models (HuBERT and wav2vec 2.0) to extract features from wearable sensor time series data, then training simple probes on these features for downstream classification tasks.

Result: The speech model features significantly outperform self-supervised models trained directly on modality-specific datasets, with convolutional feature encoders from speech models showing particularly strong relevance for wearable sensor tasks.

Conclusion: Speech foundation models provide effective domain-independent representations for time series data, improving performance and robustness for data-scarce wearable sensor tasks, paving the way for generalized time series models across speech and sensor domains.

Abstract: Both speech and sensor time series data encode information in both the time- and frequency- domains, like spectral powers and waveform shapelets. We show that speech foundation models learn representations that are domain-independent and achieve state-of-the-art performance on time series tasks from wearable sensors. Probes trained on features extracted from HuBERT and wav2vec 2.0 outperform those extracted from self-supervised models trained directly on modality specific datasets for mood classification, arrhythmia detection, and activity classification tasks. We find a particularly strong relevance of the convolutional feature encoders from speech models for wearable sensor tasks. The methods proposed here improve performance and robustness for data-scarce time series tasks, using simple probing methods. This work is a step towards generalized time series models for speech and sensor data, a topic for further exploration.

[751] Data Cartography for Detecting Memorization Hotspots and Guiding Data Interventions in Generative Models

Laksh Patel, Neel Shanbhag

Main category: cs.LG

TL;DR: GenDataCarto framework uses difficulty and memorization scores to identify and prune problematic training data, reducing data leakage by 40% with minimal performance impact.

DetailsMotivation: Modern generative models risk overfitting and unintentionally memorizing rare training examples, which can be exploited by adversaries or artificially inflate benchmark performance.

Method: Assigns each pretraining sample a difficulty score (early-epoch loss) and memorization score (frequency of forget events), then partitions examples into four quadrants to guide targeted pruning and up-/down-weighting.

Result: Reduces synthetic canary extraction success by over 40% at just 10% data pruning, while increasing validation perplexity by less than 0.5%.

Conclusion: Principled data interventions can dramatically mitigate leakage with minimal cost to generative performance, providing a data-centric approach to improve model security.

Abstract: Modern generative models risk overfitting and unintentionally memorizing rare training examples, which can be extracted by adversaries or inflate benchmark performance. We propose Generative Data Cartography (GenDataCarto), a data-centric framework that assigns each pretraining sample a difficulty score (early-epoch loss) and a memorization score (frequency of ``forget events’’), then partitions examples into four quadrants to guide targeted pruning and up-/down-weighting. We prove that our memorization score lower-bounds classical influence under smoothness assumptions and that down-weighting high-memorization hotspots provably decreases the generalization gap via uniform stability bounds. Empirically, GenDataCarto reduces synthetic canary extraction success by over 40% at just 10% data pruning, while increasing validation perplexity by less than 0.5%. These results demonstrate that principled data interventions can dramatically mitigate leakage with minimal cost to generative performance.

[752] Learning to Refine: Self-Refinement of Parallel Reasoning in LLMs

Qibin Wang, Pu Zhao, Shaohan Huang, Fangkai Yang, Lu Wang, Furu Wei, Qingwei Lin, Saravan Rajmohan, Dongmei Zhang

Main category: cs.LG

TL;DR: GSR is a parallel test-time scaling framework where LLMs generate multiple candidate responses and then self-refine them into a superior solution, achieving SOTA performance on math benchmarks.

DetailsMotivation: Existing test-time scaling methods like Best-of-N and majority voting fail when all candidate responses are incorrect, and adding separate selection models increases deployment costs.

Method: A unified model generates candidate responses in parallel, then performs self-refinement using a hybrid training pipeline that optimizes both direct problem-solving and candidate refinement objectives.

Result: Achieves state-of-the-art performance across five mathematical benchmarks, with learned self-refinement skills being model-agnostic and robust across different model scales and out-of-distribution tasks.

Conclusion: GSR provides an effective framework for enhancing LLM reasoning without additional deployment costs, demonstrating strong generalization capabilities across various model sizes and reasoning tasks.

Abstract: To further enhance the ability of Large Language Models (LLMs) to solve complex, multi-step reasoning problems, test-time scaling (TTS) methods have gained widespread attention. Existing approaches such as Best-of-N and majority voting are limited as their performance depends on the quality of candidate responses, making them unable to produce a correct solution when all candidates are incorrect. Introducing an additional model to select the best response also incurs significant deployment costs. To this end, we introduce Generative Self-Refinement (GSR), a novel parallel test-time scaling framework where a unified model first generates a set of candidate responses in parallel and then performs self-refinement to synthesize a new superior solution based on a prompt consisting of the problem and these candidates. However, LLMs struggle to perform refinement effectively when prompted directly. Therefore, we design a hybrid training pipeline by jointly optimizing for two complementary objectives, solving problems directly and refining candidate responses. Experimental results demonstrate that our method achieves state-of-the-art performance across five mathematical benchmarks. We further show that this learned self-refinement skill is a model-agnostic enhancement, robust across different model scales and generalizing to out-of-distribution reasoning tasks.

[753] Learning to Coordinate: Distributed Meta-Trajectory Optimization Via Differentiable ADMM-DDP

Bingheng Wang, Yichao Gao, Tianchen Sun, Lin Zhao

Main category: cs.LG

TL;DR: L2C is a meta-learning framework that optimizes hyperparameters for ADMM-DDP trajectory coordination in multi-agent systems, enabling adaptive performance across diverse tasks and configurations with efficient gradient computation.

DetailsMotivation: Distributed trajectory optimization via ADMM-DDP requires extensive tuning of tightly coupled hyperparameters that govern both local task performance and global coordination, which is challenging and time-consuming.

Method: Proposes Learning to Coordinate (L2C) framework with agent-wise neural networks to meta-learn hyperparameters, differentiates end-to-end through ADMM-DDP pipeline, reuses DDP components for efficient meta-gradient computation, and uses truncated iterations with optimized ADMM penalty parameters.

Result: L2C generates dynamically feasible trajectories in high-fidelity simulation, reconfigures quadrotor formations for safe 6-DoF load manipulation, adapts to varying team sizes and task conditions, and achieves up to 88% faster gradient computation than state-of-the-art methods.

Conclusion: L2C provides a general framework for meta-learning coordination hyperparameters that enables robust adaptation across diverse multi-agent tasks while significantly improving computational efficiency compared to existing approaches.

Abstract: Distributed trajectory optimization via ADMM-DDP is a powerful approach for coordinating multi-agent systems, but it requires extensive tuning of tightly coupled hyperparameters that jointly govern local task performance and global coordination. In this paper, we propose Learning to Coordinate (L2C), a general framework that meta-learns these hyperparameters, modeled by lightweight agent-wise neural networks, to adapt across diverse tasks and agent configurations. L2C differentiates end-to-end through the ADMM-DDP pipeline in a distributed manner. It also enables efficient meta-gradient computation by reusing DDP components such as Riccati recursions and feedback gains. These gradients correspond to the optimal solutions of distributed matrix-valued LQR problems, coordinated across agents via an auxiliary ADMM framework that becomes convex under mild assumptions. Training is further accelerated by truncating iterations and meta-learning ADMM penalty parameters optimized for rapid residual reduction, with provable Lipschitz-bounded gradient errors. On a challenging cooperative aerial transport task, L2C generates dynamically feasible trajectories in high-fidelity simulation using IsaacSIM, reconfigures quadrotor formations for safe 6-DoF load manipulation in tight spaces, and adapts robustly to varying team sizes and task conditions, while achieving up to $88%$ faster gradient computation than state-of-the-art methods.

[754] Centralized vs. Federated Learning for Educational Data Mining: A Comparative Study on Student Performance Prediction with SAEB Microdata

Rodrigo Tertulino

Main category: cs.LG

TL;DR: Federated Learning with FedProx algorithm achieves 61.23% accuracy for student performance prediction while preserving privacy under Brazil’s data protection laws, with only marginal performance loss compared to centralized models.

DetailsMotivation: Privacy legislation like Brazil's LGPD restricts centralized collection of sensitive student data, creating barriers for educational data mining and AI applications that require privacy-preserving computational approaches.

Method: Used Federated Learning (FedProx algorithm) with Deep Neural Network trained across 50 simulated schools, benchmarked against centralized XGBoost model on over 2 million student records from Brazilian Basic Education Assessment System (SAEB).

Result: Centralized XGBoost achieved 63.96% accuracy while federated model reached 61.23% accuracy - showing only marginal performance loss in exchange for robust privacy guarantees.

Conclusion: Federated Learning is a viable and effective solution for collaborative predictive modeling in Brazilian educational context that complies with LGPD privacy requirements.

Abstract: The application of data mining and artificial intelligence in education offers unprecedented potential for personalizing learning and early identification of at-risk students. However, the practical use of these techniques faces a significant barrier in privacy legislation, such as Brazil’s General Data Protection Law (LGPD), which restricts the centralization of sensitive student data. To resolve this challenge, privacy-preserving computational approaches are required. The present study evaluates the feasibility and effectiveness of Federated Learning, specifically the FedProx algorithm, to predict student performance using microdata from the Brazilian Basic Education Assessment System (SAEB). A Deep Neural Network (DNN) model was trained in a federated manner, simulating a scenario with 50 schools, and its performance was rigorously benchmarked against a centralized eXtreme Gradient Boosting (XGBoost) model. The analysis, conducted on a universe of over two million student records, revealed that the centralized model achieved an accuracy of 63.96%. Remarkably, the federated model reached a peak accuracy of 61.23%, demonstrating a marginal performance loss in exchange for a robust privacy guarantee. The results indicate that Federated Learning is a viable and effective solution for building collaborative predictive models in the Brazilian educational context, in alignment with the requirements of the LGPD.

[755] Yet Unnoticed in LSTM: Binary Tree Based Input Reordering, Weight Regularization, and Gate Nonlinearization

Mojtaba Moattari

Main category: cs.LG

TL;DR: The paper proposes three enhancements to LSTM models: input reordering to prioritize specific indices, weight normalization with optimal Lp norms, and nonlinearized gates using FFNNs to better emphasize past inputs.

DetailsMotivation: Current LSTM models don't optimally focus on specific old information or long-term dependencies, and existing approaches lack proper weight normalization and gate nonlinearization techniques.

Method: Three main approaches: 1) Input reordering to prioritize certain input indices, 2) Weight normalization with optimal Lp norm selection through supervised loss, 3) Nonlinearizing gates using small FFNNs analogous to attention mechanisms.

Result: The proposed approaches were implemented and compared with simple LSTM, showing improved accuracy in text classification tasks.

Conclusion: The enhancements to LSTM models through input reordering, weight normalization, and gate nonlinearization effectively improve model performance in handling long-term dependencies and text classification accuracy.

Abstract: LSTM models used in current Machine Learning literature and applications, has a promising solution for permitting long term information using gating mechanisms that forget and reduce effect of current input information. However, even with this pipeline, they do not optimally focus on specific old index or long-term information. This paper elaborates upon input reordering approaches to prioritize certain input indices. Moreover, no LSTM based approach is found in the literature that examines weight normalization while choosing the right weight and exponent of Lp norms through main supervised loss function. In this paper, we find out which norm best finds relationship between weights to either smooth or sparsify them. Lastly, gates, as weighted representations of inputs and states, which control reduction-extent of current input versus previous inputs (~ state), are not nonlinearized enough (through a small FFNN). As analogous to attention mechanisms, gates easily filter current information to bold (emphasize on) past inputs. Nonlinearized gates can more easily tune up to peculiar nonlinearities of specific input in the past. This type of nonlinearization is not proposed in the literature, to the best of author’s knowledge. The proposed approaches are implemented and compared with a simple LSTM to understand their performance in text classification tasks. The results show they improve accuracy of LSTM.

[756] Learning from Peers: Collaborative Ensemble Adversarial Training

Li Dengjin, Guo Yanming, Xie Yuxiang, Li Zheng, Chen Jiangming, Li Xiaolong, Lao Mingrui

Main category: cs.LG

TL;DR: CEAT improves ensemble adversarial training by focusing on samples with classification disparities between sub-models, using probability disparities to assign adaptive weights and enhance cooperative learning.

DetailsMotivation: Current EAT strategies train sub-models independently, missing cooperative benefits. Samples with classification disparities between sub-models are found to be near decision boundaries and crucial for ensemble robustness.

Method: Collaborative Ensemble Adversarial Training (CEAT) that gives greater attention to samples with larger predictive disparities during adversarial training. Uses probability disparities to assign adaptive weights with calibrating distance regularization.

Result: Extensive experiments show CEAT achieves state-of-the-art performance over competitive EAT methods on widely-adopted datasets.

Conclusion: CEAT is an effective, model-agnostic approach that can be seamlessly integrated into various ensemble methods, demonstrating superior robustness through enhanced cooperative learning.

Abstract: Ensemble Adversarial Training (EAT) attempts to enhance the robustness of models against adversarial attacks by leveraging multiple models. However, current EAT strategies tend to train the sub-models independently, ignoring the cooperative benefits between sub-models. Through detailed inspections of the process of EAT, we find that that samples with classification disparities between sub-models are close to the decision boundary of ensemble, exerting greater influence on the robustness of ensemble. To this end, we propose a novel yet efficient Collaborative Ensemble Adversarial Training (CEAT), to highlight the cooperative learning among sub-models in the ensemble. To be specific, samples with larger predictive disparities between the sub-models will receive greater attention during the adversarial training of the other sub-models. CEAT leverages the probability disparities to adaptively assign weights to different samples, by incorporating a calibrating distance regularization. Extensive experiments on widely-adopted datasets show that our proposed method achieves the state-of-the-art performance over competitive EAT methods. It is noteworthy that CEAT is model-agnostic, which can be seamlessly adapted into various ensemble methods with flexible applicability.

[757] VariAntNet: Learning Decentralized Control of Multi-Agent Systems

Yigal Koifman, Erez Koifman, Eran Iceland, Ariel Barel, Alfred M. Bruckstein

Main category: cs.LG

TL;DR: VariAntNet is a deep learning-based decentralized control model for simple robotic swarms that significantly outperforms analytical solutions in gathering tasks, achieving double the convergence rate while maintaining swarm cohesion in disaster response scenarios.

DetailsMotivation: To address the challenge of maintaining swarm cohesion and avoiding fragmentation in simple robotic agents (Ant Robots) with limited sensing and no communication capabilities, particularly for time-critical disaster response applications like firefighting where traditional analytical methods are too slow.

Method: Proposed VariAntNet - a deep learning model with geometric feature extraction from unordered local observations, using a neural network trained with a novel differentiable multi-objective loss function that leverages visibility graph Laplacian matrix properties to promote swarm cohesiveness.

Result: VariAntNet achieved more than double the convergence rate compared to existing analytical solutions while maintaining high swarm connectivity across varying swarm sizes, demonstrating superior performance in the fundamental multi-agent gathering task with bearing-only limited-range sensing.

Conclusion: The paper presents a trade-off between guaranteed cohesion (analytical methods) and practical speed (learning-based methods) for time-critical scenarios, justifying VariAntNet’s approach which sacrifices some theoretical guarantees for significantly improved performance in emergency response operations.

Abstract: A simple multi-agent system can be effectively utilized in disaster response applications, such as firefighting. Such a swarm is required to operate in complex environments with limited local sensing and no reliable inter-agent communication or centralized control. These simple robotic agents, also known as Ant Robots, are defined as anonymous agents that possess limited sensing capabilities, lack a shared coordinate system, and do not communicate explicitly with one another. A key challenge for simple swarms lies in maintaining cohesion and avoiding fragmentation despite limited-range sensing. Recent advances in machine learning offer effective solutions to some of the classical decentralized control challenges. We propose VariAntNet, a deep learning-based decentralized control model designed to facilitate agent swarming and collaborative task execution. VariAntNet includes geometric features extraction from unordered, variable-sized local observations. It incorporates a neural network architecture trained with a novel, differentiable, multi-objective, mathematically justified loss function that promotes swarm cohesiveness by utilizing the properties of the visibility graph Laplacian matrix. VariAntNet is demonstrated on the fundamental multi-agent gathering task, where agents with bearing-only and limited-range sensing must gather at some location. VariAntNet significantly outperforms an existing analytical solution, achieving more than double the convergence rate while maintaining high swarm connectivity across varying swarm sizes. While the analytical solution guarantees cohesion, it is often too slow in practice. In time-critical scenarios, such as emergency response operations where lives are at risk, slower analytical methods are impractical and justify the loss of some agents within the swarm. This paper presents and analyzes this trade-off in detail.

[758] Robust Detection of Synthetic Tabular Data under Schema Variability

G. Charbel N. Kindji, Elisa Fromont, Lina Maria Rojas-Barahona, Tanguy Urvoy

Main category: cs.LG

TL;DR: Novel transformer architecture for detecting synthetic tabular data with variable schemas, achieving 7-point improvements in AUC/accuracy and additional 7-point gain with table adaptation.

DetailsMotivation: Addressing the overlooked challenge of detecting synthetic tabular data despite its ubiquity, especially difficult due to heterogeneous structure and unseen formats at test time.

Method: Introduces a datum-wise transformer architecture with table-adaptation component to handle variable and previously unseen table schemas.

Result: Significantly outperforms previous baseline with 7-point improvements in both AUC and accuracy, plus additional 7 accuracy points from table adaptation.

Conclusion: First strong evidence that detecting synthetic tabular data in real-world conditions is feasible and can be done with high reliability.

Abstract: The rise of powerful generative models has sparked concerns over data authenticity. While detection methods have been extensively developed for images and text, the case of tabular data, despite its ubiquity, has been largely overlooked. Yet, detecting synthetic tabular data is especially challenging due to its heterogeneous structure and unseen formats at test time. We address the underexplored task of detecting synthetic tabular data in the wild, where tables have variable and previously unseen schemas. We introduce a novel datum-wise transformer architecture that significantly outperforms the only previously published baseline, improving both AUC and accuracy by 7 points. By incorporating a table-adaptation component, our model gains an additional 7 accuracy points, demonstrating enhanced robustness. This work provides the first strong evidence that detecting synthetic tabular data in real-world conditions is not only feasible, but can be done with high reliability.

[759] Financial Decision Making using Reinforcement Learning with Dirichlet Priors and Quantum-Inspired Genetic Optimization

Prasun Nandy, Debjit Dhar, Rik Das

Main category: cs.LG

TL;DR: Hybrid RL framework with Dirichlet stochasticity and quantum genetic optimization for dynamic budget allocation, achieving near-perfect alignment with actual Apple financial data.

DetailsMotivation: Traditional budget allocation models struggle with stochastic and nonlinear financial data, requiring more adaptive approaches.

Method: Reinforcement learning agent with Dirichlet distribution for state evolution and quantum mutation-based genetic optimization to escape local minima.

Result: Achieved cosine similarity of 0.9990 and KL divergence of 0.0023 with actual Apple allocations on unseen data.

Conclusion: Combining deep RL, stochastic modeling, and quantum-inspired heuristics shows promise for adaptive enterprise budgeting.

Abstract: Traditional budget allocation models struggle with the stochastic and nonlinear nature of real-world financial data. This study proposes a hybrid reinforcement learning (RL) framework for dynamic budget allocation, enhanced with Dirichlet-inspired stochasticity and quantum mutation-based genetic optimization. Using Apple Inc. quarterly financial data (2009 to 2025), the RL agent learns to allocate budgets between Research and Development and Selling, General and Administrative to maximize profitability while adhering to historical spending patterns, with L2 penalties discouraging unrealistic deviations. A Dirichlet distribution governs state evolution to simulate shifting financial contexts. To escape local minima and improve generalization, the trained policy is refined using genetic algorithms with quantum mutation via parameterized qubit rotation circuits. Generation-wise rewards and penalties are logged to visualize convergence and policy behavior. On unseen fiscal data, the model achieves high alignment with actual allocations (cosine similarity 0.9990, KL divergence 0.0023), demonstrating the promise of combining deep RL, stochastic modeling, and quantum-inspired heuristics for adaptive enterprise budgeting.

[760] Pruning Weights but Not Truth: Safeguarding Truthfulness While Pruning LLMs

Yao Fu, Runchao Li, Xianxuan Long, Haotian Yu, Xiaotian Han, Yu Yin, Pan Li

Main category: cs.LG

TL;DR: Neural network pruning disrupts LLMs’ internal activation features needed for lie detection. The paper proposes TPLO (Truthful Pruning aligned by Layer-wise Outliers) to preserve lie detection capabilities while pruning, achieving 88% accuracy at 50% sparsity.

DetailsMotivation: Pruning LLMs for low-resource deployment inadvertently removes crucial internal activation features needed for lie detection, creating a need for pruning methods that preserve these critical capabilities.

Method: Proposed TPLO (Truthful Pruning aligned by Layer-wise Outliers) that emphasizes layers with more activation outliers and stronger discriminative features. Also introduced a prompting rule to enrich TruthfulQA benchmark for better calibration.

Result: Achieved 88% lie detection accuracy at 50% sparsity while preserving original LLM performance. Enhanced performance on TruthfulQA benchmark.

Conclusion: TPLO successfully prunes LLMs without sacrificing lie detection capabilities by focusing on layers with activation outliers and discriminative features, maintaining both model efficiency and truthfulness assessment.

Abstract: Neural network pruning has emerged as a promising approach for deploying LLMs in low-resource scenarios while preserving downstream task performance. However, for the first time, we reveal that such pruning disrupts LLMs’ internal activation features crucial for lie detection, where probing classifiers (typically small logistic regression models) trained on these features assess the truthfulness of LLM-generated statements. This discovery raises a crucial open question: how can we prune LLMs without sacrificing these critical lie detection capabilities? Our investigation further reveals that naively adjusting layer-wise pruning sparsity based on importance inadvertently removes crucial weights, failing to improve lie detection performance despite its reliance on the most crucial LLM layer. To address this issue, we propose Truthful Pruning aligned by Layer-wise Outliers (TPLO), which places greater emphasis on layers with more activation outliers and stronger discriminative features simultaneously. This preserves LLMs’ original performance while retaining critical features of inner states needed for robust lie detection. Moreover, we introduce a prompting rule to enrich the TruthfulQA benchmark for better calibrating LLM pruning. Empirical results show that our approach improves the hallucination detection for pruned LLMs (achieving 88% accuracy at 50% sparsity) and enhances their performance on TruthfulQA.

[761] Progressive Element-wise Gradient Estimation for Neural Network Quantization

Kaiqi Zhao

Main category: cs.LG

TL;DR: PEGE is a novel gradient estimation method that improves neural network quantization by progressively replacing full-precision values with quantized ones and co-optimizing task loss with discretization error, outperforming traditional STE methods especially at low bit-widths.

DetailsMotivation: Traditional STE-based quantization methods overlook discretization errors between continuous and quantized values, leading to accuracy degradation, particularly at extremely low bit-widths where these errors become more significant.

Method: Progressive Element-wise Gradient Estimation (PEGE) uses a logarithmic curriculum-driven mixed-precision replacement strategy to gradually replace full-precision weights/activations with quantized counterparts, formulating QAT as a co-optimization problem that minimizes both task loss and discretization error.

Result: Extensive experiments on CIFAR-10 and ImageNet across various architectures (ResNet, VGG) show PEGE consistently outperforms existing backpropagation methods and enables low-precision models to match or even surpass the accuracy of full-precision counterparts.

Conclusion: PEGE provides a unified and generalizable framework for neural network quantization that effectively addresses discretization errors, making it particularly valuable for deploying deep neural networks on resource-constrained hardware with minimal accuracy loss.

Abstract: Neural network quantization aims to reduce the bit-widths of weights and activations, making it a critical technique for deploying deep neural networks on resource-constrained hardware. Most Quantization-Aware Training (QAT) methods rely on the Straight-Through Estimator (STE) to address the non-differentiability of discretization functions by replacing their derivatives with that of the identity function. While effective, STE overlooks discretization errors between continuous and quantized values, which can lead to accuracy degradation – especially at extremely low bit-widths. In this paper, we propose Progressive Element-wise Gradient Estimation (PEGE), a simple yet effective alternative to STE, which can be seamlessly integrated with any forward propagation methods and improves the quantized model accuracy. PEGE progressively replaces full-precision weights and activations with their quantized counterparts via a novel logarithmic curriculum-driven mixed-precision replacement strategy. Then it formulates QAT as a co-optimization problem that simultaneously minimizes the task loss for prediction and the discretization error for quantization, providing a unified and generalizable framework. Extensive experiments on CIFAR-10 and ImageNet across various architectures (e.g., ResNet, VGG) demonstrate that PEGE consistently outperforms existing backpropagation methods and enables low-precision models to match or even outperform the accuracy of their full-precision counterparts.

[762] LLM-QUBO: An End-to-End Framework for Automated QUBO Transformation from Natural Language Problem Descriptions

Huixiang Zhang, Mahzabeen Emu, Salimur Choudhury

Main category: cs.LG

TL;DR: LLM-QUBO is an end-to-end framework that automates quantum annealing for optimization problems by using LLMs to generate QUBO formulations from natural language and employing hybrid quantum-classical decomposition for scalability.

DetailsMotivation: To overcome the manual complexity of translating problems to QUBO format and address quantum hardware scalability limitations for practical quantum annealing applications.

Method: Uses LLM to parse natural language into mathematical representations, integrates hybrid Benders’ decomposition to partition problems into QUBO master problems (for quantum) and linear sub-problems (for classical solvers).

Result: Validated correctness of generated QUBO formulations and demonstrated scalability through classical solvers, establishing performance baseline ready for quantum hardware.

Conclusion: Provides an automated workflow that bridges classical AI and quantum computing, significantly reducing barriers to practical quantum optimization applications.

Abstract: Quantum annealing offers a promising paradigm for solving NP-hard combinatorial optimization problems, but its practical application is severely hindered by two challenges: the complex, manual process of translating problem descriptions into the requisite Quadratic Unconstrained Binary Optimization (QUBO) format and the scalability limitations of current quantum hardware. To address these obstacles, we propose a novel end-to-end framework, LLM-QUBO, that automates this entire formulation-to-solution pipeline. Our system leverages a Large Language Model (LLM) to parse natural language, automatically generating a structured mathematical representation. To overcome hardware limitations, we integrate a hybrid quantum-classical Benders’ decomposition method. This approach partitions the problem, compiling the combinatorial complex master problem into a compact QUBO format, while delegating linearly structured sub-problems to classical solvers. The correctness of the generated QUBO and the scalability of the hybrid approach are validated using classical solvers, establishing a robust performance baseline and demonstrating the framework’s readiness for quantum hardware. Our primary contribution is a synergistic computing paradigm that bridges classical AI and quantum computing, addressing key challenges in the practical application of optimization problem. This automated workflow significantly reduces the barrier to entry, providing a viable pathway to transform quantum devices into accessible accelerators for large-scale, real-world optimization challenges.

[763] Exploiting a Mixture-of-Layers in an Electrocardiography Foundation Model

Phu X. Nguyen, Huy Phan, Hieu Pham, Christos Chatzichristos, Bert Vandenberk, Maarten De Vos

Main category: cs.LG

TL;DR: PMA method aggregates multiple Transformer layer representations instead of just using final layer, improving ECG foundation model performance

DetailsMotivation: Final layer of pre-trained Transformer models may not provide optimal performance for downstream ECG tasks, and layer-wise representations are underutilized

Method: Post-pretraining Mixture-of-layers Aggregation (PMA) with gating network to selectively fuse representations from all layers of 1D Vision Transformer pre-trained via masked modeling

Result: Enhanced representation power and improved performance in downstream ECG applications compared to using only final layer

Conclusion: Leveraging representation diversity across Transformer layers through selective aggregation outperforms traditional single-layer approaches for ECG analysis

Abstract: Transformer-based foundation models for Electrocardiograms (ECGs) have recently achieved impressive performance in many downstream applications. However, the internal representations of such models across layers have not been fully understood and exploited. An important question arises: Does the final layer of the pre-trained Transformer model, the \emph{de facto} representational layer, provide optimal performance for downstream tasks? Although our answer based on empirical and theoretical analyses for this question is negative, we propose a novel approach to leverage the representation diversity of the model’s layers effectively. Specifically, we introduce a novel architecture called Post-pretraining Mixture-of-layers Aggregation (PMA), which enables a flexible combination of the layer-wise representations from the layer stack of a Transformer-based foundation model. We first pre-train the model from ECG signals using the 1-dimensional Vision Transformer (ViT) via masked modeling. In downstream applications, instead of relying solely on the last layer of the model, we employ a gating network to selectively fuse the representations from the pretrained model’s layers, thereby enhancing representation power and improving performance of the downstream applications. In addition, we extend the proposed method to the pretraining stage by aggregating all representations through group-wise averaging before feeding them into the decoder-based Transformer.

[764] Pre-trained knowledge elevates large language models beyond traditional chemical reaction optimizers

Robert MacKnight, Jose Emilio Regio, Jeffrey G. Ethier, Luke A. Baldwin, Gabe Gomes

Main category: cs.LG

TL;DR: LLM-guided optimization matches or exceeds Bayesian optimization performance in chemical reaction optimization, particularly in complex categorical spaces where high-performing conditions are scarce, while maintaining higher exploration diversity.

DetailsMotivation: To demonstrate that pre-trained knowledge in large language models fundamentally changes the paradigm of black-box optimization in experimental chemistry by enabling more effective navigation of chemical parameter spaces.

Method: Benchmarked LLM-guided optimization against Bayesian optimization and random sampling across six fully enumerated categorical reaction datasets (768-5,684 experiments), using a topology-agnostic information theory framework to quantify sampling diversity.

Result: Frontier LLMs consistently matched or exceeded BO performance across five single-objective datasets, with advantages growing as parameter complexity increases and high-performing conditions become scarce (<5% of space). LLMs maintained systematically higher exploration entropy than BO while achieving superior performance.

Conclusion: LLM-guided optimization excels where traditional methods struggle - complex categorical spaces requiring domain understanding rather than mathematical optimization, with pre-trained domain knowledge enabling more effective navigation rather than replacing structured exploration strategies.

Abstract: Modern optimization in experimental chemistry employs algorithmic search through black-box parameter spaces. Here we demonstrate that pre-trained knowledge in large language models (LLMs) fundamentally changes this paradigm. Using six fully enumerated categorical reaction datasets (768 - 5,684 experiments), we benchmark LLM-guided optimization (LLM-GO) against Bayesian optimization (BO) and random sampling. Frontier LLMs consistently match or exceed BO performance across five single-objective datasets, with advantages growing as parameter complexity increases and high-performing conditions become scarce (<5% of space). BO retains superiority only for explicit multi-objective trade-offs. To understand these contrasting behaviors, we introduce a topology-agnostic information theory framework quantifying sampling diversity throughout optimization campaigns. This analysis reveals that LLMs maintain systematically higher exploration entropy than BO across all datasets while achieving superior performance, with advantages most pronounced in solution-scarce parameter spaces where high-entropy exploration typically fails

  • suggesting that pre-trained domain knowledge enables more effective navigation of chemical parameter space rather than replacing structured exploration strategies. To enable transparent benchmarking and community validation, we release Iron Mind (https://gomes.andrew.cmu.edu/iron-mind), a no-code platform for side-by-side evaluation of human, algorithmic, and LLM optimization campaigns with public leaderboards and complete trajectories. Our findings establish that LLM-GO excels precisely where traditional methods struggle: complex categorical spaces requiring domain understanding rather than mathematical optimization.

[765] Principled Approximation Methods for Efficient and Scalable Deep Learning

Pedro Savarese

Main category: cs.LG

TL;DR: This thesis develops principled approximation methods to improve deep learning efficiency through model compression, architecture design, and optimization techniques, addressing computational and energy demands of large models.

DetailsMotivation: The computational and energy demands of increasingly larger deep learning models create significant barriers to deployment and wider adoption, necessitating more efficient approaches.

Method: Three main approaches: 1) Novel approximations for pruning and quantization that frame discrete problems as continuous and differentiable; 2) Neural architecture search with parameter sharing across layers; 3) Adaptive optimization methods with improved hyperparameter tuning.

Result: Experimental results on image classification, language modeling, and generative modeling show significant improvements in training and inference efficiency while maintaining or improving model performance.

Conclusion: The proposed scalable and principled approximations effectively tackle computationally hard problems in deep learning, enabling more efficient models without sacrificing performance.

Abstract: Recent progress in deep learning has been driven by increasingly larger models. However, their computational and energy demands have grown proportionally, creating significant barriers to their deployment and to a wider adoption of deep learning technologies. This thesis investigates principled approximation methods for improving the efficiency of deep learning systems, with a particular focus on settings that involve discrete constraints and non-differentiability. We study three main approaches toward improved efficiency: architecture design, model compression, and optimization. For model compression, we propose novel approximations for pruning and quantization that frame the underlying discrete problem as continuous and differentiable, enabling gradient-based training of compression schemes alongside the model’s parameters. These approximations allow for fine-grained sparsity and precision configurations, leading to highly compact models without significant fine-tuning. In the context of architecture design, we design an algorithm for neural architecture search that leverages parameter sharing across layers to efficiently explore implicitly recurrent architectures. Finally, we study adaptive optimization, revisiting theoretical properties of widely used methods and proposing an adaptive optimizer that allows for quick hyperparameter tuning. Our contributions center on tackling computationally hard problems via scalable and principled approximations. Experimental results on image classification, language modeling, and generative modeling tasks show that the proposed methods provide significant improvements in terms of training and inference efficiency while maintaining, or even improving, the model’s performance.

[766] FNODE: Flow-Matching for data-driven simulation of constrained multibody systems

Hongyu Wang, Jingquan Wang, Dan Negrut

Main category: cs.LG

TL;DR: FNODE is a novel neural ODE framework that learns acceleration vector fields directly from trajectory data, eliminating backpropagation through ODE solvers and achieving superior performance on multibody systems.

DetailsMotivation: Address high computational cost and limited long-term prediction accuracy in data-driven modeling of constrained multibody systems.

Method: Flow-Matching Neural ODE (FNODE) learns acceleration vector fields directly, uses numerical differentiation (FFT+FD hybrid scheme) for acceleration targets, and avoids backpropagation through ODE solvers.

Result: Outperforms MBD-NODE, LSTM, and FCNN across multiple benchmarks (mass-spring-damper, double pendulum, slider-crank, cart-pole) with good accuracy, generalization, and computational efficiency.

Conclusion: FNODE provides an effective framework for constrained multibody system modeling with improved computational efficiency and prediction accuracy compared to existing approaches.

Abstract: Data-driven modeling of constrained multibody systems faces two persistent challenges: high computational cost and limited long-term prediction accuracy. To address these issues, we introduce the Flow-Matching Neural Ordinary Differential Equation (FNODE), a framework that learns acceleration vector fields directly from trajectory data. By reformulating the training objective to supervise accelerations rather than integrated states, FNODE eliminates the need for backpropagation through an ODE solver, which represents a bottleneck in traditional Neural ODEs. Acceleration targets are computed efficiently using numerical differentiation techniques, including a hybrid Fast Fourier Transform (FFT) and Finite Difference (FD) scheme. We evaluate FNODE on a diverse set of benchmarks, including the single and triple mass-spring-damper systems, double pendulum, slider-crank, and cart-pole. Across all cases, FNODE consistently outperforms existing approaches such as Multi-Body Dynamic Neural ODE (MBD-NODE), Long Short-Term Memory (LSTM) networks, and Fully Connected Neural Networks (FCNN), demonstrating good accuracy, generalization, and computational efficiency.

[767] Democratizing Agentic AI with Fast Test-Time Scaling on the Edge

Hao Mark Chen, Zhiwen Mo, Guanxi Lu, Shuang Liang, Lingxiao Ma, Wayne Luk, Hongxiang Fan

Main category: cs.LG

TL;DR: FlashTTS is a serving system that enables efficient Test-Time Scaling for edge LLMs, achieving cloud-level accuracy on consumer GPUs through optimized memory management and scheduling.

DetailsMotivation: Edge devices need agentic AI for privacy and responsiveness, but memory constraints force the use of smaller LLMs with inferior reasoning. Existing Test-Time Scaling methods have prohibitive overhead on edge hardware.

Method: FlashTTS introduces three optimizations: Speculative Beam Extension to handle irregular reasoning paths, Asymmetric Multi-Model Memory Allocation for dynamic memory balancing, and Dynamic Prefix-Aware Scheduling for KV-cache reuse.

Result: FlashTTS enables edge LLMs on a single 24GB consumer GPU to match cloud model accuracy and latency, achieving 2.2x higher goodput and 38%-68% latency reduction compared to vLLM baseline.

Conclusion: FlashTTS makes Test-Time Scaling practical for memory-constrained edge devices, paving the way for democratized, high-performance agentic AI deployment on edge hardware.

Abstract: Deploying agentic AI on edge devices is crucial for privacy and responsiveness, but memory constraints typically relegate these systems to smaller Large Language Models (LLMs) with inferior reasoning capabilities. Test-Time Scaling (TTS) can bridge this reasoning gap by dedicating more compute during inference, but existing methods incur prohibitive overhead on edge hardware. To overcome this, we introduce FlashTTS, a serving system that makes TTS practical for memory-constrained LLM reasoning. FlashTTS introduces three synergistic optimizations: (i) Speculative Beam Extension to mitigate system stragglers from irregular reasoning paths; (ii) Asymmetric Multi-Model Memory Allocation to dynamically balance memory between generation and verification; and (iii) Dynamic Prefix-Aware Scheduling to maximize KV-cache reuse. Built as a plug-and-play library for vLLM, FlashTTS enables edge LLMs on a single consumer GPU (24 GB) to match the accuracy and latency of large cloud models. Our evaluation demonstrates that FlashTTS achieves an average 2.2x higher goodput and reduces latency by 38%-68% compared to a vLLM baseline, paving the way for democratized, high-performance agentic AI on edge devices.

[768] Decentralized Transformers with Centralized Aggregation are Sample-Efficient Multi-Agent World Models

Yang Zhang, Chenjia Bai, Bin Zhao, Junchi Yan, Xiu Li, Xuelong Li

Main category: cs.LG

TL;DR: A Transformer-based world model for multi-agent RL that combines decentralized local dynamics with centralized representation aggregation to address scalability and non-stationarity challenges in MARL.

DetailsMotivation: To improve sample efficiency in multi-agent RL by building a world model that can handle scalability issues of centralized architectures and non-stationarity issues of decentralized architectures.

Method: Proposes a novel world model using Transformer architecture for auto-regressive sequence modeling of discrete tokens to learn decentralized local dynamics, combined with a Perceiver Transformer for centralized representation aggregation from all agents.

Result: Outperforms strong model-free approaches and existing model-based methods in both sample efficiency and overall performance on Starcraft Multi-Agent Challenge (SMAC).

Conclusion: The proposed Transformer-based world model effectively addresses MARL challenges by combining decentralized dynamics learning with centralized representation, demonstrating superior performance and sample efficiency.

Abstract: Learning a world model for model-free Reinforcement Learning (RL) agents can significantly improve the sample efficiency by learning policies in imagination. However, building a world model for Multi-Agent RL (MARL) can be particularly challenging due to the scalability issue in a centralized architecture arising from a large number of agents, and also the non-stationarity issue in a decentralized architecture stemming from the inter-dependency among agents. To address both challenges, we propose a novel world model for MARL that learns decentralized local dynamics for scalability, combined with a centralized representation aggregation from all agents. We cast the dynamics learning as an auto-regressive sequence modeling problem over discrete tokens by leveraging the expressive Transformer architecture, in order to model complex local dynamics across different agents and provide accurate and consistent long-term imaginations. As the first pioneering Transformer-based world model for multi-agent systems, we introduce a Perceiver Transformer as an effective solution to enable centralized representation aggregation within this context. Results on Starcraft Multi-Agent Challenge (SMAC) show that it outperforms strong model-free approaches and existing model-based methods in both sample efficiency and overall performance.

[769] From TLinFormer to TConstFormer: The Leap to Constant-Time Transformer Attention: Achieving O(1) Computation and O(1) KV Cache during Autoregressive Inference

Zhongpan Tang

Main category: cs.LG

TL;DR: TConstFormer introduces a periodic state update mechanism to achieve constant-size KV cache and O(1) computational complexity for Transformer inference on long sequences.

DetailsMotivation: To overcome the limitations of autoregressive Transformer inference which suffers from linearly growing KV Cache and O(N²d) computational complexity that hinders processing ultra-long sequences.

Method: Uses an innovative periodic state update mechanism that performs constant-time computations for k-1 consecutive steps and executes a single linear-time global information synchronization only on the k-th step (e.g., k=256).

Result: Demonstrates overwhelming advantage over baseline models in speed, memory efficiency, and overall performance on long-text inference tasks.

Conclusion: This breakthrough enables efficient and robust streaming language model applications by achieving truly constant-size O(1) KV Cache.

Abstract: Although the Transformer has become the cornerstone of modern AI, its autoregressive inference suffers from a linearly growing KV Cache and a computational complexity of O(N^2 d), severely hindering its ability to process ultra-long sequences. To overcome this limitation, this paper introduces the TConstFormer architecture, building upon our previous work, TLinFormer. TConstFormer employs an innovative periodic state update mechanism to achieve a truly constant-size O(1) KV Cache. The computational complexity of this mechanism is also O(1) in an amortized sense: it performs purely constant-time computations for $k-1$ consecutive steps (e.g., $k=256$) and executes a single linear-time global information synchronization only on the $k$-th step. Theoretical calculations and experimental results demonstrate that TConstFormer exhibits an overwhelming advantage over baseline models in terms of speed, memory efficiency, and overall performance on long-text inference tasks. This breakthrough paves the way for efficient and robust streaming language model applications.

[770] Estimating Parameter Fields in Multi-Physics PDEs from Scarce Measurements

Xuyang Li, Mahdi Masmoudi, Rami Gharbi, Nizar Lajnef, Vishnu Naresh Boddeti

Main category: cs.LG

TL;DR: Neptune is a novel method that uses independent coordinate neural networks to accurately infer spatiotemporal parameter fields from sparse measurements, significantly outperforming existing approaches like PINNs in complex PDE systems.

DetailsMotivation: Existing parameter estimation methods struggle with nonlinear, spatiotemporal parameter variations in PDE systems, especially with limited observations, multiphysics interactions, and nonlinear dynamics.

Method: Employs independent coordinate neural networks to continuously represent each parameter field in physical space or state variables, enabling robust inference from sparse measurements.

Result: Achieves robust parameter estimation from as few as 50 observations, reduces parameter errors by two orders of magnitude and dynamic response errors by 10x compared to PINNs, with superior extrapolation capabilities.

Conclusion: Neptune enables reliable and data-efficient parameter inference for complex PDE systems, promising transformative impacts across engineering, healthcare, and physics applications.

Abstract: Parameterized partial differential equations (PDEs) underpin the mathematical modeling of complex systems in diverse domains, including engineering, healthcare, and physics. A central challenge in using PDEs for real-world applications is to accurately infer the parameters, particularly when the parameters exhibit non-linear and spatiotemporal variations. Existing parameter estimation methods, such as sparse identification and physics-informed neural networks (PINNs), struggle in such cases, especially with nonlinear dynamics, multiphysics interactions, or limited observations of the system response. To address these challenges, we introduce Neptune, a general-purpose method capable of inferring parameter fields from sparse measurements of system responses. Neptune employs independent coordinate neural networks to continuously represent each parameter field in physical space or in state variables. Across various physical and biomedical problems, where direct parameter measurements are prohibitively expensive or unattainable, Neptune significantly outperforms existing methods, achieving robust parameter estimation from as few as 50 observations, reducing parameter estimation errors by two orders of magnitude and dynamic response prediction errors by a factor of ten compared to PINNs. Furthermore, Neptune exhibits superior extrapolation capabilities, enabling accurate predictions in regimes beyond training data where PINN fail. By facilitating reliable and data-efficient parameter inference, Neptune promises broad transformative impacts in engineering, healthcare, and beyond.

[771] Agent Trading Arena: A Study on Numerical Understanding in LLM-Based Agents

Tianmi Ma, Jiawei Du, Wenxin Huang, Wenjie Wang, Liang Xie, Xian Zhong, Joey Tianyi Zhou

Main category: cs.LG

TL;DR: Agent Trading Arena: A virtual stock market platform where LLM-based agents compete in trading with realistic price impact, showing that visual chart inputs and reflection modules significantly improve trading performance over text-only approaches.

DetailsMotivation: Existing LLM trading approaches are limited to historical backtesting where agents cannot influence market prices and train on static data, creating a gap between training and real-world financial environments.

Method: Created a virtual zero-sum stock market platform with competitive multi-agent trading where LLM agents can directly impact price dynamics through realistic bid-ask interactions, and tested both text-based and chart-based visual inputs with reflection modules.

Result: LLMs struggle with numerical reasoning using plain-text data, often overfitting to local patterns. Chart visualizations significantly enhance numerical reasoning and trading performance. Reflection modules provide additional improvements, especially with visual inputs. Superior performance demonstrated on NASDAQ and CSI datasets, particularly under high volatility.

Conclusion: The Agent Trading Arena successfully bridges the gap between training and real markets, showing that visual inputs and reflection mechanisms are crucial for improving LLM trading performance in dynamic financial environments.

Abstract: Large language models (LLMs) have demonstrated remarkable capabilities in natural language tasks, yet their performance in dynamic, real-world financial environments remains underexplored. Existing approaches are limited to historical backtesting, where trading actions cannot influence market prices and agents train only on static data. To address this limitation, we present the Agent Trading Arena, a virtual zero-sum stock market in which LLM-based agents engage in competitive multi-agent trading and directly impact price dynamics. By simulating realistic bid-ask interactions, our platform enables training in scenarios that closely mirror live markets, thereby narrowing the gap between training and evaluation. Experiments reveal that LLMs struggle with numerical reasoning when given plain-text data, often overfitting to local patterns and recent values. In contrast, chart-based visualizations significantly enhance both numerical reasoning and trading performance. Furthermore, incorporating a reflection module yields additional improvements, especially with visual inputs. Evaluations on NASDAQ and CSI datasets demonstrate the superiority of our method, particularly under high volatility. All code and data are available at https://github.com/wekjsdvnm/Agent-Trading-Arena.

[772] Learning to Shard: RL for Co-optimizing the Parallelism Degrees and Per-operator Sharding Dimensions in Distributed LLM Inference

Ruokai Yin, Sattwik Deb Mishra, Xuan Zuo, Hokchhay Tann, Preyas Shah, Apala Guha

Main category: cs.LG

TL;DR: Learn to Shard is an RL-based approach that co-optimizes parallelism degrees and per-operator sharding dimensions for distributed LLM inference, achieving up to 3.5x throughput improvement over baseline methods.

DetailsMotivation: Current systems like Megatron-LM use static heuristics that separately configure parallelism strategies, leaving significant performance untapped as models scale and hardware topologies diversify.

Method: Uses reinforcement learning with an attention-based policy over an elite history to learn from high-performing strategies and efficiently navigate the combinatorial search space.

Result: Achieves up to 3.5x throughput improvement over metaheuristic baselines and 1.06x over Megatron heuristics on H100 clusters with MoE models up to 1.6T parameters.

Conclusion: The RL-based co-optimization approach significantly outperforms existing static heuristics for distributed LLM inference, demonstrating the value of learned strategies over hand-crafted rules.

Abstract: Distributed LLM inference requires careful coordination of parallelization strategies across hundreds to thousands of NPUs to meet production SLOs. Current systems like Megatron-LM rely on static heuristics that separately configure parallelism degrees and per-operator sharding dimensions, leaving significant performance on the table as models scale and hardware topologies diversify. We introduce Learn to Shard, to our knowledge, the first RL-based approach to co-optimize both coarse-grained parallelism degrees and fine-grained per-operator sharding dimensions for distributed LLM inference. Our method employs an attention-based policy over an elite history that learns from high-performing strategies to efficiently navigate the vast combinatorial search space. Evaluated on H100 clusters with MoE models up to 1.6T parameters, Learn to Shard achieves up to 3.5x throughput improvement over metaheuristic baselines and 1.06x over Megatron heuristics.

[773] The challenge of hidden gifts in multi-agent reinforcement learning

Dane Malenfant, Blake A. Richards

Main category: cs.LG

TL;DR: The paper studies ‘hidden gifts’ in MARL - beneficial actions by others that are unobservable. Using a simple grid-world task where agents must share a key to get collective rewards, it shows state-of-the-art MARL algorithms fail. Independent policy gradient agents succeed with action history, and a variance-reducing correction term helps them converge reliably.

DetailsMotivation: To understand how 'hidden gifts' - beneficial actions by others that are unobservable - pose challenges for credit assignment in multi-agent reinforcement learning, and to develop solutions for these scenarios.

Method: A simple grid-world MARL task where agents must unlock individual doors with a shared key and drop it for others to obtain collective rewards. Tested various state-of-the-art RL algorithms including MARL approaches, and developed a correction term for independent agents inspired by learning-aware approaches.

Result: Standard MARL algorithms failed to solve the task. Independent model-free policy gradient agents succeeded when provided with action history information. A derived correction term reduced learning variance and improved convergence to collective success.

Conclusion: Credit assignment is particularly challenging with hidden gifts in multi-agent settings. Learning awareness in independent agents can benefit these scenarios, and the proposed correction term helps independent agents converge more reliably to cooperative solutions.

Abstract: Sometimes we benefit from actions that others have taken even when we are unaware that they took those actions. For example, if your neighbor chooses not to take a parking spot in front of your house when you are not there, you can benefit, even without being aware that they took this action. These “hidden gifts” represent an interesting challenge for multi-agent reinforcement learning (MARL), since assigning credit when the beneficial actions of others are hidden is non-trivial. Here, we study the impact of hidden gifts with a very simple MARL task. In this task, agents in a grid-world environment have individual doors to unlock in order to obtain individual rewards. As well, if all the agents unlock their door the group receives a larger collective reward. However, there is only one key for all of the doors, such that the collective reward can only be obtained when the agents drop the key for others after they use it. Notably, there is nothing to indicate to an agent that the other agents have dropped the key, thus the act of dropping the key for others is a “hidden gift”. We show that several different state-of-the-art RL algorithms, including MARL algorithms, fail to learn how to obtain the collective reward in this simple task. Interestingly, we find that independent model-free policy gradient agents can solve the task when we provide them with information about their own action history, but MARL agents still cannot solve the task with action history. Finally, we derive a correction term for these independent agents, inspired by learning aware approaches, which reduces the variance in learning and helps them to converge to collective success more reliably. These results show that credit assignment in multi-agent settings can be particularly challenging in the presence of “hidden gifts”, and demonstrate that learning awareness in independent agents can benefit these settings.

[774] Quantum-Optimized Selective State Space Model for Efficient Time Series Prediction

Stefan-Alexandru Jura, Mihai Udrescu, Alexandru Topirceanu

Main category: cs.LG

TL;DR: Q-SSM is a quantum-optimized state space model that uses variational quantum gates to improve long-range time series forecasting, outperforming traditional Transformers and state space models with better stability and efficiency.

DetailsMotivation: Address limitations of Transformer-based models (quadratic complexity, degraded long-horizon performance) and state space models (unstable training, sensitivity to initialization) in long-range time series forecasting.

Method: Hybrid quantum-optimized approach integrating state space dynamics with variational quantum gate (RY-RX ansatz) that regulates memory updates adaptively, replacing expensive attention mechanisms with lightweight quantum circuits.

Result: Q-SSM consistently outperforms strong baselines (LSTM, TCN, Reformer), Transformer-based models, and S-Mamba on ETT, Traffic, and Exchange Rate benchmarks.

Conclusion: Variational quantum gating effectively addresses current limitations in long-range forecasting, providing accurate and robust multivariate predictions with improved convergence stability and long-term dependency modeling.

Abstract: Long-range time series forecasting remains challenging, as it requires capturing non-stationary and multi-scale temporal dependencies while maintaining noise robustness, efficiency, and stability. Transformer-based architectures such as Autoformer and Informer improve generalization but suffer from quadratic complexity and degraded performance on very long time horizons. State space models, notably S-Mamba, provide linear-time updates but often face unstable training dynamics, sensitivity to initialization, and limited robustness for multivariate forecasting. To address such challenges, we propose the Quantum-Optimized Selective State Space Model (Q-SSM), a hybrid quantum-optimized approach that integrates state space dynamics with a variational quantum gate. Instead of relying on expensive attention mechanisms, Q-SSM employs a simple parametrized quantum circuit (RY-RX ansatz) whose expectation values regulate memory updates adaptively. This quantum gating mechanism improves convergence stability, enhances the modeling of long-term dependencies, and provides a lightweight alternative to attention. We empirically validate Q-SSM on three widely used benchmarks, i.e., ETT, Traffic, and Exchange Rate. Results show that Q-SSM consistently improves over strong baselines (LSTM, TCN, Reformer), Transformer-based models, and S-Mamba. These findings demonstrate that variational quantum gating can address current limitations in long-range forecasting, leading to accurate and robust multivariate predictions.

[775] ReLATE: Learning Efficient Sparse Encoding for High-Performance Tensor Decomposition

Ahmed E. Helal, Fabio Checconi, Jan Laukemann, Yongseok Soh, Jesmin Jahan Tithi, Fabrizio Petrini, Jee Choi

Main category: cs.LG

TL;DR: ReLATE is a reinforcement learning framework that automatically generates optimized sparse tensor encodings for tensor decomposition, achieving up to 2x speedup over expert-designed formats.

DetailsMotivation: Traditional expert-designed sparse tensor formats fail to adapt to irregular tensor shapes and variable data distributions, creating performance bottlenecks for tensor decomposition on modern processors.

Method: Uses reinforcement learning with an autonomous agent that discovers efficient tensor encodings through interaction with the TD environment, employing hybrid model-free/model-based learning and incorporating rule-driven action masking and dynamics-informed filtering.

Result: Achieves up to 2x speedup compared to best sparse formats, with geometric-mean speedup of 1.4-1.46x across diverse sparse tensor datasets.

Conclusion: ReLATE demonstrates that learning-based approaches can automatically generate superior sparse tensor representations that outperform expert-designed formats while ensuring functional correctness and bounded execution time.

Abstract: Tensor decomposition (TD) is essential for analyzing high-dimensional sparse data, yet its irregular computations and memory-access patterns pose major performance challenges on modern parallel processors. Prior works rely on expert-designed sparse tensor formats that fail to adapt to irregular tensor shapes and/or highly variable data distributions. We present the reinforcement-learned adaptive tensor encoding (ReLATE) framework, a novel learning-augmented method that automatically constructs efficient sparse tensor representations without labeled training samples. ReLATE employs an autonomous agent that discovers optimized tensor encodings through direct interaction with the TD environment, leveraging a hybrid model-free and model-based algorithm to learn from both real and imagined actions. Moreover, ReLATE introduces rule-driven action masking and dynamics-informed action filtering mechanisms that ensure functionally correct tensor encoding with bounded execution time, even during early learning stages. By automatically adapting to both irregular tensor shapes and data distributions, ReLATE generates sparse tensor representations that consistently outperform expert-designed formats across diverse sparse tensor data sets, achieving up to 2X speedup compared to the best sparse format, with a geometric-mean speedup of 1.4-1.46X.

[776] Continuously Tempered Diffusion Samplers

Ezra Erives, Bowen Jing, Peter Holderrieth, Tommi Jaakkola

Main category: cs.LG

TL;DR: Proposes continuously tempered diffusion samplers that use temperature-based exploration to improve neural sampling from unnormalized distributions, addressing insufficient exploration issues in previous methods.

DetailsMotivation: Previous annealing-based neural samplers suffer from insufficient exploration due to isolated modes and pathological properties in the annealing path, leading to poor performance after training.

Method: Introduces a family of distributions across different temperatures to lower energy barriers at higher temperatures and drive exploration at the target temperature, leveraging techniques from molecular dynamics.

Result: Empirical validation shows improved sampler performance driven by extended exploration capabilities compared to previous methods.

Conclusion: Continuously tempered diffusion samplers effectively address exploration limitations in neural sampling by incorporating temperature-based exploration strategies, leading to better performance.

Abstract: Annealing-based neural samplers seek to amortize sampling from unnormalized distributions by training neural networks to transport a family of densities interpolating from source to target. A crucial design choice in the training phase of such samplers is the proposal distribution by which locations are generated at which to evaluate the loss. Previous work has obtained such a proposal distribution by combining a partially learned transport with annealed Langevin dynamics. However, isolated modes and other pathological properties of the annealing path imply that such proposals achieve insufficient exploration and thereby lower performance post training. To remedy this, we propose continuously tempered diffusion samplers, which leverage exploration techniques developed in the context of molecular dynamics to improve proposal distributions. Specifically, a family of distributions across different temperatures is introduced to lower energy barriers at higher temperatures and drive exploration at the lower temperature of interest. We empirically validate improved sampler performance driven by extended exploration. Code is available at https://github.com/eje24/ctds.

[777] Chunked TabPFN: Exact Training-Free In-Context Learning for Long-Context Tabular Data

Renat Sergazinov, Shao-An Yin

Main category: cs.LG

TL;DR: TabPFN v2 outperforms tree-based models on tabular benchmarks but has 10K token limit. This paper introduces tiled-block attention strategy to handle long contexts without pre-processing, enabling TabPFN to process more data efficiently.

DetailsMotivation: TabPFN v2 shows superior performance over traditional tree-based models for tabular data but is limited by transformer's quadratic computation costs that restrict context to 10K tokens. Existing compression methods require pre-processing, so a more efficient solution is needed.

Method: The authors introduce a tiled-block strategy to compute attention within the TabPFN framework. This approach is compatible with standard GPU setups and eliminates the need for pre-processing while enabling long context processing.

Result: The method successfully enables TabPFN to process long contexts without any pre-processing requirements. The effectiveness is demonstrated on the standard TabArena benchmark.

Conclusion: The tiled-block attention strategy provides an efficient solution to TabPFN’s context limitation problem, making it practical for handling larger tabular datasets without compromising performance or requiring additional pre-processing steps.

Abstract: TabPFN v2 achieves better results than tree-based models on several tabular benchmarks, which is notable since tree-based models are usually the strongest choice for tabular data. However, it cannot handle more than 10K context tokens because transformers have quadratic computation and memory costs. Unlike existing approaches that rely on context compression, such as selecting representative samples via K-nearest neighbors (KNN), we introduce a \textbf{tiled-block} strategy to compute attention within the TabPFN framework. This design is compatible with standard GPU setups and, to the best of our knowledge, is the first to enable TabPFN to \textbf{process long contexts without any pre-processing}. We demonstrate the effectiveness of our approach on the standard TabArena benchmark.

[778] Counterfactual Risk Minimization with IPS-Weighted BPR and Self-Normalized Evaluation in Recommender Systems

Rahul Raja, Arpita Vats

Main category: cs.LG

TL;DR: A pipeline combining IPS-weighted training with BPR objective and propensity regularizer to address exposure bias in recommender systems, reducing variance while maintaining effectiveness.

DetailsMotivation: Learning from logged implicit feedback suffers from exposure bias, and standard IPS correction methods often have high variance and instability issues.

Method: Integrates IPS-weighted training with IPS-weighted Bayesian Personalized Ranking objective augmented by a Propensity Regularizer. Compares DM, IPS, and SNIPS for offline evaluation.

Result: Experiments on synthetic and MovieLens 100K data show better generalization under unbiased exposure and reduced evaluation variance compared to standard methods.

Conclusion: The approach provides practical guidance for counterfactual learning and evaluation in real-world recommendation settings by mitigating variance from extreme propensity weights.

Abstract: Learning and evaluating recommender systems from logged implicit feedback is challenging due to exposure bias. While inverse propensity scoring (IPS) corrects this bias, it often suffers from high variance and instability. In this paper, we present a simple and effective pipeline that integrates IPS-weighted training with an IPS-weighted Bayesian Personalized Ranking (BPR) objective augmented by a Propensity Regularizer (PR). We compare Direct Method (DM), IPS, and Self-Normalized IPS (SNIPS) for offline policy evaluation, and demonstrate how IPS-weighted training improves model robustness under biased exposure. The proposed PR further mitigates variance amplification from extreme propensity weights, leading to more stable estimates. Experiments on synthetic and MovieLens 100K data show that our approach generalizes better under unbiased exposure while reducing evaluation variance compared to naive and standard IPS methods, offering practical guidance for counterfactual learning and evaluation in real-world recommendation settings.

[779] Are We Really Learning the Score Function? Reinterpreting Diffusion Models Through Wasserstein Gradient Flow Matching

An B. Vuong, Michael T. McCann, Javier E. Santos, Yen Ting Lin

Main category: cs.LG

TL;DR: Diffusion models don’t actually learn true score functions as commonly assumed, but still work well because they effectively learn velocity fields for Wasserstein Gradient Flows.

DetailsMotivation: To resolve the paradox that diffusion models perform well despite violating mathematical constraints of true score functions, and to provide a better theoretical framework for understanding diffusion training.

Method: Numerical analysis showing trained diffusion networks violate integral and differential constraints of conservative vector fields, and proposing a new theoretical perspective based on Wasserstein Gradient Flow (WGF) framework.

Result: Demonstrated that neural vector fields in diffusion models are not conservative (not true scores), but models still generate successfully. Showed that non-conservative errors don’t harm density transport.

Conclusion: The WGF perspective provides a more principled and elegant theoretical framework for understanding diffusion models, explaining why they work despite not learning true score functions.

Abstract: Diffusion models are commonly interpreted as learning the score function, i.e., the gradient of the log-density of noisy data. However, this assumption implies that the target of learning is a conservative vector field, which is not enforced by the neural network architectures used in practice. We present numerical evidence that trained diffusion networks violate both integral and differential constraints required of true score functions, demonstrating that the learned vector fields are not conservative. Despite this, the models perform remarkably well as generative mechanisms. To explain this apparent paradox, we advocate a new theoretical perspective: diffusion training is better understood as flow matching to the velocity field of a Wasserstein Gradient Flow (WGF), rather than as score learning for a reverse-time stochastic differential equation. Under this view, the “probability flow” arises naturally from the WGF framework, eliminating the need to invoke reverse-time SDE theory and clarifying why generative sampling remains successful even when the neural vector field is not a true score. We further show that non-conservative errors from neural approximation do not necessarily harm density transport. Our results advocate for adopting the WGF perspective as a principled, elegant, and theoretically grounded framework for understanding diffusion generative models.

[780] Scalable Option Learning in High-Throughput Environments

Mikael Henaff, Scott Fujimoto, Michael Rabbat

Main category: cs.LG

TL;DR: SOL is a highly scalable hierarchical RL algorithm that achieves 25x higher throughput than existing methods, successfully trained on 20B frames in NetHack and validated on MiniHack and Mujoco environments.

DetailsMotivation: To address the challenge of scaling hierarchical reinforcement learning to high-throughput environments and enable effective decision-making over long timescales.

Method: Proposes Scalable Option Learning (SOL), a hierarchical RL algorithm designed for high scalability and throughput.

Result: Achieves 25x higher throughput compared to existing hierarchical methods, successfully trains on 20B frames in NetHack, surpassing flat agents and showing positive scaling trends.

Conclusion: SOL demonstrates general applicability across different environments and represents a significant advancement in scaling hierarchical RL for complex decision-making tasks.

Abstract: Hierarchical reinforcement learning (RL) has the potential to enable effective decision-making over long timescales. Existing approaches, while promising, have yet to realize the benefits of large-scale training. In this work, we identify and solve several key challenges in scaling hierarchical RL to high-throughput environments. We propose Scalable Option Learning (SOL), a highly scalable hierarchical RL algorithm which achieves a 25x higher throughput compared to existing hierarchical methods. We train our hierarchical agents using 20 billion frames of experience on the complex game of NetHack, significantly surpassing flat agents and demonstrating positive scaling trends. We also validate our algorithm on MiniHack and Mujoco environments, showcasing its general applicability. Our code is open sourced at github.com/facebookresearch/sol.

[781] LLM-Driven Policy Diffusion: Enhancing Generalization in Offline Reinforcement Learning

Hanping Zhang, Yuhong Guo

Main category: cs.LG

TL;DR: LLMDPD enhances offline RL generalization using text and trajectory prompts processed by LLMs and transformers to guide policy diffusion for unseen tasks.

DetailsMotivation: Offline RL struggles with generalization due to limited datasets and lack of online environments. Need better methods to adapt to new tasks without expert-designed environments.

Method: Uses LLM for text prompts (task descriptions) and transformer for trajectory prompts. Combines both as conditional inputs to context-aware policy diffusion model.

Result: Outperforms state-of-the-art offline RL methods on unseen tasks, demonstrating improved generalization and adaptability.

Conclusion: LLMDPD effectively addresses offline RL generalization challenges by leveraging multimodal prompts and diffusion models for better task adaptation.

Abstract: Reinforcement Learning (RL) is known for its strong decision-making capabilities and has been widely applied in various real-world scenarios. However, with the increasing availability of offline datasets and the lack of well-designed online environments from human experts, the challenge of generalization in offline RL has become more prominent. Due to the limitations of offline data, RL agents trained solely on collected experiences often struggle to generalize to new tasks or environments. To address this challenge, we propose LLM-Driven Policy Diffusion (LLMDPD), a novel approach that enhances generalization in offline RL using task-specific prompts. Our method incorporates both text-based task descriptions and trajectory prompts to guide policy learning. We leverage a large language model (LLM) to process text-based prompts, utilizing its natural language understanding and extensive knowledge base to provide rich task-relevant context. Simultaneously, we encode trajectory prompts using a transformer model, capturing structured behavioral patterns within the underlying transition dynamics. These prompts serve as conditional inputs to a context-aware policy-level diffusion model, enabling the RL agent to generalize effectively to unseen tasks. Our experimental results demonstrate that LLMDPD outperforms state-of-the-art offline RL methods on unseen tasks, highlighting its effectiveness in improving generalization and adaptability in diverse settings.

[782] Theory Foundation of Physics-Enhanced Residual Learning

Shixiao Liang, Wang Chen, Keke Long, Peng Zhang, Xiaopeng Li, Jintao Ke

Main category: cs.LG

TL;DR: This paper provides theoretical justification for Physics-Enhanced Residual Learning (PERL), explaining its advantages in parameter reduction, faster convergence, and reduced training data requirements through rigorous mathematical proofs and numerical validation.

DetailsMotivation: Previous numerical studies showed PERL has advantages but lacked theoretical explanation. The authors aim to provide rigorous mathematical justification for why PERL reduces parameters, converges faster, and needs fewer training samples.

Method: The study investigates problems with Lipschitz continuity properties, examines relationships between loss function bounds and residual learning structure, and proves theorems explaining PERL’s advantages. Numerical examples in vehicle trajectory prediction validate the theorems.

Result: The theoretical analysis proves PERL’s three advantages. Numerical experiments confirm PERL achieves higher accuracy with significantly fewer training samples compared to pure neural networks, demonstrating practical value in autonomous driving applications.

Conclusion: PERL improves predictive performance while reducing data requirements, making it valuable for real-world applications like autonomous driving where corner case data is costly or hard to obtain.

Abstract: Intensive studies have been conducted in recent years to integrate neural networks with physics models to balance model accuracy and interpretability. One recently proposed approach, named Physics-Enhanced Residual Learning (PERL), is to use learning to estimate the residual between the physics model prediction and the ground truth. Numeral examples suggested that integrating such residual with physics models in PERL has three advantages: (1) a reduction in the number of required neural network parameters; (2) faster convergence rates; and (3) fewer training samples needed for the same computational precision. However, these numerical results lack theoretical justification and cannot be adequately explained. This paper aims to explain these advantages of PERL from a theoretical perspective. We investigate a general class of problems with Lipschitz continuity properties. By examining the relationships between the bounds to the loss function and residual learning structure, this study rigorously proves a set of theorems explaining the three advantages of PERL. Several numerical examples in the context of automated vehicle trajectory prediction are conducted to illustrate the proposed theorems. The results confirm that, even with significantly fewer training samples, PERL consistently achieves higher accuracy than a pure neural network. These results demonstrate the practical value of PERL in real world autonomous driving applications where corner case data are costly or hard to obtain. PERL therefore improves predictive performance while reducing the amount of data required.

[783] Optimized Weight Initialization on the Stiefel Manifold for Deep ReLU Neural Networks

Hyungu Lee, Taehyeong Kim, Hayoung Choi

Main category: cs.LG

TL;DR: A novel orthogonal initialization method optimized for ReLU networks that prevents neuron death and gradient vanishing in deep architectures by preserving scale and calibrating pre-activation statistics through Stiefel manifold optimization.

DetailsMotivation: Existing initialization methods like He, Xavier, and orthogonal initialization fail to properly regulate pre-activation mean and control activation sparsity, especially in very deep ReLU networks, leading to dying ReLU problems and gradient instability.

Method: Develops an orthogonal initialization specifically for ReLU by solving an optimization problem on the Stiefel manifold, deriving closed-form solutions and an efficient sampling scheme to preserve scale and calibrate pre-activation statistics.

Result: Theoretical analysis shows prevention of dying ReLU, slower decay of activation variance, and mitigation of gradient vanishing. Empirical results across multiple datasets (MNIST, Fashion-MNIST, tabular data, few-shot settings) demonstrate superior performance over previous initializations.

Conclusion: The proposed orthogonal initialization method enables stable training in deep ReLU networks by addressing fundamental issues with existing initialization approaches, providing both theoretical guarantees and empirical effectiveness across diverse applications.

Abstract: Stable and efficient training of ReLU networks with large depth is highly sensitive to weight initialization. Improper initialization can cause permanent neuron inactivation dying ReLU and exacerbate gradient instability as network depth increases. Methods such as He, Xavier, and orthogonal initialization preserve variance or promote approximate isometry. However, they do not necessarily regulate the pre-activation mean or control activation sparsity, and their effectiveness often diminishes in very deep architectures. This work introduces an orthogonal initialization specifically optimized for ReLU by solving an optimization problem on the Stiefel manifold, thereby preserving scale and calibrating the pre-activation statistics from the outset. A family of closed-form solutions and an efficient sampling scheme are derived. Theoretical analysis at initialization shows that prevention of the dying ReLU problem, slower decay of activation variance, and mitigation of gradient vanishing, which together stabilize signal and gradient flow in deep architectures. Empirically, across MNIST, Fashion-MNIST, multiple tabular datasets, few-shot settings, and ReLU-family activations, our method outperforms previous initializations and enables stable training in deep networks.

[784] Unifying Adversarial Perturbation for Graph Neural Networks

Jinluan Yang, Ruihao Zhang, Zhengyu Chen, Fei Wu, Kun Kuang

Main category: cs.LG

TL;DR: PerturbEmbedding is a novel adversarial training method that applies perturbations directly to GNN hidden embeddings, providing a unified framework for enhancing robustness and generalization against both random and adversarial attacks.

DetailsMotivation: Current adversarial training methods for GNNs are limited to specific datasets and model types, lacking a unified approach to handle various perturbation strategies and improve both robustness and generalization.

Method: PerturbEmbedding performs perturbation operations directly on every hidden embedding of GNNs, offering a unified framework that encompasses most existing perturbation strategies and handles both random and adversarial perturbations.

Result: Experiments across various datasets and backbone models show that PerturbEmbedding significantly improves both robustness and generalization abilities of GNNs, outperforming existing methods and enhancing model performance against both random and adversarial perturbations.

Conclusion: The proposed PerturbEmbedding method provides an effective unified framework for adversarial training in GNNs, successfully enhancing model resilience to attacks while improving generalization capabilities across diverse datasets and model architectures.

Abstract: This paper studies the vulnerability of Graph Neural Networks (GNNs) to adversarial attacks on node features and graph structure. Various methods have implemented adversarial training to augment graph data, aiming to bolster the robustness and generalization of GNNs. These methods typically involve applying perturbations to the node feature, weights, or graph structure and subsequently minimizing the loss by learning more robust graph model parameters under the adversarial perturbations. Despite the effectiveness of adversarial training in enhancing GNNs’ robustness and generalization abilities, its application has been largely confined to specific datasets and GNN types. In this paper, we propose a novel method, PerturbEmbedding, that integrates adversarial perturbation and training, enhancing GNNs’ resilience to such attacks and improving their generalization ability. PerturbEmbedding performs perturbation operations directly on every hidden embedding of GNNs and provides a unified framework for most existing perturbation strategies/methods. We also offer a unified perspective on the forms of perturbations, namely random and adversarial perturbations. Through experiments on various datasets using different backbone models, we demonstrate that PerturbEmbedding significantly improves both the robustness and generalization abilities of GNNs, outperforming existing methods. The rejection of both random (non-targeted) and adversarial (targeted) perturbations further enhances the backbone model’s performance.

[785] Curriculum Guided Personalized Subgraph Federated Learning

Minku Kang, Hogun Park

Main category: cs.LG

TL;DR: CUFL introduces curriculum learning and improved client similarity estimation to address overfitting and bias in subgraph federated learning, achieving better performance through paced training and enhanced aggregation.

DetailsMotivation: Subgraph FL suffers from severe data heterogeneity and rapid overfitting to sparse, biased subgraphs, causing client similarity matrices to collapse and aggregation to lose effectiveness as clients reinforce their own biases.

Method: CUFL uses curriculum learning to adaptively select edges based on reconstruction scores, exposing GNNs to easier cross-client substructures first and harder client-specific ones later. It also improves weighted aggregation by estimating client similarity using fine-grained structural indicators on a random reference graph.

Result: Extensive experiments on six benchmark datasets confirm that CUFL achieves superior performance compared to relevant baselines.

Conclusion: CUFL effectively mitigates data heterogeneity and overfitting in subgraph FL through curriculum-guided training and improved similarity estimation, enabling better knowledge exchange and personalization.

Abstract: Subgraph Federated Learning (FL) aims to train Graph Neural Networks (GNNs) across distributed private subgraphs, but it suffers from severe data heterogeneity. To mitigate data heterogeneity, weighted model aggregation personalizes each local GNN by assigning larger weights to parameters from clients with similar subgraph characteristics inferred from their current model states. However, the sparse and biased subgraphs often trigger rapid overfitting, causing the estimated client similarity matrix to stagnate or even collapse. As a result, aggregation loses effectiveness as clients reinforce their own biases instead of exploiting diverse knowledge otherwise available. To this end, we propose a novel personalized subgraph FL framework called Curriculum guided personalized sUbgraph Federated Learning (CUFL). On the client side, CUFL adopts Curriculum Learning (CL) that adaptively selects edges for training according to their reconstruction scores, exposing each GNN first to easier, generic cross-client substructures and only later to harder, client-specific ones. This paced exposure prevents early overfitting to biased patterns and enables gradual personalization. By regulating personalization, the curriculum also reshapes server aggregation from exchanging generic knowledge to propagating client-specific knowledge. Further, CUFL improves weighted aggregation by estimating client similarity using fine-grained structural indicators reconstructed on a random reference graph. Extensive experiments on six benchmark datasets confirm that CUFL achieves superior performance compared to relevant baselines. Code is available at https://github.com/Kang-Min-Ku/CUFL.git.

[786] Metis: Training Large Language Models with Advanced Low-Bit Quantization

Hengjie Cao, Mengyi Chen, Yifeng Yang, Ruijun Huang, Fang Dong, Jixian Zhou, Anrui Chen, Mingzhi Dong, Yujiang Wang, Jinlong Hou, Yuan Cheng, Fan Wu, Fan Yang, Tun Lu, Ning Gu, Li Shang

Main category: cs.LG

TL;DR: Metis is a training framework that enables stable low-bit quantization (FP8/FP4) for large language models by addressing anisotropic parameter distributions through spectral decomposition, adaptive learning rates, and dual-range regularization.

DetailsMotivation: Anisotropic parameter distributions with dominant singular values create wide numerical ranges that conflict with block-wise quantization bias, causing training instability and poor performance in low-bit LLM training.

Method: Combines (i) spectral decomposition with random embedding to compress broad distributions, (ii) adaptive learning rates in spectral domain to amplify underrepresented features, and (iii) dual-range regularizer to constrain numerical precision and parameter range distribution.

Result: FP8 training surpasses FP32 baselines, and FP4 training achieves accuracy comparable to FP32, enabling robust and scalable LLM training with advanced low-bit quantization.

Conclusion: Metis successfully overcomes the fundamental barrier of anisotropic parameter distributions in low-bit quantization, providing a framework for stable and high-performance LLM training with significantly reduced precision requirements.

Abstract: This work identifies anisotropic parameter distributions as a fundamental barrier to training large language models (LLMs) with low-bit quantization: a few dominant singular values create wide numerical ranges that conflict with the inherent bias of block-wise quantization. This bias disproportionately preserves high-magnitude values while discarding smaller ones, causing training instability and low model performance. This work introduces Metis, a training framework that combines (i) spectral decomposition with random embedding to efficiently disentangle dominant from long-tail components, compressing broad distributions into quantization-friendly narrow ranges; (ii) adaptive learning rates in the spectral domain to amplify underrepresented directions and better capture diverse features critical for performance; and (iii) a dual-range regularizer that jointly constrains numerical precision and parameter range distribution, ensuring stable, unbiased low-bit training. With Metis, FP8 training surpasses FP32 baselines, and FP4 training achieves accuracy comparable to FP32, paving the way for robust and scalable LLM training under advanced low-bit quantization. The code implementation for Metis is available at: https://github.com/typename-yyf/Metis-quantization.

[787] Lagrangian Relaxation for Multi-Action Partially Observable Restless Bandits: Heuristic Policies and Indexability

Rahul Meshram, Kesav Kaza

Main category: cs.LG

TL;DR: This paper studies multi-action partially observable restless multi-armed bandits with finite states and actions, motivated by public-health intervention planning applications. It analyzes Lagrangian bounds, presents approximation methods using PBVI and online rollout policies, and discusses Whittle index policies with their limitations.

DetailsMotivation: The research is motivated by applications in recommendation systems, communication systems, public healthcare outreach, and operations research, particularly focusing on public-health intervention planning where multiple actions are available for each bandit with budget constraints.

Method: The paper formulates a discounted optimization problem for partially observable restless bandits, analyzes Lagrangian bound methods, and develops approximations using point-based value iteration (PBVI) and online rollout policies. It also studies heuristic policies and Whittle index approaches.

Result: The research provides theoretical insights on PBVI and online rollout policies, analyzes properties of value functions, and presents computational approximations for Lagrangian bounds in partially observable settings with multiple actions.

Conclusion: The paper addresses the challenging problem of multi-action partially observable restless bandits, offering approximation methods and theoretical analysis, while also highlighting limitations of Whittle index policies in this extended model with multiple actions per bandit.

Abstract: Partially observable restless multi-armed bandits have found numerous applications including in recommendation systems, communication systems, public healthcare outreach systems, and in operations research. We study multi-action partially observable restless multi-armed bandits, it is a generalization of the classical restless multi-armed bandit problem – 1) each bandit has finite states, and the current state is not observable, 2) each bandit has finite actions. In particular, we assume that more than two actions are available for each bandit. We motivate our problem with the application of public-health intervention planning. We describe the model and formulate a long term discounted optimization problem, where the state of each bandit evolves according to a Markov process, and this evolution is action dependent. The state of a bandit is not observable but one of finitely many feedback signals are observable. Each bandit yields a reward, based on the action taken on that bandit. The agent is assumed to have a budget constraint. The bandits are assumed to be independent. However, they are weakly coupled at the agent through the budget constraint. We first analyze the Lagrangian bound method for our partially observable restless bandits. The computation of optimal value functions for finite-state, finite-action POMDPs is non-trivial. Hence, the computation of Lagrangian bounds is also challenging. We describe approximations for the computation of Lagrangian bounds using point based value iteration (PBVI) and online rollout policy. We further present various properties of the value functions and provide theoretical insights on PBVI and online rollout policy. We study heuristic policies for multi-actions PORMAB. Finally, we discuss present Whittle index policies and their limitations in our model.

[788] Memory Limitations of Prompt Tuning in Transformers

Maxime Meyer, Mario Michelessa, Caroline Chaux, Vincent Y. F. Tan

Main category: cs.LG

TL;DR: Prompt tuning’s memorization capability is theoretically analyzed, showing linear scaling with prompt length and proving transformers have inherent memory limitations that cause performance degradation with extended contexts.

DetailsMotivation: Despite empirical success of prompt tuning, theoretical understanding of its memorization capabilities and limitations remains limited. The paper aims to provide formal theoretical analysis of transformers' memory constraints.

Method: Theoretical analysis proving two main results: 1) information memorization scales linearly with prompt length, and 2) formal proof of performance degradation with extended contexts due to inherent memory limitations.

Result: Transformers cannot memorize information faster than linearly with prompt length, and they inherently have limited memory capacity that constrains information retention regardless of context size.

Conclusion: Transformers have fundamental architectural limitations in handling long sequences, providing theoretical explanation for observed performance degradation with extended contexts in large language models.

Abstract: Despite the empirical success of prompt tuning in adapting pretrained language models to new tasks, theoretical analyses of its capabilities remain limited. Existing theoretical work primarily addresses universal approximation properties, demonstrating results comparable to standard weight tuning. In this paper, we explore a different aspect of the theory of transformers: the memorization capability of prompt tuning. We provide two principal theoretical contributions. First, we prove that the amount of information memorized by a transformer cannot scale faster than linearly with the prompt length. Second, and more importantly, we present the first formal proof of a phenomenon empirically observed in large language models: performance degradation in transformers with extended contexts. We rigorously demonstrate that transformers inherently have limited memory, constraining the amount of information they can retain, regardless of the context size. This finding offers a fundamental understanding of the intrinsic limitations of transformer architectures, particularly their ability to handle long sequences.

[789] Universal Properties of Activation Sparsity in Modern Large Language Models

Filip Szatkowski, Patryk Będkowski, Alessio Devoto, Jan Dubiński, Pasquale Minervini, Mikołaj Piórczyński, Simone Scardapane, Bartosz Wójcik

Main category: cs.LG

TL;DR: The paper proposes a general framework to study activation sparsity in modern LLMs, revealing universal patterns and providing practical guidelines for model design and acceleration.

DetailsMotivation: Input-dependent activation sparsity has been well-studied in ReLU-based models but not in modern LLMs that use different activation functions, leading to fragmented and model-specific approaches without consensus.

Method: The authors propose a general framework to assess sparsity robustness and conduct a systematic study of activation sparsity in FFN layers of modern LLMs, including diffusion LLMs.

Result: The study reveals universal patterns of activation sparsity in LLMs, providing insights into this phenomenon across different model architectures.

Conclusion: The findings offer practical guidelines for exploiting activation sparsity in LLM design and acceleration, addressing the gap in understanding sparsity patterns in modern activation functions beyond ReLU.

Abstract: Input-dependent activation sparsity is a notable property of deep learning models, which has been extensively studied in networks with ReLU activations and is associated with efficiency, robustness, and interpretability. However, the approaches developed for ReLU-based models depend on exact zero activations and do not transfer directly to modern large language models~(LLMs), which have abandoned ReLU in favor of other activation functions. As a result, current work on activation sparsity in LLMs is fragmented, model-specific, and lacks consensus on which components to target. We propose a general framework to assess sparsity robustness and present a systematic study of the phenomenon in the FFN layers of modern LLMs, including diffusion LLMs. Our findings reveal universal patterns of activation sparsity in LLMs, provide insights into this phenomenon, and offer practical guidelines for exploiting it in model design and acceleration.

[790] Localizing and Mitigating Memorization in Image Autoregressive Models

Aditya Kasliwal, Franziska Boenisch, Adam Dziedzic

Main category: cs.LG

TL;DR: This paper analyzes memorization patterns in Image AutoRegressive (IAR) models, finding different memorization behaviors across architectures and proposing interventions to reduce data leakage while maintaining image quality.

DetailsMotivation: IAR models achieve state-of-the-art image generation performance but raise privacy concerns due to potential memorization of training data, requiring investigation into where and how this memorization occurs.

Method: The researchers measure fine-grained memorization across different IAR architectures, analyzing hierarchical per-resolution models vs standard autoregressive per token prediction models to identify where memorization emerges.

Result: Memorization patterns differ significantly - hierarchical architectures show early emergence that deepens with resolution, while standard autoregressive models concentrate memorization in later processing stages. Interventions on memorizing components significantly reduce data extraction capacity with minimal impact on generated image quality.

Conclusion: The findings provide insights into IAR model behavior and offer practical strategies for mitigating privacy risks by targeting specific memorization components, enabling better privacy-preserving image generation models.

Abstract: Image AutoRegressive (IAR) models have achieved state-of-the-art performance in speed and quality of generated images. However, they also raise concerns about memorization of their training data and its implications for privacy. This work explores where and how such memorization occurs within different image autoregressive architectures by measuring a fine-grained memorization. The analysis reveals that memorization patterns differ across various architectures of IARs. In hierarchical per-resolution architectures, it tends to emerge early and deepen with resolutions, while in IARs with standard autoregressive per token prediction, it concentrates in later processing stages. These localization of memorization patterns are further connected to IARs’ ability to memorize and leak training data. By intervening on their most memorizing components, we significantly reduce the capacity for data extraction from IARs with minimal impact on the quality of generated images. These findings offer new insights into the internal behavior of image generative models and point toward practical strategies for mitigating privacy risks.

[791] Graph Convolutional Network With Pattern-Spatial Interactive and Regional Awareness for Traffic Forecasting

Xinyu Ji, Chengcheng Yan, Jibiao Yuan, Fiefie Zhao

Main category: cs.LG

TL;DR: PSIRAGCN is a novel graph convolutional network that addresses limitations in traffic forecasting by modeling pattern-spatial interactions and regional heterogeneity through interactive fusion and data-driven message passing.

DetailsMotivation: Existing spatial-temporal models struggle with effectively modeling spatial-temporal correlations across different perceptual perspectives and neglect the interactive fusion between traffic patterns and spatial correlations. Most studies also fail to consider regional heterogeneity during message-passing due to spatial constraints.

Method: Proposed Pattern-Spatial Interactive and Regional Awareness GCN (PSIRAGCN) with: 1) Pattern-spatial interactive fusion framework capturing correlations from global to local levels with mutual feedback, 2) Graph convolutional network using regional characteristics bank for data-driven message passing with regional awareness to reveal heterogeneity.

Result: Extensive experiments on three real-world traffic datasets show PSIRAGCN outperforms state-of-the-art baselines while maintaining computational efficiency.

Conclusion: PSIRAGCN successfully addresses key limitations in traffic forecasting by effectively modeling pattern-spatial interactions and regional heterogeneity, achieving superior performance with balanced computational costs.

Abstract: Traffic forecasting is significant for urban traffic management, intelligent route planning, and real-time flow monitoring. Recent advances in spatial-temporal models have markedly improved the modeling of intricate spatial-temporal correlations for traffic forecasting. Unfortunately, most previous studies have encountered challenges in effectively modeling spatial-temporal correlations across various perceptual perspectives, which have neglected the interactive fusion between traffic patterns and spatial correlations. Additionally, constrained by spatial heterogeneity, most studies fail to consider distinct regional heterogeneity during message-passing. To overcome these limitations, we propose a Pattern-Spatial Interactive and Regional Awareness Graph Convolutional Network (PSIRAGCN) for traffic forecasting. Specifically, we propose a pattern-spatial interactive fusion framework composed of pattern and spatial modules. This framework aims to capture patterns and spatial correlations by adopting a perception perspective from the global to the local level and facilitating mutual utilization with positive feedback. In the spatial module, we designed a graph convolutional network based on message-passing. The network is designed to leverage a regional characteristics bank to reconstruct data-driven message-passing with regional awareness. Reconstructed message passing can reveal the regional heterogeneity between nodes in the traffic network. Extensive experiments on three real-world traffic datasets demonstrate that PSIRAGCN outperforms the State-of-the-art baseline while balancing computational costs.

[792] Biological Pathway Informed Models with Graph Attention Networks (GATs)

Gavin Wong, Ping Shu Ho, Ivan Au Yeung, Ka Chun Cheung, Simon See

Main category: cs.LG

TL;DR: GAT framework models biological pathways at gene level, outperforming MLP approaches with 81% MSE reduction and successfully rediscovers known gene interactions from raw data.

DetailsMotivation: Current ML models treat genes as unstructured tokens or pathways as 'bags of genes', discarding known pathway structure and gene-gene interactions that are crucial for understanding biological processes.

Method: Proposes a Graph Attention Network (GAT) framework that models pathways at the gene level, encoding drug mechanisms via edge interventions to boost model robustness.

Result: GATs achieve 81% reduction in MSE when predicting pathway dynamics under unseen treatment conditions and successfully rediscover all five gene-gene interactions in the canonical TP53-MDM2-MDM4 feedback loop from raw time-series mRNA data.

Conclusion: The GAT framework demonstrates superior generalization over MLP approaches and shows potential for generating novel biological hypotheses directly from experimental data by properly capturing pathway topology and gene interactions.

Abstract: Biological pathways map gene-gene interactions that govern all human processes. Despite their importance, most ML models treat genes as unstructured tokens, discarding known pathway structure. The latest pathway-informed models capture pathway-pathway interactions, but still treat each pathway as a “bag of genes” via MLPs, discarding its topology and gene-gene interactions. We propose a Graph Attention Network (GAT) framework that models pathways at the gene level. We show that GATs generalize much better than MLPs, achieving an 81% reduction in MSE when predicting pathway dynamics under unseen treatment conditions. We further validate the correctness of our biological prior by encoding drug mechanisms via edge interventions, boosting model robustness. Finally, we show that our GAT model is able to correctly rediscover all five gene-gene interactions in the canonical TP53-MDM2-MDM4 feedback loop from raw time-series mRNA data, demonstrating potential to generate novel biological hypotheses directly from experimental data.

[793] FedThief: Harming Others to Benefit Oneself in Self-Centered Federated Learning

Xiangyu Zhang, Mang Ye

Main category: cs.LG

TL;DR: FedThief is a self-centered federated learning attack that degrades global model performance while enhancing the attacker’s private model through divergence-aware ensemble techniques.

DetailsMotivation: Existing FL attacks degrade both global and attacker models, but real attackers want competitive advantage - better models than others, not just disruption.

Method: FedThief framework modifies uploads to degrade global model while using divergence-aware ensemble techniques to integrate global updates and local knowledge for private model enhancement.

Result: Extensive experiments show effective global model degradation and attacker obtaining ensemble model that significantly outperforms the global model.

Conclusion: SCFL attack paradigm enables attackers to gain competitive advantage by simultaneously degrading global performance and enhancing private models through strategic update manipulation.

Abstract: In federated learning, participants’ uploaded model updates cannot be directly verified, leaving the system vulnerable to malicious attacks. Existing attack strategies have adversaries upload tampered model updates to degrade the global model’s performance. However, attackers also degrade their own private models, gaining no advantage. In real-world scenarios, attackers are driven by self-centered motives: their goal is to gain a competitive advantage by developing a model that outperforms those of other participants, not merely to cause disruption. In this paper, we study a novel Self-Centered Federated Learning (SCFL) attack paradigm, in which attackers not only degrade the performance of the global model through attacks but also enhance their own models within the federated learning process. We propose a framework named FedThief, which degrades the performance of the global model by uploading modified content during the upload stage. At the same time, it enhances the private model’s performance through divergence-aware ensemble techniques, where “divergence” quantifies the deviation between private and global models, that integrate global updates and local knowledge. Extensive experiments show that our method effectively degrades the global model performance while allowing the attacker to obtain an ensemble model that significantly outperforms the global model.

[794] Advanced spectral clustering for heterogeneous data in credit risk monitoring systems

Lu Han, Mengyan Li, Jiping Qiang, Zhi Su

Main category: cs.LG

TL;DR: ASC integrates financial and textual data for credit monitoring using optimized spectral clustering, achieving 18% higher Silhouette scores and identifying meaningful risk patterns like 30% lower default risk in recruitment-focused SMEs.

DetailsMotivation: Heterogeneous data (numerical financial variables and textual records) present challenges for credit monitoring of SMEs, requiring methods that can effectively integrate both data types for better risk assessment.

Method: Advanced Spectral Clustering (ASC) integrates financial and textual similarities through optimized weight parameters and uses eigenvalue-silhouette optimization for eigenvector selection.

Result: ASC achieved 18% higher Silhouette score than baseline, identified that 51% of low-risk firms include ‘social recruitment’ in records, and showed recruitment-focused SMEs have 30% lower default risk. Robust across multiple algorithms with {Δ}Intra/Inter < 0.13.

Conclusion: ASC effectively bridges spectral clustering theory with heterogeneous data applications, enabling identification of meaningful clusters and supporting more targeted credit interventions for SMEs.

Abstract: Heterogeneous data, which encompass both numerical financial variables and textual records, present substantial challenges for credit monitoring. To address this issue, we propose Advanced Spectral Clustering (ASC), a method that integrates financial and textual similarities through an optimized weight parameter and selects eigenvectors using a novel eigenvalue-silhouette optimization approach. Evaluated on a dataset comprising 1,428 small and medium-sized enterprises (SMEs), ASC achieves a Silhouette score that is 18% higher than that of a single-type data baseline method. Furthermore, the resulting clusters offer actionable insights; for instance, 51% of low-risk firms are found to include the term ‘social recruitment’ in their textual records. The robustness of ASC is confirmed across multiple clustering algorithms, including k-means, k-medians, and k-medoids, with {\Delta}Intra/Inter < 0.13 and {\Delta}Silhouette Coefficient < 0.02. By bridging spectral clustering theory with heterogeneous data applications, ASC enables the identification of meaningful clusters, such as recruitment-focused SMEs exhibiting a 30% lower default risk, thereby supporting more targeted and effective credit interventions.

[795] Integrated Multivariate Segmentation Tree for the Analysis of Heterogeneous Credit Data in Small and Medium-Sized Enterprises

Lu Han, Xiuying Wang

Main category: cs.LG

TL;DR: IMST framework integrates financial and textual data for SME credit evaluation, achieving 88.9% accuracy with improved interpretability over traditional models.

DetailsMotivation: Traditional decision trees struggle with high-dimensional data and cannot effectively incorporate textual information, limiting their effectiveness in SME credit evaluation.

Method: Three-stage approach: (1) matrix factorization to convert text to numerical matrices, (2) Lasso regression for financial feature selection, (3) multivariate segmentation tree construction using Gini/Entropy with weakest-link pruning.

Result: 88.9% accuracy on 1,428 Chinese SMEs dataset, outperforming baseline decision trees (87.4%), logistic regression, and SVM models.

Conclusion: IMST provides superior accuracy, interpretability, computational efficiency, and risk detection capabilities for SME credit evaluation compared to traditional methods.

Abstract: Traditional decision tree models, which rely exclusively on numerical variables, often encounter difficulties in handling high-dimensional data and fail to effectively incorporate textual information. To address these limitations, we propose the Integrated Multivariate Segmentation Tree (IMST), a comprehensive framework designed to enhance credit evaluation for small and medium-sized enterprises (SMEs) by integrating financial data with textual sources. The methodology comprises three core stages: (1) transforming textual data into numerical matrices through matrix factorization; (2) selecting salient financial features using Lasso regression; and (3) constructing a multivariate segmentation tree based on the Gini index or Entropy, with weakest-link pruning applied to regulate model complexity. Experimental results derived from a dataset of 1,428 Chinese SMEs demonstrate that IMST achieves an accuracy of 88.9%, surpassing baseline decision trees (87.4%) as well as conventional models such as logistic regression and support vector machines (SVM). Furthermore, the proposed model exhibits superior interpretability and computational efficiency, featuring a more streamlined architecture and enhanced risk detection capabilities.

[796] An Efficient GNNs-to-KANs Distillation via Self-Attention Dynamic Sampling with Potential for Consumer Electronics Edge Deployment

Can Cui, Zilong Fu, Penghe Huang, Yuanyuan Li, Wu Deng, Dongyan Li

Main category: cs.LG

TL;DR: Proposes SA-DSD framework for distilling GNN knowledge to KANs instead of MLPs, achieving significant performance gains and efficiency improvements for edge deployment.

DetailsMotivation: Edge devices need efficient models but MLPs struggle to capture GNN's complex neighborhood dependencies. KANs offer better nonlinear fitting with lower complexity.

Method: Improved Fourier KAN (FR-KAN+) as student model with learnable frequency bases and phase-shift. Uses margin-level sampling probability matrix and adaptive weighted loss for distillation.

Result: 3.05%-3.62% improvement over GNN teachers, 15.61% over FR-KAN+, 16.96x parameter reduction, 55.75% faster inference compared to benchmarks.

Conclusion: SA-DSD effectively transfers GNN knowledge to efficient KAN models, enabling high-performance deployment on resource-constrained edge devices.

Abstract: Knowledge distillation (KD) is crucial for deploying deep learning models in resource-constrained edge environments, particularly within the consumer electronics sector, including smart home devices, wearable technology, and mobile terminals. These applications place higher demands on model compression and inference speed, necessitating the transfer of knowledge from Graph Neural Networks (GNNs) to more efficient Multi-Layer Perceptron (MLP) models. However, due to their fixed activation functions and fully connected architecture, MLPs face challenges in rapidly capturing the complex neighborhood dependencies learned by GNNs, thereby limiting their performance in edge environments. To address these limitations, this paper introduces an innovative from GNNs to Kolmogorov-Arnold Networks (KANs) knowledge distillation framework-Self Attention Dynamic Sampling Distillation (SA-DSD). This study improved Fourier KAN (FR-KAN) and replaced MLP with the improved FR-KAN+ as the student model. Through the incorporation of learnable frequency bases and phase-shift mechanisms, along with algorithmic optimization, FR-KAN significantly improves its nonlinear fitting capability while effectively reducing computational complexity. Building on this, a margin-level sampling probability matrix, based on teacher-student prediction consistency, is constructed, and an adaptive weighted loss mechanism is designed to mitigate performance degradation in the student model due to the lack of explicit neighborhood aggregation. Extensive experiments conducted on six real-world datasets demonstrate that SA-DSD achieves performance improvements of 3.05%-3.62% over three GNN teacher models and 15.61% over the FR-KAN+ model. Moreover, when compared with key benchmark models, SA-DSD achieves a 16.96x reduction in parameter count and a 55.75% decrease in inference time.

[797] TranCIT: Transient Causal Interaction Toolbox

Salar Nouri, Kaidi Shao, Shervin Safavi

Main category: cs.LG

TL;DR: TranCIT is an open-source Python toolbox for detecting transient causal interactions in neural signals, overcoming limitations of traditional methods for brief neural events.

DetailsMotivation: Traditional methods are inadequate for brief neural events, and existing advanced techniques lack accessible Python implementations for quantifying transient causal interactions in neuroscience.

Method: Implements comprehensive analysis pipeline including Granger Causality, Transfer Entropy, and robust Structural Causal Model-based methods (DCS and rDCS) for detecting event-driven causal effects.

Result: Successfully captures causality in high-synchrony regimes where traditional methods fail, and identifies known transient information flow from hippocampal CA3 to CA1 during sharp-wave ripple events in real data.

Conclusion: TranCIT provides a user-friendly, validated solution for investigating transient causal dynamics in complex neural systems, bridging the implementation gap in Python ecosystem.

Abstract: Quantifying transient causal interactions from non-stationary neural signals is a fundamental challenge in neuroscience. Traditional methods are often inadequate for brief neural events, and advanced, event-specific techniques have lacked accessible implementations within the Python ecosystem. Here, we introduce trancit (Transient Causal Interaction Toolbox), an open-source Python package designed to bridge this gap. TranCIT implements a comprehensive analysis pipeline, including Granger Causality, Transfer Entropy, and the more robust Structural Causal Model-based Dynamic Causal Strength (DCS) and relative Dynamic Causal Strength (rDCS) for accurately detecting event-driven causal effects. We demonstrate TranCIT’s utility by successfully capturing causality in high-synchrony regimes where traditional methods fail and by identifying the known transient information flow from hippocampal CA3 to CA1 during sharp-wave ripple events in real-world data. The package offers a user-friendly, validated solution for investigating the transient causal dynamics that govern complex systems.

[798] RoFt-Mol: Benchmarking Robust Fine-Tuning with Molecular Graph Foundation Models

Shikun Liu, Deyu Zou, Nima Shoghi, Victor Fung, Kai Liu, Pan Li

Main category: cs.LG

TL;DR: This paper analyzes fine-tuning methods for molecular graph foundation models, classifies 8 methods into 3 categories, benchmarks them on various tasks, and proposes ROFT-MOL - a robust fine-tuning approach combining weight interpolation and ensemble methods.

DetailsMotivation: Molecular graph foundation models face unique challenges including smaller pre-training datasets, severe data scarcity for downstream tasks, and need to handle both regression and classification objectives, requiring enhanced fine-tuning methods.

Method: Classified 8 fine-tuning methods into three mechanisms (weight-based, representation-based, partial fine-tuning), benchmarked them on downstream tasks across supervised and self-supervised pre-trained models, then designed ROFT-MOL combining weight interpolation with weight ensemble methods.

Result: Extensive evaluation provided insights leading to ROFT-MOL, which delivers improved performance across both regression and classification tasks while maintaining ease of use.

Conclusion: ROFT-MOL effectively addresses the unique fine-tuning challenges in molecular graph foundation models by combining the strengths of different fine-tuning approaches for robust performance across diverse objectives.

Abstract: In the era of foundation models, fine-tuning pre-trained models for specific downstream tasks has become crucial. This drives the need for robust fine-tuning methods to address challenges such as model overfitting and sparse labeling. Molecular graph foundation models (MGFMs) face unique difficulties that complicate fine-tuning. These models are limited by smaller pre-training datasets and more severe data scarcity for downstream tasks, both of which require enhanced model generalization. Moreover, MGFMs must accommodate diverse objectives, including both regression and classification tasks. To better understand and improve fine-tuning techniques under these conditions, we classify eight fine-tuning methods into three mechanisms: weight-based, representation-based, and partial fine-tuning. We benchmark these methods on downstream regression and classification tasks across supervised and self-supervised pre-trained models in diverse labeling settings. This extensive evaluation provides valuable insights and informs the design of a refined robust fine-tuning method, ROFT-MOL. This approach combines the strengths of simple post-hoc weight interpolation with more complex weight ensemble fine-tuning methods, delivering improved performance across both task types while maintaining the ease of use inherent in post-hoc weight interpolation.

[799] TimeCopilot

Azul Garza, Reneé Rosillo

Main category: cs.LG

TL;DR: TimeCopilot is the first open-source agentic framework that combines Time Series Foundation Models with LLMs through a unified API to automate forecasting pipelines with natural language explanations.

DetailsMotivation: To create a practical, reproducible, and accessible agentic forecasting system that automates the entire forecasting pipeline while providing explainable results through natural language interfaces.

Method: Combines multiple Time Series Foundation Models with Large Language Models through a unified API, automating feature analysis, model selection, cross-validation, and forecast generation. The framework is LLM-agnostic and supports ensemble methods across diverse forecasting families.

Result: Achieves state-of-the-art probabilistic forecasting performance on the large-scale GIFT-Eval benchmark at low cost.

Conclusion: TimeCopilot provides a practical foundation for reproducible, explainable, and accessible agentic forecasting systems that can handle complex forecasting tasks with natural language interfaces.

Abstract: We introduce TimeCopilot, the first open-source agentic framework for forecasting that combines multiple Time Series Foundation Models (TSFMs) with Large Language Models (LLMs) through a single unified API. TimeCopilot automates the forecasting pipeline: feature analysis, model selection, cross-validation, and forecast generation, while providing natural language explanations and supporting direct queries about the future. The framework is LLM-agnostic, compatible with both commercial and open-source models, and supports ensembles across diverse forecasting families. Results on the large-scale GIFT-Eval benchmark show that TimeCopilot achieves state-of-the-art probabilistic forecasting performance at low cost. Our framework provides a practical foundation for reproducible, explainable, and accessible agentic forecasting systems.

[800] Forecasting the Ionosphere from Sparse GNSS Data with Temporal-Fusion Transformers

Giacomo Acciarini, Simone Mestici, Halil Kelebek, Linnea Wolniewicz, Michael Vergalla, Madhulika Guhathakurta, Umaa Rebbapragada, Bala Poduval, Atılım Güneş Baydin, Frank Soboczenski

Main category: cs.LG

TL;DR: Machine learning framework using Temporal Fusion Transformers for accurate 24-hour ionospheric TEC forecasting with interpretability and open-source toolkit.

DetailsMotivation: Accurate ionospheric prediction is challenging due to nonlinear couplings between solar, geomagnetic, and thermospheric drivers, and limited forecasting capabilities during space weather events.

Method: Leverages Temporal Fusion Transformers to predict sparse ionosphere data, incorporating heterogeneous inputs (solar irradiance, geomagnetic indices, GNSS-derived TEC) with preprocessing and temporal alignment strategies.

Result: Achieves robust predictions up to 24 hours ahead with root mean square errors as low as 3.33 TECU, with solar EUV irradiance identified as strongest predictive signal.

Conclusion: Framework provides both accurate forecasting and interpretability through attention-based analysis, released as open-source toolkit ionopy to support operational applications and scientific discovery.

Abstract: The ionosphere critically influences Global Navigation Satellite Systems (GNSS), satellite communications, and Low Earth Orbit (LEO) operations, yet accurate prediction of its variability remains challenging due to nonlinear couplings between solar, geomagnetic, and thermospheric drivers. Total Electron Content (TEC), a key ionospheric parameter, is derived from GNSS observations, but its reliable forecasting is limited by the sparse nature of global measurements and the limited accuracy of empirical models, especially during strong space weather conditions. In this work, we present a machine learning framework for ionospheric TEC forecasting that leverages Temporal Fusion Transformers (TFT) to predict sparse ionosphere data. Our approach accommodates heterogeneous input sources, including solar irradiance, geomagnetic indices, and GNSS-derived vertical TEC, and applies preprocessing and temporal alignment strategies. Experiments spanning 2010-2025 demonstrate that the model achieves robust predictions up to 24 hours ahead, with root mean square errors as low as 3.33 TECU. Results highlight that solar EUV irradiance provides the strongest predictive signals. Beyond forecasting accuracy, the framework offers interpretability through attention-based analysis, supporting both operational applications and scientific discovery. To encourage reproducibility and community-driven development, we release the full implementation as the open-source toolkit \texttt{ionopy}.

[801] Disentangling Slow and Fast Temporal Dynamics in Degradation Inference with Hierarchical Differential Models

Mengjie Zhao, Olga Fink

Main category: cs.LG

TL;DR: Proposes Hierarchical Controlled Differential Equations (H-CDE) framework to disentangle degradation from operational dynamics in sensor data, addressing limitations of existing methods for condition monitoring.

DetailsMotivation: Existing methods struggle to separate subtle long-term degradation from dominant short-term operational variations in sensor data, particularly in dynamic systems where residuals remain entangled with operational history.

Method: H-CDE framework with slow (degradation) and fast (operation) CDE components, multi-scale time integration, learnable path transformation for degradation drivers, and monotonic activation function for regularization.

Result: Outperforms residual-based baselines in both dynamic response and steady state systems, providing more accurate, robust, and interpretable degradation inference.

Conclusion: H-CDE effectively addresses numerical stiffness and degradation disentanglement challenges, offering a superior framework for condition monitoring and prognostics in engineered systems.

Abstract: Reliable inference of system degradation from sensor data is fundamental to condition monitoring and prognostics in engineered systems. Since degradation is rarely observable and measurable, it must be inferred to enable accurate health assessment and decision-making. This is particularly challenging because operational variations dominate system behavior, while degradation introduces only subtle, long-term changes. Consequently, sensor data mainly reflect short-term operational variability, making it difficult to disentangle the underlying degradation process. Residual-based methods are widely employed, but the residuals remain entangled with operational history, often resulting in noisy and unreliable degradation estimation, particularly in systems with dynamic responses. Neural Ordinary Equations (NODEs) offer a promising framework for inferring latent dynamics, but the time-scale separation in slow-fast systems introduces numerical stiffness and complicates training, while degradation disentanglement remains difficult. To address these limitations, we propose a novel Hierarchical Controlled Differential Equation (H-CDE) framework that incorporates a slow (degradation) and a fast (operation) CDE component in a unified architecture. It introduces three key innovations: a multi-scale time integration scheme to mitigate numerical stiffness; a learnable path transformation that extracts latent degradation drivers to control degradation evolution; and a novel activation function that enforces monotonicity on inferred degradation as a regularizer for disentanglement. Through comprehensive evaluations on both dynamic response (e.g., bridges) and steady state (e.g., aero-engine) systems, we demonstrate that H-CDE effectively disentangles degradation from operational dynamics and outperforms residual-based baselines, yielding more accurate, robust, and interpretable inference.

Zhipeng Yin, Zichong Wang, Avash Palikhe, Zhen Liu, Jun Liu, Wenbin Zhang

Main category: cs.LG

TL;DR: AMCR framework addresses copyright risks in text-to-image models through systematic prompt restructuring, attention-based similarity detection, and adaptive risk mitigation during generation.

DetailsMotivation: Existing prompt-based copyright mitigation methods fail to handle subtle infringement cases where seemingly benign prompts still produce copyrighted content, creating legal and ethical challenges for generative model deployment.

Method: Three-pronged approach: 1) Systematic restructuring of risky prompts into safe forms, 2) Attention-based similarity analysis to detect partial infringements, 3) Adaptive risk mitigation during generation process to reduce violations while maintaining image quality.

Result: Extensive experiments validate AMCR’s effectiveness in revealing and mitigating latent copyright risks, providing practical benchmarks for safer generative model deployment.

Conclusion: AMCR offers a comprehensive framework that addresses limitations of existing methods by handling both obvious and subtle copyright infringement cases, enabling safer real-world deployment of text-to-image generative models.

Abstract: Generative models have achieved impressive results in text to image tasks, significantly advancing visual content creation. However, this progress comes at a cost, as such models rely heavily on large-scale training data and may unintentionally replicate copyrighted elements, creating serious legal and ethical challenges for real-world deployment. To address these concerns, researchers have proposed various strategies to mitigate copyright risks, most of which are prompt based methods that filter or rewrite user inputs to prevent explicit infringement. While effective in handling obvious cases, these approaches often fall short in more subtle situations, where seemingly benign prompts can still lead to infringing outputs. To address these limitations, this paper introduces Assessing and Mitigating Copyright Risks (AMCR), a comprehensive framework which i) builds upon prompt-based strategies by systematically restructuring risky prompts into safe and non-sensitive forms, ii) detects partial infringements through attention-based similarity analysis, and iii) adaptively mitigates risks during generation to reduce copyright violations without compromising image quality. Extensive experiments validate the effectiveness of AMCR in revealing and mitigating latent copyright risks, offering practical insights and benchmarks for the safer deployment of generative models.

[803] Context-Action Embedding Learning for Off-Policy Evaluation in Contextual Bandits

Kushagra Chandak, Vincent Liu, Haanvid Lee

Main category: cs.LG

TL;DR: CAEL-MIPS learns context-action embeddings to minimize MSE in off-policy evaluation for contextual bandits, outperforming existing methods.

DetailsMotivation: Existing IPS estimators suffer from high variance with large action spaces or underexplored contexts, and MIPS estimators don't minimize MSE or use context information effectively.

Method: Proposes CAEL-MIPS that learns context-action embeddings from offline data specifically to minimize the mean squared error of the MIPS estimator, building on theoretical bias-variance analysis.

Result: Empirical studies on synthetic and real-world datasets show that CAEL-MIPS outperforms baseline estimators in terms of mean squared error.

Conclusion: Learning context-action embeddings that directly minimize MSE leads to improved off-policy evaluation performance compared to existing methods.

Abstract: We consider off-policy evaluation (OPE) in contextual bandits with finite action space. Inverse Propensity Score (IPS) weighting is a widely used method for OPE due to its unbiased, but it suffers from significant variance when the action space is large or when some parts of the context-action space are underexplored. Recently introduced Marginalized IPS (MIPS) estimators mitigate this issue by leveraging action embeddings. However, these embeddings do not minimize the mean squared error (MSE) of the estimators and do not consider context information. To address these limitations, we introduce Context-Action Embedding Learning for MIPS, or CAEL-MIPS, which learns context-action embeddings from offline data to minimize the MSE of the MIPS estimator. Building on the theoretical analysis of bias and variance of MIPS, we present an MSE-minimizing objective for CAEL-MIPS. In the empirical studies on a synthetic dataset and a real-world dataset, we demonstrate that our estimator outperforms baselines in terms of MSE.

[804] Missing Data Imputation using Neural Cellular Automata

Tin Luu, Binh Nguyen, Man Ngo

Main category: cs.LG

TL;DR: Proposes a novel neural cellular automata (NCA) based method for tabular data imputation that outperforms state-of-the-art approaches.

DetailsMotivation: Missing data is a persistent problem in tabular data analysis, and while recent generative models like VAEs and GANs have been explored for imputation, Neural Cellular Automata (NCA) have been overlooked despite their computational power.

Method: Developed an NCA-based imputation model with appropriate adaptations to handle missing data in tabular datasets, leveraging the computational capabilities of neural cellular automata.

Result: The proposed NCA-based model demonstrates superior performance compared to state-of-the-art methods, achieving lower imputation error and better post-imputation task performance.

Conclusion: Neural Cellular Automata represent a promising and previously overlooked approach for tabular data imputation, offering competitive advantages over existing generative model-based methods.

Abstract: When working with tabular data, missingness is always one of the most painful problems. Throughout many years, researchers have continuously explored better and better ways to impute missing data. Recently, with the rapid development evolution in machine learning and deep learning, there is a new trend of leveraging generative models to solve the imputation task. While the imputing version of famous models such as Variational Autoencoders or Generative Adversarial Networks were investigated, prior work has overlooked Neural Cellular Automata (NCA), a powerful computational model. In this paper, we propose a novel imputation method that is inspired by NCA. We show that, with some appropriate adaptations, an NCA-based model is able to address the missing data imputation problem. We also provide several experiments to evidence that our model outperforms state-of-the-art methods in terms of imputation error and post-imputation performance.

[805] IndiaWeatherBench: A Dataset and Benchmark for Data-Driven Regional Weather Forecasting over India

Tung Nguyen, Harkanwar Singh, Nilay Naharas, Lucas Bandarkar, Aditya Grover

Main category: cs.LG

TL;DR: IndiaWeatherBench is a new benchmark for regional weather forecasting focused on India, providing curated datasets, evaluation metrics, and baseline models to standardize research in this domain.

DetailsMotivation: Regional weather forecasting is crucial for climate adaptation and disaster mitigation, but existing efforts lack standardized datasets and evaluation frameworks, making fair comparisons difficult.

Method: Created a comprehensive benchmark with curated high-resolution regional reanalysis data, implemented various models (UNets, Transformers, Graph networks) with different boundary conditioning strategies and training objectives.

Result: Established strong baselines for regional weather forecasting and provided all datasets, model implementations, and evaluation pipelines as open-source resources.

Conclusion: IndiaWeatherBench serves as a foundation for advancing regional weather forecasting research and is easily extensible to other geographic regions.

Abstract: Regional weather forecasting is a critical problem for localized climate adaptation, disaster mitigation, and sustainable development. While machine learning has shown impressive progress in global weather forecasting, regional forecasting remains comparatively underexplored. Existing efforts often use different datasets and experimental setups, limiting fair comparison and reproducibility. We introduce IndiaWeatherBench, a comprehensive benchmark for data-driven regional weather forecasting focused on the Indian subcontinent. IndiaWeatherBench provides a curated dataset built from high-resolution regional reanalysis products, along with a suite of deterministic and probabilistic metrics to facilitate consistent training and evaluation. To establish strong baselines, we implement and evaluate a range of models across diverse architectures, including UNets, Transformers, and Graph-based networks, as well as different boundary conditioning strategies and training objectives. While focused on India, IndiaWeatherBench is easily extensible to other geographic regions. We open-source all raw and preprocessed datasets, model implementations, and evaluation pipelines to promote accessibility and future development. We hope IndiaWeatherBench will serve as a foundation for advancing regional weather forecasting research. Code is available at https://github.com/tung-nd/IndiaWeatherBench.

[806] An Evolutionary Multi-objective Optimization for Replica-Exchange-based Physics-informed Operator Learning Network

Binghang Lu, Changhong Mou, Guang Lin

Main category: cs.LG

TL;DR: Evolutionary multi-objective optimization framework for physics-informed operator learning that balances operator/physics losses, improves exploration via replica exchange dynamics, and provides Bayesian uncertainty quantification.

DetailsMotivation: Existing physics-informed neural networks and operator learning methods struggle with balancing losses, robustness under noisy/sparse data, and uncertainty quantification.

Method: Integrates evolutionary multi-objective optimization for Pareto-optimal loss balancing, replica exchange stochastic gradient Langevin dynamics for global exploration, and Bayesian uncertainty quantification through stochastic sampling.

Result: Outperforms general operator learning methods in accuracy, noise robustness, and uncertainty quantification on 1D Burgers equation and time-fractional mixed diffusion-wave equation.

Conclusion: The proposed framework effectively addresses limitations of current operator learning approaches by providing adaptive loss balancing, improved convergence, and built-in uncertainty quantification.

Abstract: In this paper, we propose an evolutionary Multi-objective Optimization for Replica-Exchange-based Physics-informed Operator learning Network, which is a novel operator learning network to efficiently solve parametric partial differential equations. In forward and inverse settings, this operator learning network only admits minimum requirement of noisy observational data. While physics-informed neural networks and operator learning approaches such as Deep Operator Networks and Fourier Neural Operators offer promising alternatives to traditional numerical solvers, they struggle with balancing operator and physics losses, maintaining robustness under noisy or sparse data, and providing uncertainty quantification. The proposed framework addresses these limitations by integrating: (i) evolutionary multi-objective optimization to adaptively balance operator and physics-based losses in the Pareto front; (ii) replica exchange stochastic gradient Langevin dynamics to improve global parameter-space exploration and accelerate convergence; and (iii) built-in Bayesian uncertainty quantification from stochastic sampling. The proposed operator learning method is tested numerically on several different problems including one-dimensional Burgers equation and the time-fractional mixed diffusion-wave equation. The results indicate that our framework consistently outperforms the general operator learning methods in accuracy, noise robustness, and the ability to quantify uncertainty.

[807] Valid Property-Enhanced Contrastive Learning for Targeted Optimization & Resampling for Novel Drug Design

Amartya Banerjee, Somnath Kar, Anirban Pal, Debabrata Maiti

Main category: cs.LG

TL;DR: VECTOR+ is a framework that combines property-guided contrastive learning with controllable molecule generation for efficient drug discovery in low-data regimes, demonstrating superior performance in generating novel, synthetically tractable compounds with improved binding properties compared to existing methods.

DetailsMotivation: Efficiently steering generative models toward pharmacologically relevant regions of chemical space remains challenging in molecular drug discovery under low-data conditions, requiring methods that can explore functional chemical space with limited training data.

Method: VECTOR+ couples property-guided representation learning with controllable molecule generation using contrastive learning, applicable to both regression and classification tasks for interpretable, data-efficient chemical space exploration.

Result: On PD-L1 inhibitors (296 compounds), VECTOR+ generated 100 of 8,374 molecules surpassing docking threshold (-15.0 kcal/mol), with best score -17.6 kcal/mol vs reference -15.4 kcal/mol. Also outperformed established drugs like brigatinib and sorafenib in kinase inhibitors, with molecular dynamics confirming binding stability (RMSD < 2.5 angstroms).

Conclusion: VECTOR+ provides a robust, extensible approach for property-conditioned molecular design in low-data settings, effectively bridging contrastive learning and generative modeling for reproducible, AI-accelerated drug discovery with superior performance over benchmarks like JT-VAE and MolGPT.

Abstract: Efficiently steering generative models toward pharmacologically relevant regions of chemical space remains a major obstacle in molecular drug discovery under low-data regimes. We present VECTOR+: Valid-property-Enhanced Contrastive Learning for Targeted Optimization and Resampling, a framework that couples property-guided representation learning with controllable molecule generation. VECTOR+ applies to both regression and classification tasks and enables interpretable, data-efficient exploration of functional chemical space. We evaluate on two datasets: a curated PD-L1 inhibitor set (296 compounds with experimental $IC_{50}$ values) and a receptor kinase inhibitor set (2,056 molecules by binding mode). Despite limited training data, VECTOR+ generates novel, synthetically tractable candidates. Against PD-L1 (PDB 5J89), 100 of 8,374 generated molecules surpass a docking threshold of $-15.0$ kcal/mol, with the best scoring $-17.6$ kcal/mol compared to the top reference inhibitor ($-15.4$ kcal/mol). The best-performing molecules retain the conserved biphenyl pharmacophore while introducing novel motifs. Molecular dynamics (250 ns) confirm binding stability (ligand RMSD < $2.5$ angstroms). VECTOR+ generalizes to kinase inhibitors, producing compounds with stronger docking scores than established drugs such as brigatinib and sorafenib. Benchmarking against JT-VAE and MolGPT across docking, novelty, uniqueness, and Tanimoto similarity highlights the superior performance of our method. These results position our work as a robust, extensible approach for property-conditioned molecular design in low-data settings, bridging contrastive learning and generative modeling for reproducible, AI-accelerated discovery.

[808] DELTA: Variational Disentangled Learning for Privacy-Preserving Data Reprogramming

Arun Vignesh Malarkkan, Haoyue Bai, Anjali Kaushik, Yanjie Fu

Main category: cs.LG

TL;DR: DELTA framework for privacy-preserving data reprogramming that transforms features to maximize target prediction accuracy while minimizing sensitive attribute leakage using two-phase variational disentangled learning.

DetailsMotivation: Real-world domain data contains sensitive attributes subject to regulations (HIPAA, GDPR) and requires interpretable feature engineering, but existing methods focus on downstream performance risking privacy leakage.

Method: Two-phase framework: Phase I uses policy-guided RL to discover utility feature transformations; Phase II employs variational LSTM seq2seq with utility-privacy disentangled latent space and adversarial-causal regularization to suppress privacy signals.

Result: Experiments on eight datasets show DELTA improves predictive performance by ~9.3% and reduces privacy leakage by ~35%.

Conclusion: DELTA demonstrates robust, privacy-aware data transformation that effectively balances utility and privacy requirements in feature engineering.

Abstract: In real-world applications, domain data often contains identifiable or sensitive attributes, is subject to strict regulations (e.g., HIPAA, GDPR), and requires explicit data feature engineering for interpretability and transparency. Existing feature engineering primarily focuses on advancing downstream task performance, often risking privacy leakage. We generalize this learning task under such new requirements as Privacy-Preserving Data Reprogramming (PPDR): given a dataset, transforming features to maximize target attribute prediction accuracy while minimizing sensitive attribute prediction accuracy. PPDR poses challenges for existing systems: 1) generating high-utility feature transformations without being overwhelmed by a large search space, and 2) disentangling and eliminating sensitive information from utility-oriented features to reduce privacy inferability. To tackle these challenges, we propose DELTA, a two-phase variational disentangled generative learning framework. Phase I uses policy-guided reinforcement learning to discover feature transformations with downstream task utility, without any regard to privacy inferability. Phase II employs a variational LSTM seq2seq encoder-decoder with a utility-privacy disentangled latent space design and adversarial-causal disentanglement regularization to suppress privacy signals during feature generation. Experiments on eight datasets show DELTA improves predictive performance by ~9.3% and reduces privacy leakage by ~35%, demonstrating robust, privacy-aware data transformation.

[809] Robust Spatiotemporal Forecasting Using Adaptive Deep-Unfolded Variational Mode Decomposition

Osama Ahmad, Lukas Wesemann, Fabian Waschkowski, Zubair Khalid

Main category: cs.LG

TL;DR: MAGN transforms iterative variational mode decomposition into a trainable neural module for efficient spatiotemporal forecasting, achieving 85-95% error reduction and 250x speedup.

DetailsMotivation: Existing decomposition-integrated approaches like VMGCN suffer from computational inefficiency and manual hyperparameter tuning despite improving accuracy through signal decomposition.

Method: Proposes mode adaptive graph network (MAGN) with: (1) unfolded VMD module replacing iterative optimization with fixed-depth network, and (2) mode-specific learnable bandwidth constraints that adapt spatial heterogeneity and eliminate manual tuning.

Result: Achieves 85-95% reduction in prediction error over VMGCN and 250x decomposition time reduction on LargeST benchmark (6,902 sensors, 241M observations). Outperforms state-of-the-art baselines.

Conclusion: MAGN successfully addresses computational inefficiency and manual tuning limitations of previous decomposition methods while significantly improving spatiotemporal forecasting accuracy.

Abstract: Accurate spatiotemporal forecasting is critical for numerous complex systems but remains challenging due to complex volatility patterns and spectral entanglement in conventional graph neural networks (GNNs). While decomposition-integrated approaches like variational mode graph convolutional network (VMGCN) improve accuracy through signal decomposition, they suffer from computational inefficiency and manual hyperparameter tuning. To address these limitations, we propose the mode adaptive graph network (MAGN) that transforms iterative variational mode decomposition (VMD) into a trainable neural module. Our key innovations include (1) an unfolded VMD (UVMD) module that replaces iterative optimization with a fixed-depth network to reduce the decomposition time (by 250x for the LargeST benchmark), and (2) mode-specific learnable bandwidth constraints ({\alpha}k ) adapt spatial heterogeneity and eliminate manual tuning while preventing spectral overlap. Evaluated on the LargeST benchmark (6,902 sensors, 241M observations), MAGN achieves an 85-95% reduction in the prediction error over VMGCN and outperforms state-of-the-art baselines.

[810] Why Pool When You Can Flow? Active Learning with GFlowNets

Renfei Zhang, Mohit Pandey, Artem Cherkasov, Martin Ester

Main category: cs.LG

TL;DR: BALD-GFlowNet combines generative flow networks with active learning to enable scalable virtual screening by directly sampling molecules proportional to BALD reward instead of evaluating large unlabeled pools.

DetailsMotivation: Traditional pool-based active learning becomes computationally prohibitive when scaling to billion-sample molecular libraries in drug discovery, particularly with methods like BALD that require evaluating all unlabeled samples.

Method: Leverages Generative Flow Networks (GFlowNets) to directly sample objects in proportion to the BALD reward, replacing traditional pool-based acquisition with generative sampling.

Result: Achieves performance comparable to standard BALD baseline while generating more structurally diverse molecules, with scalability independent of unlabeled pool size.

Conclusion: BALD-GFlowNet offers a promising direction for efficient and scalable molecular discovery by circumventing computational bottlenecks of traditional active learning approaches.

Abstract: The scalability of pool-based active learning is limited by the computational cost of evaluating large unlabeled datasets, a challenge that is particularly acute in virtual screening for drug discovery. While active learning strategies such as Bayesian Active Learning by Disagreement (BALD) prioritize informative samples, it remains computationally intensive when scaled to libraries containing billions samples. In this work, we introduce BALD-GFlowNet, a generative active learning framework that circumvents this issue. Our method leverages Generative Flow Networks (GFlowNets) to directly sample objects in proportion to the BALD reward. By replacing traditional pool-based acquisition with generative sampling, BALD-GFlowNet achieves scalability that is independent of the size of the unlabeled pool. In our virtual screening experiment, we show that BALD-GFlowNet achieves a performance comparable to that of standard BALD baseline while generating more structurally diverse molecules, offering a promising direction for efficient and scalable molecular discovery.

[811] Task-Aware Adaptive Modulation: A Replay-Free and Resource-Efficient Approach For Continual Graph Learning

Jingtao Liu, Xinming Zhang

Main category: cs.LG

TL;DR: TAAM is a replay-free continual graph learning method that uses lightweight neural synapse modulators to dynamically control a frozen GNN backbone, achieving state-of-the-art performance without extensive pre-training or data replay.

DetailsMotivation: Current continual graph learning methods struggle with the stability-plasticity dilemma and require resource-heavy pre-training. The authors aim to overcome these challenges by modulating a frozen backbone rather than replaying data or fine-tuning the entire network.

Method: Propose Task-Aware Adaptive Modulation (TAAM) with Neural Synapse Modulators (NSM) that are trained and frozen for each task. Uses prototype-guided strategy: deep-copying similar past modulators for training and selecting relevant frozen NSMs for inference to perform node-attentive modulation of a frozen GNN backbone.

Result: TAAM comprehensively outperforms state-of-the-art methods across six GCIL benchmark datasets, demonstrating superior performance without replay or extensive pre-training.

Conclusion: TAAM provides an effective replay-free and resource-efficient solution for continual graph learning by dynamically modulating internal computational flow of frozen backbones, successfully navigating the stability-plasticity dilemma.

Abstract: Continual Graph Learning(CGL)focuses on acquiring new knowledge while retaining previously learned information, essential for real-world graph applications. Current methods grapple with two main issues:1) The Stability-Plasticity Dilemma: Replay-based methods often create an imbalance between the Dilemma, while incurring significant storage costs.2) The Resource-Heavy Pre-training: Leading replay-free methods critically depend on extensively pre-trained backbones, this reliance imposes a substantial resource burden.In this paper, we argue that the key to overcoming these challenges lies not in replaying data or fine-tuning the entire network, but in dynamically modulating the internal computational flow of a frozen backbone. We posit that lightweight, task-specific modules can effectively steer a GNN’s reasoning process. Motivated by this insight, we propose Task-Aware Adaptive Modulation(TAAM), a replay-free, resource-efficient approach that charts a new path for navigating the stability-plasticity dilemma. TAAM’s core is its Neural Synapse Modulators(NSM), which are trained and then frozen for each task to store expert knowledge. A pivotal prototype-guided strategy governs these modulators: 1) For training, it initializes a new NSM by deep-copying from a similar past modulator to boost knowledge transfer. 2) For inference, it selects the most relevant frozen NSM for each task. These NSMs insert into a frozen GNN backbone to perform fine-grained, node-attentive modulation of its internal flow-different from the static perturbations of prior methods. Extensive experiments show that TAAM comprehensively outperforms state-of-the-art methods across six GCIL benchmark datasets. The code will be released upon acceptance of the paper.

[812] Attribute Fusion-based Classifier on Framework of Belief Structure

Qiying Hu, Yingying Liang, Qianli Zhou, Witold Pedrycz

Main category: cs.LG

TL;DR: Enhanced DST-based classifier with improved membership modeling and BPA transformation, achieving 4.84% accuracy improvement over existing methods.

DetailsMotivation: Traditional DST classifiers suffer from oversimplified membership functions and limited belief structure exploitation, reducing effectiveness in complex scenarios.

Method: Selective modeling using Gaussian/GMM for membership functions, novel BPA transformation from possibility distributions, and enhanced evidential K-NN classifier.

Result: Outperforms best existing evidential classifier with 4.84% average accuracy improvement and maintains low variance.

Conclusion: Proposed approach provides superior effectiveness and robustness for uncertainty modeling in multi-attribute classification tasks.

Abstract: Dempster-Shafer Theory (DST) provides a powerful framework for modeling uncertainty and has been widely applied to multi-attribute classification tasks. However, traditional DST-based attribute fusion-based classifiers suffer from oversimplified membership function modeling and limited exploitation of the belief structure brought by basic probability assignment (BPA), reducing their effectiveness in complex real-world scenarios. This paper presents an enhanced attribute fusion-based classifier that addresses these limitations through two key innovations. First, we adopt a selective modeling strategy that utilizes both single Gaussian and Gaussian Mixture Models (GMMs) for membership function construction, with model selection guided by cross-validation and a tailored evaluation metric. Second, we introduce a novel method to transform the possibility distribution into a BPA by combining simple BPAs derived from normalized possibility distributions, enabling a much richer and more flexible representation of uncertain information. Furthermore, we apply the belief structure-based BPA generation method to the evidential K-Nearest Neighbors classifier, enhancing its ability to incorporate uncertainty information into decision-making. Comprehensive experiments on benchmark datasets are conducted to evaluate the performance of the proposed attribute fusion-based classifier and the enhanced evidential K-Nearest Neighbors classifier in comparison with both evidential classifiers and conventional machine learning classifiers. The results demonstrate that our proposed classifier outperforms the best existing evidential classifier, achieving an average accuracy improvement of 4.84%, while maintaining low variance, thus confirming its superior effectiveness and robustness.

[813] Flow Matters: Directional and Expressive GNNs for Heterophilic Graphs

Arman Gupta, Govind Waghmare, Gaurav Oberoi, Nitish Srivastava

Main category: cs.LG

TL;DR: This paper proposes two GNN architectures that combine polynomial expressiveness with edge directionality for improved node classification in heterophilic graphs, achieving state-of-the-art results.

DetailsMotivation: Conventional GNNs struggle with heterophilic graphs where neighboring nodes belong to different classes, and prior work suggests edge directionality and polynomial expressiveness could help but haven't been combined.

Method: Two architectures: (1) Poly - polynomially expressive GAT baseline, and (2) Dir-Poly - direction-aware variant that separately aggregates incoming and outgoing edges. Both learn permutation-equivariant high-degree polynomials over input features with no added time complexity.

Result: Experiments on five heterophilic datasets show Poly consistently outperforms existing baselines, and Dir-Poly offers additional gains on graphs with inherent directionality (e.g., Roman Empire), achieving state-of-the-art results. Artificial directionality on undirected graphs doesn’t always help.

Conclusion: The findings highlight the complementary roles of edge direction and expressive feature modeling in heterophilic graph learning, with benefits being context-dependent on graph directionality.

Abstract: In heterophilic graphs, where neighboring nodes often belong to different classes, conventional Graph Neural Networks (GNNs) struggle due to their reliance on local homophilous neighborhoods. Prior studies suggest that modeling edge directionality in such graphs can increase effective homophily and improve classification performance. Simultaneously, recent work on polynomially expressive GNNs shows promise in capturing higher-order interactions among features. In this work, we study the combined effect of edge directionality and expressive message passing on node classification in heterophilic graphs. Specifically, we propose two architectures: (1) a polynomially expressive GAT baseline (Poly), and (2) a direction-aware variant (Dir-Poly) that separately aggregates incoming and outgoing edges. Both models are designed to learn permutation-equivariant high-degree polynomials over input features, while remaining scalable with no added time complexity. Experiments on five benchmark heterophilic datasets show that our Poly model consistently outperforms existing baselines, and that Dir-Poly offers additional gains on graphs with inherent directionality (e.g., Roman Empire), achieving state-of-the-art results. Interestingly, on undirected graphs, introducing artificial directionality does not always help, suggesting that the benefit of directional message passing is context-dependent. Our findings highlight the complementary roles of edge direction and expressive feature modeling in heterophilic graph learning.

[814] ProCause: Generating Counterfactual Outcomes to Evaluate Prescriptive Process Monitoring Methods

Jakob De Moor, Hans Weytjens, Johannes De Smedt

Main category: cs.LG

TL;DR: ProCause is a new generative approach for evaluating Prescriptive Process Monitoring methods that addresses limitations of RealCause by supporting sequential models and multiple CI architectures, with ensemble models proving more reliable than single architectures.

DetailsMotivation: Existing evaluation methods for Prescriptive Process Monitoring (PresPM) lack ground-truth outcomes and RealCause overlooks temporal dependencies while being limited to a single CI model architecture (TARNet).

Method: ProCause introduces a generative approach that supports both sequential (e.g., LSTMs) and non-sequential models while integrating multiple CI architectures including S-Learner, T-Learner, TARNet, and an ensemble approach.

Result: Research using a simulator with known ground truths shows TARNet is not always optimal, ensemble models offer more consistent reliability, and LSTMs show potential for improved evaluations when temporal dependencies exist.

Conclusion: ProCause provides a more reliable evaluation framework for PresPM methods, validated through both simulation with known ground truths and real-world data analysis.

Abstract: Prescriptive Process Monitoring (PresPM) is the subfield of Process Mining that focuses on optimizing processes through real-time interventions based on event log data. Evaluating PresPM methods is challenging due to the lack of ground-truth outcomes for all intervention actions in datasets. A generative deep learning approach from the field of Causal Inference (CI), RealCause, has been commonly used to estimate the outcomes for proposed intervention actions to evaluate a new policy. However, RealCause overlooks the temporal dependencies in process data, and relies on a single CI model architecture, TARNet, limiting its effectiveness. To address both shortcomings, we introduce ProCause, a generative approach that supports both sequential (e.g., LSTMs) and non-sequential models while integrating multiple CI architectures (S-Learner, T-Learner, TARNet, and an ensemble). Our research using a simulator with known ground truths reveals that TARNet is not always the best choice; instead, an ensemble of models offers more consistent reliability, and leveraging LSTMs shows potential for improved evaluations when temporal dependencies are present. We further validate ProCause’s practical effectiveness through a real-world data analysis, ensuring a more reliable evaluation of PresPM methods.

Noorain Mukhtiar, Adnan Mahmood, Quan Z. Sheng

Main category: cs.LG

TL;DR: Survey paper exploring fairness issues in Federated Learning systems, covering sources of bias, mitigation techniques, evaluation metrics, and future research directions.

DetailsMotivation: Federated Learning enables collaborative model training while preserving privacy, but suffers from fairness concerns due to various sources of heterogeneity that cause biases, skewed predictions, reduced accuracy, and inefficient convergence.

Method: Comprehensive survey approach analyzing diverse sources of bias (data, client, model biases), reviewing state-of-the-art mitigation techniques, discussing theoretical foundations, and examining evaluation metrics for measuring fairness quantitatively.

Result: Provides a thorough overview of fairness notions, technical aspects, and existing techniques in FL systems, highlighting both strengths and limitations of current approaches to address bias disparities.

Conclusion: Identifies exciting open research directions for achieving fairer FL frameworks and establishes a strong foundation for future research in this critical area of privacy-preserving collaborative learning.

Abstract: At the intersection of the cutting-edge technologies and privacy concerns, Federated Learning (FL) with its distributed architecture, stands at the forefront in a bid to facilitate collaborative model training across multiple clients while preserving data privacy. However, the applicability of FL systems is hindered by fairness concerns arising from numerous sources of heterogeneity that can result in biases and undermine a system’s effectiveness, with skewed predictions, reduced accuracy, and inefficient model convergence. This survey thus explores the diverse sources of bias, including but not limited to, data, client, and model biases, and thoroughly discusses the strengths and limitations inherited within the array of the state-of-the-art techniques utilized in the literature to mitigate such disparities in the FL training process. We delineate a comprehensive overview of the several notions, theoretical underpinnings, and technical aspects associated with fairness and their adoption in FL-based multidisciplinary environments. Furthermore, we examine salient evaluation metrics leveraged to measure fairness quantitatively. Finally, we envisage exciting open research directions that have the potential to drive future advancements in achieving fairer FL frameworks, in turn, offering a strong foundation for future research in this pivotal area.

[816] XAI-Driven Machine Learning System for Driving Style Recognition and Personalized Recommendations

Feriel Amel Sellal, Ahmed Ayoub Bellachia, Meryem Malak Dif, Enguerrand De Rautlin De La Roy, Mouhamed Amine Bouchiha, Yacine Ghamri-Doudane

Main category: cs.LG

TL;DR: Proposes interpretable ML models (RF, XGBoost, SVM) for driving style classification that match DL performance (0.92 accuracy) while providing transparency through SHAP explanations.

DetailsMotivation: Address the black-box nature of deep learning models in driving style classification by developing interpretable ML approaches that maintain high accuracy while enabling trust and transparency for real-world deployment in automotive AI systems.

Method: Uses machine learning techniques (Random Forest, XGBoost, SVM) with SHAP explainability on a new CARLA-Drive dataset for three-class driving style classification, focusing on efficient, lightweight, and interpretable models.

Result: Achieved 0.92 accuracy with both Random Forest and XGBoost classifiers, matching deep learning model performance while providing interpretability through SHAP-based personalized driving recommendations.

Conclusion: The proposed ML-based approach successfully balances high accuracy with interpretability, offering transparent and practical solutions for real-world intelligent transportation systems without sacrificing performance compared to black-box DL models.

Abstract: Artificial intelligence (AI) is increasingly used in the automotive industry for applications such as driving style classification, which aims to improve road safety, efficiency, and personalize user experiences. While deep learning (DL) models, such as Long Short-Term Memory (LSTM) networks, excel at this task, their black-box nature limits interpretability and trust. This paper proposes a machine learning (ML)-based method that balances high accuracy with interpretability. We introduce a high-quality dataset, CARLA-Drive, and leverage ML techniques like Random Forest (RF), Gradient Boosting (XGBoost), and Support Vector Machine (SVM), which are efficient, lightweight, and interpretable. In addition, we apply the SHAP (Shapley Additive Explanations) explainability technique to provide personalized recommendations for safer driving. Achieving an accuracy of 0.92 on a three-class classification task with both RF and XGBoost classifiers, our approach matches DL models in performance while offering transparency and practicality for real-world deployment in intelligent transportation systems.

[817] Crystal Structure Prediction with a Geometric Permutation-Invariant Loss Function

Emmanuel Jehanno, Romain Menegaux, Julien Mairal, Sergei Grudinin

Main category: cs.LG

TL;DR: SinkFast - a novel method for molecular crystal structure prediction using differentiable linear assignment with Sinkhorn algorithm, outperforming complex flow-matching approaches on COD-Cluster17 benchmark.

DetailsMotivation: Crystalline structure prediction remains challenging, especially for organic materials. Existing methods rely on computationally expensive iterative flow-matching approaches, creating a need for more efficient and accurate solutions.

Method: Proposed a novel loss function that captures key geometric molecular properties while maintaining permutation invariance. Uses a differentiable linear assignment scheme based on the Sinkhorn algorithm.

Result: Significantly outperforms more complex flow-matching approaches on the COD-Cluster17 benchmark (curated subset of Crystallography Open Database). Even simple regression with SinkFast achieves superior results.

Conclusion: SinkFast provides an effective and efficient alternative to computationally expensive flow-matching methods for molecular crystal structure prediction, demonstrating strong performance on standard benchmarks.

Abstract: Crystalline structure prediction remains an open challenge in materials design. Despite recent advances in computational materials science, accurately predicting the three-dimensional crystal structures of organic materials–an essential first step for designing materials with targeted properties–remains elusive. In this work, we address the problem of molecular assembly, where a set $\mathcal{S}$ of identical rigid molecules is packed to form a crystalline structure. Existing state-of-the-art models typically rely on computationally expensive, iterative flow-matching approaches. We propose a novel loss function that correctly captures key geometric molecular properties while maintaining permutation invariance over $\mathcal{S}$. We achieve this via a differentiable linear assignment scheme based on the Sinkhorn algorithm. Remarkably, we show that even a simple regression using our method {\em SinkFast} significantly outperforms more complex flow-matching approaches on the COD-Cluster17 benchmark, a curated subset of the Crystallography Open Database (COD).

[818] Causal SHAP: Feature Attribution with Dependency Awareness through Causal Discovery

Woon Yee Ng, Li Rong Wang, Siyuan Liu, Xiuyi Fan

Main category: cs.LG

TL;DR: Causal SHAP integrates causal relationships into feature attribution to address SHAP’s limitation of confusing correlation with causality, providing more accurate explanations for ML predictions.

DetailsMotivation: SHAP explanations fail to differentiate between causality and correlation, which is problematic in high-stakes domains like healthcare where understanding true causal relationships is critical for informed decision-making.

Method: Proposes Causal SHAP framework that combines Peter-Clark (PC) algorithm for causal discovery and Intervention Calculus when the DAG is Absent (IDA) algorithm for causal strength quantification, while preserving SHAP’s desirable properties.

Result: Causal SHAP reduces attribution scores for features that are merely correlated with the target, as validated through experiments on both synthetic and real-world datasets.

Conclusion: Provides a practical framework for causal-aware model explanations in Explainable AI, particularly valuable in domains like healthcare where understanding true causal relationships is essential.

Abstract: Explaining machine learning (ML) predictions has become crucial as ML models are increasingly deployed in high-stakes domains such as healthcare. While SHapley Additive exPlanations (SHAP) is widely used for model interpretability, it fails to differentiate between causality and correlation, often misattributing feature importance when features are highly correlated. We propose Causal SHAP, a novel framework that integrates causal relationships into feature attribution while preserving many desirable properties of SHAP. By combining the Peter-Clark (PC) algorithm for causal discovery and the Intervention Calculus when the DAG is Absent (IDA) algorithm for causal strength quantification, our approach addresses the weakness of SHAP. Specifically, Causal SHAP reduces attribution scores for features that are merely correlated with the target, as validated through experiments on both synthetic and real-world datasets. This study contributes to the field of Explainable AI (XAI) by providing a practical framework for causal-aware model explanations. Our approach is particularly valuable in domains such as healthcare, where understanding true causal relationships is critical for informed decision-making.

[819] Predicting Multi-Type Talented Students in Secondary School Using Semi-Supervised Machine Learning

Xinzhe Zheng, Zhen-Qun Yang, Jiannong Cao, Jiabei Cheng

Main category: cs.LG

TL;DR: TalentPredictor is a semi-supervised multi-modal neural network that combines Transformer, LSTM, and ANN to predict 7 talent types in secondary students using offline educational data, achieving 90.8% accuracy.

DetailsMotivation: Traditional talent identification relies on manual processes and academic focus, missing non-academic talents and early intervention opportunities in secondary education.

Method: Semi-supervised multi-modal neural network combining Transformer, LSTM, and ANN architectures, using clustering of award records and feature extraction from diverse learning behaviors of 1,041 secondary students.

Result: Achieved high prediction accuracy with 0.908 classification accuracy and 0.908 ROCAUC across seven talent categories (academic, sport, art, leadership, service, technology, others).

Conclusion: Machine learning can effectively identify diverse student talents early in development, overcoming limitations of traditional methods and enabling timely intervention.

Abstract: Talent identification plays a critical role in promoting student development. However, traditional approaches often rely on manual processes or focus narrowly on academic achievement, and typically delaying intervention until the higher education stage. This oversight overlooks diverse non-academic talents and misses opportunities for early intervention. To address this gap, this study introduces TalentPredictor, a novel semi-supervised multi-modal neural network that combines Transformer, LSTM, and ANN architectures. This model is designed to predict seven different talent types–academic, sport, art, leadership, service, technology, and others–in secondary school students within an offline educational setting. Drawing on existing offline educational data from 1,041 local secondary students, TalentPredictor overcomes the limitations of traditional talent identification methods. By clustering various award records into talent categories and extracting features from students’ diverse learning behaviors, it achieves high prediction accuracy (0.908 classification accuracy, 0.908 ROCAUC). This demonstrates the potential of machine learning to identify diverse talents early in student development.

[820] Tabular Diffusion Counterfactual Explanations

Wei Zhang, Brian Barr, John Paisley

Main category: cs.LG

TL;DR: Novel guided reverse process using Gumbel-softmax approximation for counterfactual explanations on tabular data, outperforming baselines in interpretability and realism.

DetailsMotivation: Existing counterfactual explanation methods focus on computer vision problems, leaving a gap for tabular data typical in finance and social sciences that requires specialized approaches for categorical features.

Method: Proposed a guided reverse process for categorical features based on Gumbel-softmax distribution approximation, with theoretical analysis of temperature parameter τ effects.

Result: Experiments on credit lending and other tabular datasets show superior performance in interpretability, diversity, instability, and validity metrics compared to baseline methods.

Conclusion: The approach produces robust and realistic counterfactual explanations for tabular data, effectively addressing the gap in counterfactual explanation methods for categorical features in financial and social science applications.

Abstract: Counterfactual explanations methods provide an important tool in the field of {interpretable machine learning}. Recent advances in this direction have focused on diffusion models to explain a deep classifier. However, these techniques have predominantly focused on problems in computer vision. In this paper, we focus on tabular data typical in finance and the social sciences and propose a novel guided reverse process for categorical features based on an approximation to the Gumbel-softmax distribution. Furthermore, we study the effect of the temperature $\tau$ and derive a theoretical bound between the Gumbel-softmax distribution and our proposed approximated distribution. We perform experiments on several large-scale credit lending and other tabular datasets, assessing their performance in terms of the quantitative measures of interpretability, diversity, instability, and validity. These results indicate that our approach outperforms popular baseline methods, producing robust and realistic counterfactual explanations.

[821] An Explainable Gaussian Process Auto-encoder for Tabular Data

Wei Zhang, Brian Barr, John Paisley

Main category: cs.LG

TL;DR: Proposes a novel Gaussian process-based autoencoder for generating counterfactual explanations with fewer parameters and better in-distribution samples.

DetailsMotivation: Address the need for explainable machine learning in high-stakes scenarios by improving counterfactual explanation methods that leverage generative models like autoencoders.

Method: Uses Gaussian process to construct autoencoder architecture, introduces novel density estimator for in-distribution sample search, and develops algorithm for optimal regularization rate selection.

Result: Experiments on large-scale tabular datasets show the method generates diversified and in-distribution counterfactual samples, outperforming other autoencoder-based approaches.

Conclusion: The proposed Gaussian process-based autoencoder provides an effective solution for counterfactual explanation generation with reduced overfitting and improved sample quality.

Abstract: Explainable machine learning has attracted much interest in the community where the stakes are high. Counterfactual explanations methods have become an important tool in explaining a black-box model. The recent advances have leveraged the power of generative models such as an autoencoder. In this paper, we propose a novel method using a Gaussian process to construct the auto-encoder architecture for generating counterfactual samples. The resulting model requires fewer learnable parameters and thus is less prone to overfitting. We also introduce a novel density estimator that allows for searching for in-distribution samples. Furthermore, we introduce an algorithm for selecting the optimal regularization rate on density estimator while searching for counterfactuals. We experiment with our method in several large-scale tabular datasets and compare with other auto-encoder-based methods. The results show that our method is capable of generating diversified and in-distribution counterfactual samples.

[822] DTRNet: Dynamic Token Routing Network to Reduce Quadratic Costs in Transformers

Aman Sharma, Saeed Najafi, Parsa Farinneya, Benyamin Jamialahmadi, Marzieh S. Tahaei, Yuhe Fan, Mehdi Rezagholizadeh, Boxing Chen, Aref Jafari

Main category: cs.LG

TL;DR: DTRNet is a Transformer variant that dynamically routes only ~10% of tokens through quadratic attention per layer while maintaining full Transformer performance, achieving significant computational savings especially for long sequences.

DetailsMotivation: Standard Transformers apply quadratic self-attention to every token at every layer, making them computationally expensive. There's a need for more efficient architectures that maintain performance while reducing computation costs.

Method: DTRNet uses dynamic token routing to selectively apply quadratic attention to only ~10% of tokens per layer, while other tokens receive lightweight linear updates. It preserves the MLP module and ensures every token is explicitly updated.

Result: DTRNet maintains performance comparable to full Transformers while routing fewer tokens to full attention. It outperforms routing-based layer skipping methods (MoD, D-LLM) in both accuracy and memory at matched FLOPs, with efficiency gains scaling with sequence length.

Conclusion: DTRNet provides a simple, efficient, and scalable alternative to standard Transformers by decoupling token updates from attention mixing, substantially reducing the quadratic computation share while maintaining performance.

Abstract: Transformers achieve state-of-the-art results across many tasks, but their uniform application of quadratic self-attention to every token at every layer makes them computationally expensive. We introduce DTRNet (Dynamic Token Routing Network), an improved Transformer architecture that allows tokens to dynamically skip the quadratic cost of cross-token mixing while still receiving lightweight linear updates. By preserving the MLP module and reducing the attention cost for most tokens to linear, DTRNet ensures that every token is explicitly updated while significantly lowering overall computation. This design offers an efficient and effective alternative to standard dense attention. Once trained, DTRNet blocks routes only ~10% of tokens through attention at each layer while maintaining performance comparable to a full Transformer. It consistently outperforms routing-based layer skipping methods such as MoD and D-LLM in both accuracy and memory at matched FLOPs, while routing fewer tokens to full attention. Its efficiency gains, scales with sequence length, offering significant reduction in FLOPs for long-context inputs. By decoupling token updates from attention mixing, DTRNet substantially reduces the quadratic share of computation, providing a simple, efficient, and scalable alternative to Transformers.

[823] Superposition in Graph Neural Networks

Lukas Pertl, Han Xuanyuan, Pietro Liò

Main category: cs.LG

TL;DR: Analysis of superposition in GNNs shows how width, pooling, and activations affect feature geometry and interpretability.

DetailsMotivation: Interpreting GNNs is challenging due to message passing mixing signals and lack of alignment with human concepts. The paper aims to study superposition directly in GNN latent spaces to understand feature geometry.

Method: Used controlled experiments with unambiguous graph concepts, extracting features as class-conditional centroids (graph level) and linear-probe directions (node level), then analyzed geometry with basis-invariant diagnostics across GCN/GIN/GAT architectures.

Result: Found that increasing width produces phase patterns in overlap, topology imprints overlap onto node features, pooling remixes features into task-aligned axes, sharper pooling increases axis alignment and reduces channel sharing, and shallow models can settle into metastable low-rank embeddings.

Conclusion: Results connect representational geometry with design choices (width, pooling, activations) and suggest practical approaches for more interpretable GNNs.

Abstract: Interpreting graph neural networks (GNNs) is difficult because message passing mixes signals and internal channels rarely align with human concepts. We study superposition, the sharing of directions by multiple features, directly in the latent space of GNNs. Using controlled experiments with unambiguous graph concepts, we extract features as (i) class-conditional centroids at the graph level and (ii) linear-probe directions at the node level, and then analyze their geometry with simple basis-invariant diagnostics. Across GCN/GIN/GAT we find: increasing width produces a phase pattern in overlap; topology imprints overlap onto node-level features that pooling partially remixes into task-aligned graph axes; sharper pooling increases axis alignment and reduces channel sharing; and shallow models can settle into metastable low-rank embeddings. These results connect representational geometry with concrete design choices (width, pooling, and final-layer activations) and suggest practical approaches for more interpretable GNNs.

[824] SCOUT: Toward Sub-Quadratic Attention via Segment Compression for Optimized Utility in Transformers

Aref Jafari, Yuhe Fan, Benyamin Jamialahmadi, Parsa Farinneya, Boxing Chen, Marzieh S. Tahaei

Main category: cs.LG

TL;DR: SCOUT is a hybrid architecture that compresses tokens locally within segments and applies attention only over compressed representations, achieving near full-attention performance with sub-quadratic complexity.

DetailsMotivation: Transformers have quadratic attention complexity limiting scalability to long sequences, while linear models like Mamba and SWA risk performance degradation on long sequences due to inability to retain detailed information from distant tokens.

Method: Each token embedding is enriched via linear local mixer (Mamba or SWA), then tokens sparsely attend to a small number of compressed checkpoint tokens that summarize input history, rather than all previous tokens.

Result: SCOUT outperforms strong long-sequence baselines under same computational budget, matches full-attention Transformers on language modeling and reasoning tasks at 400M and 1.3B scales, with higher end-to-end throughput than SOTA models.

Conclusion: SCOUT retains much of full attention’s expressivity while substantially reducing computational and memory costs, achieving sub-quadratic growth rate that is far more scalable than full Transformers.

Abstract: Transformers have demonstrated strong performance across a wide range of sequence modeling tasks, but their quadratic attention complexity limits scalability to long sequences. Linear models such as Mamba and sliding-window attention (SWA) address this by mixing tokens through recurrent or localized operations with fixed-size memory, achieving efficient inference. However, these methods risk degrading performance on long sequences due to their inability to retain detailed information from distant tokens. We propose SCOUT (Segment Compression for Optimized Utility in Transformers), a hybrid architecture that compresses tokens locally within fixed-size segments and applies attention only over these compressed representations. Each token embedding is first enriched via a linear local mixer, Mamba or SWA, that integrates recent context. Then, instead of attending to all previous tokens, each token sparsely attends to a small number of compressed checkpoint tokens that summarize the input history. This design retains much of the expressivity of full attention while substantially reducing the computational and memory cost. By attending to compressed history rather than all previous tokens, SCOUT incurs slightly higher memory than purely linear models, but its growth rate remains sub-quadratic and far more scalable than that of full Transformers. We analyze SCOUT’s computational and memory efficiency and evaluate it empirically on long-context language modeling and reasoning tasks. SCOUT with both Mamba and SWA mixers outperforms strong long-sequence baselines under the same computational budget, matches full-attention Transformers on language modeling and common-sense reasoning tasks at 400M and 1.3B scales. Moreover, our SCOUT achieves higher end-to-end throughput than SOTA models, while delivering comparable results on long sequence benchmarks.

[825] ART: Adaptive Resampling-based Training for Imbalanced Classification

Arjun Basandrai, Shourya Jain, K. Ilanthenral

Main category: cs.LG

TL;DR: ART is an adaptive resampling method that dynamically adjusts training data distribution based on class-wise performance metrics (macro F1 scores), outperforming traditional static resampling methods on imbalanced classification tasks.

DetailsMotivation: Traditional resampling methods use fixed distributions that ignore changes in class-wise learning difficulty during training, which limits overall model performance on imbalanced datasets.

Method: Periodically updates training data distribution using class-wise macro F1 scores computed at fixed intervals. Adapts at class level rather than instance level to avoid noise and outlier sensitivity, incrementally shifting attention to underperforming classes.

Result: Consistently outperforms both resampling-based (SMOTE, NearMiss) and algorithm-level methods on binary and multi-class tasks. Achieves average 2.64 percentage point improvement in macro F1 across tabular datasets, with statistically significant gains (p < 0.05). Performs well across diverse benchmarks including Pima Indians Diabetes and Yeast datasets.

Conclusion: ART provides a reliable and consistently strong solution for imbalanced classification, delivering the strongest macro F1 performance across various tasks unlike existing methods whose performance varies by application.

Abstract: Traditional resampling methods for handling class imbalance typically uses fixed distributions, undersampling the majority or oversampling the minority. These static strategies ignore changes in class-wise learning difficulty, which can limit the overall performance of the model. This paper proposes an Adaptive Resampling-based Training (ART) method that periodically updates the distribution of the training data based on the class-wise performance of the model. Specifically, ART uses class-wise macro F1 scores, computed at fixed intervals, to determine the degree of resampling to be performed. Unlike instance-level difficulty modeling, which is noisy and outlier-sensitive, ART adapts at the class level. This allows the model to incrementally shift its attention towards underperforming classes in a way that better aligns with the optimization objective. Results on diverse benchmarks, including Pima Indians Diabetes and Yeast dataset demonstrate that ART consistently outperforms both resampling-based and algorithm-level methods, including Synthetic Minority Oversampling Technique (SMOTE), NearMiss Undersampling, and Cost-sensitive Learning on binary as well as multi-class classification tasks with varying degrees of imbalance. In most settings, these improvements are statistically significant. On tabular datasets, gains are significant under paired t-tests and Wilcoxon tests (p < 0.05), while results on text and image tasks remain favorable. Compared to training on the original imbalanced data, ART improves macro F1 by an average of 2.64 percentage points across all tested tabular datasets. Unlike existing methods, whose performance varies by task, ART consistently delivers the strongest macro F1, making it a reliable choice for imbalanced classification.

[826] Online Decentralized Federated Multi-task Learning With Trustworthiness in Cyber-Physical Systems

Olusola Odeyomi, Sofiat Olaosebikan, Ajibuwa Opeyemi, Oluwadoyinsola Ige

Main category: cs.LG

TL;DR: Online decentralized federated multi-task learning algorithm that provides model personalization and Byzantine resilience even when Byzantine clients dominate honest clients, using cyber-physical properties for trust assignment.

DetailsMotivation: Address challenges of model personalization in federated learning with high data heterogeneity, time-varying distributions, and Byzantine clients that can dominate the system in real-world applications like autonomous systems.

Method: Develops an online decentralized federated multi-task learning algorithm that leverages cyber-physical properties (e.g., received signal strength or side information) to assign trust probabilities to local models from neighbors in each iteration.

Result: Simulation results show the proposed algorithm performs close to a Byzantine-free setting, demonstrating effective resilience against dominating Byzantine clients.

Conclusion: The approach successfully extends multi-task learning to online decentralized federated learning with strong Byzantine resilience, enabling practical deployment in real-world systems where malicious clients may outnumber honest ones.

Abstract: Multi-task learning is an effective way to address the challenge of model personalization caused by high data heterogeneity in federated learning. However, extending multi-task learning to the online decentralized federated learning setting is yet to be explored. The online decentralized federated learning setting considers many real-world applications of federated learning, such as autonomous systems, where clients communicate peer-to-peer and the data distribution of each client is time-varying. A more serious problem in real-world applications of federated learning is the presence of Byzantine clients. Byzantine-resilient approaches used in federated learning work only when the number of Byzantine clients is less than one-half the total number of clients. Yet, it is difficult to put a limit on the number of Byzantine clients within a system in reality. However, recent work in robotics shows that it is possible to exploit cyber-physical properties of a system to predict clients’ behavior and assign a trust probability to received signals. This can help to achieve resiliency in the presence of a dominating number of Byzantine clients. Therefore, in this paper, we develop an online decentralized federated multi-task learning algorithm to provide model personalization and resiliency when the number of Byzantine clients dominates the number of honest clients. Our proposed algorithm leverages cyber-physical properties, such as the received signal strength in wireless systems or side information, to assign a trust probability to local models received from neighbors in each iteration. Our simulation results show that the proposed algorithm performs close to a Byzantine-free setting.

[827] MEPT: Mixture of Expert Prompt Tuning as a Manifold Mapper

Runjia Zeng, Guangyan Sun, Qifan Wang, Tong Geng, Sohail Dianat, Xiaotian Han, Raghuveer Rao, Xueling Zhang, Cheng Han, Lifu Huang, Dongfang Liu

Main category: cs.LG

TL;DR: MEPT is a novel Mixture of Expert Prompt Tuning framework that dynamically activates appropriate neural pathways to adapt to diverse data distributions, outperforming SOTA methods on SuperGLUE with 1.94% accuracy improvement and 79.25% prompt reduction.

DetailsMotivation: Current fine-tuning approaches have rigid parameter spaces that limit their ability to dynamically activate appropriate neural pathways for adapting to diverse and evolving data distributions.

Method: Proposes Mixture of Expert Prompt Tuning (MEPT) that integrates multiple prompt experts within a Mixture of Experts architecture to adaptively learn diverse and non-stationary data distributions.

Result: Outperforms state-of-the-art parameter efficient baselines on SuperGLUE with 1.94% mean accuracy improvement and 79.25% reduction in activated prompts.

Conclusion: MEPT provides an effective and efficient manifold-mapping framework supported by theoretical insights from manifold learning and validated through neural activation pathway visualization.

Abstract: Considering deep neural networks as manifold mappers, the pretrain-then-fine-tune paradigm can be interpreted as a two-stage process: pretrain establishes a broad knowledge base, and fine-tune adjusts the model parameters to activate specific neural pathways to align with the target manifold. Although prior fine-tuning approaches demonstrate success, their rigid parameter space limits their ability to dynamically activate appropriate neural pathways, rendering them ill-equipped to adapt flexibly to the diverse and evolving data distributions. In light of this view, we propose a novel approach, Mixture of Expert Prompt Tuning (MEPT), as an effective and efficient manifold-mapping framework. MEPT leverages the Mixture of Experts architecture by integrating multiple prompt experts to adaptively learn diverse and non-stationary data distributions. Empirical evaluations demonstrate that MEPT outperforms several state-of-the-art parameter efficient baselines on SuperGLUE, achieving notable improvements in mean accuracy (e.g., 1.94%) while significantly reducing activated prompts by 79.25%. The effectiveness of MEPT is further supported by theoretical insights from manifold learning and validated through neural activation pathway visualization results. Our code is avaliable at https://github.com/runtsang/MEPT.

[828] Any-Order Flexible Length Masked Diffusion

Jaeyeon Kim, Lee Cheuk-Kit, Carles Domingo-Enrich, Yilun Du, Sham Kakade, Timothy Ngotiaoco, Sitan Chen, Michael Albergo

Main category: cs.LG

TL;DR: FlexMDMs extend masked diffusion models to support flexible-length sequence generation while maintaining parallel inference capabilities, achieving better performance on tasks requiring variable-length outputs.

DetailsMotivation: Masked diffusion models (MDMs) are limited to fixed-length generations and don't support token insertions, which restricts their applicability to tasks requiring flexible sequence lengths.

Method: FlexMDMs extend the stochastic interpolant framework to generate sequences by inserting mask tokens and unmasking them, enabling modeling of variable-length sequences while retaining any-order inference flexibility.

Result: FlexMDMs match MDMs in perplexity while better modeling length statistics, achieve ≈60% higher success rate on maze planning, and enable efficient fine-tuning of existing models (3 days on 16 H100s) with significant performance improvements on math (58%→67%) and code infilling (52%→65%).

Conclusion: FlexMDMs successfully overcome the fixed-length limitation of MDMs while preserving their parallel inference advantages, demonstrating strong performance across various tasks requiring flexible sequence generation.

Abstract: Masked diffusion models (MDMs) have recently emerged as a promising alternative to autoregressive models over discrete domains. MDMs generate sequences in an any-order, parallel fashion, enabling fast inference and strong performance on non-causal tasks. However, a crucial limitation is that they do not support token insertions and are thus limited to fixed-length generations. To this end, we introduce Flexible Masked Diffusion Models (FlexMDMs), a discrete diffusion paradigm that simultaneously can model sequences of flexible length while provably retaining MDMs’ flexibility of any-order inference. Grounded in an extension of the stochastic interpolant framework, FlexMDMs generate sequences by inserting mask tokens and unmasking them. Empirically, we show that FlexMDMs match MDMs in perplexity while modeling length statistics with much higher fidelity. On a synthetic maze planning task, they achieve $\approx 60 %$ higher success rate than MDM baselines. Finally, we show pretrained MDMs can easily be retrofitted into FlexMDMs: on 16 H100s, it takes only three days to fine-tune LLaDA-8B into a FlexMDM, achieving superior performance on math (GSM8K, $58% \to 67%$) and code infilling performance ($52% \to 65%$).

[829] Reinforcement Learning Driven Generalizable Feature Representation for Cross-User Activity Recognition

Xiaozhou Ye, Kevin I-Kai Wang

Main category: cs.LG

TL;DR: TPRL-DG is a reinforcement learning framework for human activity recognition that uses Transformer-based tokenization to capture user-invariant temporal patterns, achieving state-of-the-art cross-user generalization without requiring target user annotations.

DetailsMotivation: Cross-user variability in wearable sensor data (due to different motion patterns, sensor placements, and physiological traits) limits generalization of conventional supervised learning methods in real-world HAR applications. Existing domain generalization approaches often ignore temporal dependencies or require impractical domain labels.

Method: Temporal-Preserving Reinforcement Learning Domain Generalization (TPRL-DG) redefines feature extraction as a sequential decision-making process using RL. It employs a Transformer-based autoregressive generator to produce temporal tokens capturing user-invariant activity dynamics, optimized via a multi-objective reward function balancing class discrimination and cross-user invariance.

Result: Evaluations on DSADS and PAMAP2 datasets show TPRL-DG surpasses state-of-the-art methods in cross-user generalization, achieving superior accuracy without per-user calibration.

Conclusion: TPRL-DG enables scalable HAR systems by learning robust, user-invariant temporal patterns, facilitating advancements in personalized healthcare, adaptive fitness tracking, and context-aware environments.

Abstract: Human Activity Recognition (HAR) using wearable sensors is crucial for healthcare, fitness tracking, and smart environments, yet cross-user variability – stemming from diverse motion patterns, sensor placements, and physiological traits – hampers generalization in real-world settings. Conventional supervised learning methods often overfit to user-specific patterns, leading to poor performance on unseen users. Existing domain generalization approaches, while promising, frequently overlook temporal dependencies or depend on impractical domain-specific labels. We propose Temporal-Preserving Reinforcement Learning Domain Generalization (TPRL-DG), a novel framework that redefines feature extraction as a sequential decision-making process driven by reinforcement learning. TPRL-DG leverages a Transformer-based autoregressive generator to produce temporal tokens that capture user-invariant activity dynamics, optimized via a multi-objective reward function balancing class discrimination and cross-user invariance. Key innovations include: (1) an RL-driven approach for domain generalization, (2) autoregressive tokenization to preserve temporal coherence, and (3) a label-free reward design eliminating the need for target user annotations. Evaluations on the DSADS and PAMAP2 datasets show that TPRL-DG surpasses state-of-the-art methods in cross-user generalization, achieving superior accuracy without per-user calibration. By learning robust, user-invariant temporal patterns, TPRL-DG enables scalable HAR systems, facilitating advancements in personalized healthcare, adaptive fitness tracking, and context-aware environments.

[830] MatPROV: A Provenance Graph Dataset of Material Synthesis Extracted from Scientific Literature

Hirofumi Tsuruta, Masaya Kumagai

Main category: cs.LG

TL;DR: MatPROV introduces a graph-based approach using PROV-DM standard to extract and represent materials synthesis procedures from scientific literature, capturing complex structural relationships through machine-interpretable directed graphs.

DetailsMotivation: Existing methods for extracting synthesis procedures rely on rigid schemas or assume linear sequences, limiting their ability to capture the structural complexity of real-world procedures in materials research.

Method: Adopts PROV-DM international standard for provenance information to create flexible, graph-based modeling of procedures. Uses large language models to extract PROV-DM-compliant synthesis procedures from scientific literature, representing them as directed graphs showing causal relationships.

Result: Developed MatPROV dataset that captures structural complexities and causal relationships among materials, operations, and conditions through visually intuitive directed graphs.

Conclusion: The graph-based representation enables machine-interpretable synthesis knowledge, opening opportunities for future research such as automated synthesis planning and optimization in materials discovery.

Abstract: Synthesis procedures play a critical role in materials research, as they directly affect material properties. With data-driven approaches increasingly accelerating materials discovery, there is growing interest in extracting synthesis procedures from scientific literature as structured data. However, existing studies often rely on rigid, domain-specific schemas with predefined fields for structuring synthesis procedures or assume that synthesis procedures are linear sequences of operations, which limits their ability to capture the structural complexity of real-world procedures. To address these limitations, we adopt PROV-DM, an international standard for provenance information, which supports flexible, graph-based modeling of procedures. We present MatPROV, a dataset of PROV-DM-compliant synthesis procedures extracted from scientific literature using large language models. MatPROV captures structural complexities and causal relationships among materials, operations, and conditions through visually intuitive directed graphs. This representation enables machine-interpretable synthesis knowledge, opening opportunities for future research such as automated synthesis planning and optimization.

[831] IMU-Enhanced EEG Motion Artifact Removal with Fine-Tuned Large Brain Models

Yuhong Zhang, Xusheng Zhu, Yuchen Xu, ChiaEn Lu, Hsinyu Shih, Gert Cauwenberghs, Tzyy-Ping Jung

Main category: cs.LG

TL;DR: A novel EEG motion artifact removal method using fine-tuned large brain model with IMU correlation attention mapping, outperforming traditional single-modality approaches.

DetailsMotivation: EEG signals suffer from low signal-to-noise ratios due to motion artifacts, hindering real-world BCI deployment. Most existing methods use single-modality approaches without leveraging motion data from IMUs.

Method: Proposes a fine-tuned large brain model (LaBraM)-based correlation attention mapping method that uses spatial channel relationships in IMU data to identify motion artifacts in EEG signals. The model has 9.2M parameters and was trained on 5.9 hours of EEG-IMU data.

Result: The method significantly improves robustness under diverse motion scenarios compared to the ASR-ICA benchmark, showing better performance across varying time scales and motion activities.

Conclusion: Incorporating IMU reference signals with correlation attention mapping effectively enhances EEG motion artifact removal, demonstrating the value of multi-modal approaches for real-world BCI applications.

Abstract: Electroencephalography (EEG) is a non-invasive method for measuring brain activity with high temporal resolution; however, EEG signals often exhibit low signal-to-noise ratios because of contamination from physiological and environmental artifacts. One of the major challenges hindering the real-world deployment of brain-computer interfaces (BCIs) involves the frequent occurrence of motion-related EEG artifacts. Most prior studies on EEG motion artifact removal rely on single-modality approaches, such as Artifact Subspace Reconstruction (ASR) and Independent Component Analysis (ICA), without incorporating simultaneously recorded modalities like inertial measurement units (IMUs), which directly capture the extent and dynamics of motion. This work proposes a fine-tuned large brain model (LaBraM)-based correlation attention mapping method that leverages spatial channel relationships in IMU data to identify motion-related artifacts in EEG signals. The fine-tuned model contains approximately 9.2 million parameters and uses 5.9 hours of EEG and IMU recordings for training, just 0.2346% of the 2500 hours used to train the base model. We compare our results against the established ASR-ICA benchmark across varying time scales and motion activities, showing that incorporating IMU reference signals significantly improves robustness under diverse motion scenarios.

[832] REFINESTAT: Efficient Exploration for Probabilistic Program Synthesis

Madhav Kanda, Shubham Ugare, Sasa Misailovic

Main category: cs.LG

TL;DR: RefineStat is a framework that improves small language models’ ability to generate statistically reliable probabilistic programs by enforcing semantic constraints and applying diagnostic-aware refinement.

DetailsMotivation: Small language models often produce probabilistic programs with syntactic and semantic errors when generating statistical models, despite probabilistic programmers' domain expertise and debugging strategies.

Method: Introduces RefineStat, a language model-driven framework that enforces semantic constraints to ensure valid distributions and well-formed parameters, and applies diagnostic-aware refinement by resampling prior or likelihood components when reliability checks fail.

Result: RefineStat produces programs that are both syntactically sound and statistically reliable, often matching or surpassing the performance of closed-source large language models like OpenAI o3 on probabilistic-programming code-generation tasks.

Conclusion: The framework successfully addresses the challenges of statistical model discovery in probabilistic programming by leveraging domain expertise and systematic refinement techniques to improve program generation quality.

Abstract: Probabilistic programming offers a powerful framework for modeling uncertainty, yet statistical model discovery in this domain entails navigating an immense search space under strict domain-specific constraints. When small language models are tasked with generating probabilistic programs, they frequently produce outputs that suffer from both syntactic and semantic errors, such as flawed inference constructs. Motivated by probabilistic programmers’ domain expertise and debugging strategies, we introduce RefineStat, a language model–driven framework that enforces semantic constraints ensuring synthesized programs contain valid distributions and well-formed parameters, and then applies diagnostic-aware refinement by resampling prior or likelihood components whenever reliability checks fail. We evaluate RefineStat on multiple probabilistic-programming code-generation tasks using smaller language models (SLMs) and find that it produces programs that are both syntactically sound and statistically reliable, often matching or surpassing those from closed-source large language models (e.g., OpenAI o3).

[833] A Class of Random-Kernel Network Models

James Tian

Main category: cs.LG

TL;DR: Random-kernel networks use deterministic kernel composition with randomness only in the outermost layer, showing deeper networks can approximate functions with fewer samples than shallow ones.

DetailsMotivation: To extend random feature models to multilayer architectures while maintaining theoretical guarantees and demonstrating the advantage of depth in reducing sample complexity.

Method: Develop random-kernel networks where depth is created through deterministic kernel composition, with randomness introduced only in the final layer. Prove theoretical bounds on sample complexity.

Result: Established a depth separation theorem showing that deeper random-kernel networks can approximate certain functions with significantly fewer Monte Carlo samples compared to any shallow counterpart.

Conclusion: Depth in random-kernel networks provides concrete advantages in sample efficiency, demonstrating that properly constructed deep architectures can outperform shallow models in terms of sample complexity for function approximation.

Abstract: We introduce random-kernel networks, a multilayer extension of random feature models where depth is created by deterministic kernel composition and randomness enters only in the outermost layer. We prove that deeper constructions can approximate certain functions with fewer Monte Carlo samples than any shallow counterpart, establishing a depth separation theorem in sample complexity.

[834] CCE: Confidence-Consistency Evaluation for Time Series Anomaly Detection

Zhijie Zhong, Zhiwen Yu, Yiu-ming Cheung, Kaixiang Yang

Main category: cs.LG

TL;DR: CCE is a novel time series anomaly detection evaluation metric that measures prediction confidence and uncertainty consistency using Bayesian estimation, offering better discriminative power, robustness, and efficiency than existing metrics.

DetailsMotivation: Existing time series anomaly detection metrics suffer from insufficient discriminative power, strong hyperparameter dependency, sensitivity to perturbations, and high computational overhead, limiting their effectiveness in model evaluation.

Method: The paper introduces Confidence-Consistency Evaluation (CCE) which employs Bayesian estimation to quantify uncertainty of anomaly scores, constructing both global and event-level confidence and consistency scores for model predictions.

Result: CCE demonstrates strict boundedness, Lipschitz robustness against score perturbations, and linear time complexity O(n). The paper also establishes RankEval, the first standardized benchmark for comparing metric ranking capabilities.

Conclusion: CCE provides a superior evaluation metric for time series anomaly detection with better theoretical properties and practical performance, complemented by RankEval benchmark for objective metric comparison. Both implementations are open-source.

Abstract: Time Series Anomaly Detection metrics serve as crucial tools for model evaluation. However, existing metrics suffer from several limitations: insufficient discriminative power, strong hyperparameter dependency, sensitivity to perturbations, and high computational overhead. This paper introduces Confidence-Consistency Evaluation (CCE), a novel evaluation metric that simultaneously measures prediction confidence and uncertainty consistency. By employing Bayesian estimation to quantify the uncertainty of anomaly scores, we construct both global and event-level confidence and consistency scores for model predictions, resulting in a concise CCE metric. Theoretically and experimentally, we demonstrate that CCE possesses strict boundedness, Lipschitz robustness against score perturbations, and linear time complexity $\mathcal{O}(n)$. Furthermore, we establish RankEval, a benchmark for comparing the ranking capabilities of various metrics. RankEval represents the first standardized and reproducible evaluation pipeline that enables objective comparison of evaluation metrics. Both CCE and RankEval implementations are fully open-source.

[835] SC-GIR: Goal-oriented Semantic Communication via Invariant Representation Learning

Senura Hansaja Wanasekara, Van-Dinh Nguyen, Kok-Seng, M. -Duong Nguyen, Symeon Chatzinotas, Octavia A. Dobre

Main category: cs.LG

TL;DR: Proposes SC-GIR framework for goal-oriented semantic communication using self-supervised learning to extract invariant representations for efficient image transmission and downstream task execution.

DetailsMotivation: Current goal-oriented semantic communication approaches face challenges with joint transceiver training, redundant data exchange, and reliance on labeled datasets, limiting task-agnostic utility.

Method: Uses self-supervised learning with covariance-based contrastive learning to extract invariant representations from source data, independent of specific downstream tasks, for efficient compressed transmission.

Result: Outperforms baseline schemes by nearly 10% and achieves over 85% classification accuracy for compressed data under different SNR conditions across multiple datasets.

Conclusion: SC-GIR framework effectively learns compact and informative latent representations for goal-oriented semantic communication, demonstrating superior performance in machine-to-machine image transmission tasks.

Abstract: Goal-oriented semantic communication (SC) aims to revolutionize communication systems by transmitting only task-essential information. However, current approaches face challenges such as joint training at transceivers, leading to redundant data exchange and reliance on labeled datasets, which limits their task-agnostic utility. To address these challenges, we propose a novel framework called Goal-oriented Invariant Representation-based SC (SC-GIR) for image transmission. Our framework leverages self-supervised learning to extract an invariant representation that encapsulates crucial information from the source data, independent of the specific downstream task. This compressed representation facilitates efficient communication while retaining key features for successful downstream task execution. Focusing on machine-to-machine tasks, we utilize covariance-based contrastive learning techniques to obtain a latent representation that is both meaningful and semantically dense. To evaluate the effectiveness of the proposed scheme on downstream tasks, we apply it to various image datasets for lossy compression. The compressed representations are then used in a goal-oriented AI task. Extensive experiments on several datasets demonstrate that SC-GIR outperforms baseline schemes by nearly 10%,, and achieves over 85% classification accuracy for compressed data under different SNR conditions. These results underscore the effectiveness of the proposed framework in learning compact and informative latent representations.

[836] MATL-DC: A Multi-domain Aggregation Transfer Learning Framework for EEG Emotion Recognition with Domain-Class Prototype under Unseen Targets

Guangli Li, Canbiao Wu, Zhehao Zhou, Na Tian, Zhen Liang

Main category: cs.LG

TL;DR: Proposes MATL-DC framework for EEG emotion recognition using multi-domain aggregation transfer learning with domain-class prototypes for unseen target domains.

DetailsMotivation: Current transfer learning models for EEG-based emotion recognition heavily depend on both source and target domain data, limiting practical applications. The need for methods that work with completely unseen target domains during training.

Method: Feature decoupling module separates class-invariant domain features from domain-invariant class features. Multi-domain aggregation creates superdomains from source domains. Extracts class prototypes and uses pairwise learning strategy to transform classification into similarity problems, reducing label noise impact.

Result: Achieved accuracies of 84.70% on SEED, 68.11% on SEED-IV, and 61.08% on SEED-V databases. Outperforms methods that require both source and target domain data during training.

Conclusion: MATL-DC framework effectively handles EEG emotion recognition with completely unseen target domains, demonstrating superior performance and practical applicability for real-world affective BCI systems.

Abstract: Emotion recognition based on electroencephalography (EEG) signals is increasingly becoming a key research hotspot in affective Brain-Computer Interfaces (aBCIs). However, the current transfer learning model greatly depends on the source domain and target domain data, which hinder the practical application of emotion recognition. Therefore, we propose a Multi-domain Aggregation Transfer Learning framework for EEG emotion recognition with Domain-Class prototype under unseen targets (MATL-DC). We design the feature decoupling module to decouple class-invariant domain features from domain-invariant class features from shallow features. In the model training stage, the multi-domain aggregation mechanism aggregates the domain feature space to form a superdomain, which enhances the characteristics of emotional EEG signals. In each superdomain, we further extract the class prototype representation by class features. In addition, we adopt the pairwise learning strategy to transform the sample classification problem into the similarity problem between sample pairs, which effectively alleviates the influence of label noise. It is worth noting that the target domain is completely unseen during the training process. In the inference stage, we use the trained domain-class prototypes for inference, and then realize emotion recognition. We rigorously validate it on the publicly available databases (SEED, SEED-IV and SEED-V). The results show that the accuracy of MATL-DC model is 84.70%, 68.11% and 61.08%, respectively. MATL-DC achieves comparable or even better performance than methods that rely on both source and target domains. The source code is available at https://github.com/WuCB-BCI/MATL-DC.

[837] Nonlinear Performative Prediction

Guangzheng Zhong, Yang Liu, Jiming Liu

Main category: cs.LG

TL;DR: This paper introduces a novel approach to performative prediction that extends it to nonlinear cases using kernel methods and maximum margin formulation, while providing theoretical guarantees for performative stability.

DetailsMotivation: Current performative prediction methods rely on unrealistic linear assumptions and uncontrollable conditions like bounded gradients, which don't hold in real-world applications with complex nonlinear data distributions.

Method: The authors formulate performative prediction loss using maximum margin approach and extend to nonlinear spaces via kernel methods. They quantify distribution shift using prediction error discrepancy and derive conditions for performative stability.

Result: The method achieves performative stability for both linear and nonlinear cases, with superior performance demonstrated on synthetic and real-world datasets compared to state-of-the-art baselines.

Conclusion: The proposed framework successfully generalizes performative prediction to nonlinear scenarios while preserving essential theoretical properties, making it more applicable to real-world problems.

Abstract: Performative prediction is an emerging paradigm in machine learning that addresses scenarios where the model’s prediction may induce a shift in the distribution of the data it aims to predict. Current works in this field often rely on uncontrollable assumptions, such as bounded gradients of performative loss, and primarily focus on linear cases in their examples and evaluations to maintain consistency between theoretical guarantees and empirical validations. However, such linearity rarely holds in real-world applications, where the data usually exhibit complex nonlinear characteristics. In this paper, we relax these out-of-control assumptions and present a novel design that generalizes performative prediction to nonlinear cases while preserving essential theoretical properties. Specifically, we formulate the loss function of performative prediction using a maximum margin approach and extend it to nonlinear spaces through kernel methods. To quantify the data distribution shift, we employ the discrepancy between prediction errors on these two distributions as an indicator, which characterizes the impact of the performative effect on specific learning tasks. By doing so, we can derive, for both linear and nonlinear cases, the conditions for performative stability, a critical and desirable property in performative contexts. Building on these theoretical insights, we develop an algorithm that guarantees the performative stability of the predictive model. We validate the effectiveness of our method through experiments on synthetic and real-world datasets with both linear and nonlinear data distributions, demonstrating superior performance compared to state-of-the-art baselines.

[838] Multi-Modal Machine Learning Framework for Predicting Early Recurrence of Brain Tumors Using MRI and Clinical Biomarkers

Cheng Cheng, Zeping Chen, Rui Xie, Peiyao Zheng, Xavier Wang

Main category: cs.LG

TL;DR: Multi-modal ML framework combining MRI features and clinical biomarkers improves brain tumor recurrence prediction after surgery using four algorithms with strong validation metrics.

DetailsMotivation: Accurate prediction of early recurrence in brain tumor patients after surgical resection is a clinical challenge that needs better tools for risk stratification and personalized care.

Method: Integrated structural MRI features with clinical biomarkers using four machine learning algorithms (GBM, RSF, CoxBoost, XGBoost) and validated with C-index, time-dependent AUC, calibration curves, and decision curve analysis.

Result: The model demonstrates promising performance for postoperative recurrence prediction.

Conclusion: The framework offers a potential tool for improved risk stratification and personalized follow-up planning in brain tumor patients.

Abstract: Accurately predicting early recurrence in brain tumor patients following surgical resection remains a clinical challenge. This study proposes a multi-modal machine learning framework that integrates structural MRI features with clinical biomarkers to improve postoperative recurrence prediction. We employ four machine learning algorithms – Gradient Boosting Machine (GBM), Random Survival Forest (RSF), CoxBoost, and XGBoost – and validate model performance using concordance index (C-index), time-dependent AUC, calibration curves, and decision curve analysis. Our model demonstrates promising performance, offering a potential tool for risk stratification and personalized follow-up planning.

[839] ADMP-GNN: Adaptive Depth Message Passing GNN

Yassine Abbahaddou, Fragkiskos D. Malliaros, Johannes F. Lutzeyer, Michalis Vazirgiannis

Main category: cs.LG

TL;DR: ADMP-GNN is a novel GNN framework that dynamically adjusts message-passing layers per node instead of using fixed layers, improving performance on node classification tasks.

DetailsMotivation: Standard GNNs use fixed message-passing steps for all nodes, but empirical analysis shows optimal layer count varies by node characteristics, indicating a need for adaptive depth.

Method: Proposes Adaptive Depth Message Passing GNN (ADMP-GNN) that dynamically adjusts the number of message passing layers for each node individually, applicable to any message-passing model.

Result: ADMP-GNN shows performance improvements over baseline GNN models on node classification tasks, with findings supported by both real-world data analysis and synthetic dataset experiments.

Conclusion: Adaptive depth message passing that tailors computational steps to individual node characteristics significantly enhances GNN performance compared to fixed-depth approaches.

Abstract: Graph Neural Networks (GNNs) have proven to be highly effective in various graph learning tasks. A key characteristic of GNNs is their use of a fixed number of message-passing steps for all nodes in the graph, regardless of each node’s diverse computational needs and characteristics. Through empirical real-world data analysis, we demonstrate that the optimal number of message-passing layers varies for nodes with different characteristics. This finding is further supported by experiments conducted on synthetic datasets. To address this, we propose Adaptive Depth Message Passing GNN (ADMP-GNN), a novel framework that dynamically adjusts the number of message passing layers for each node, resulting in improved performance. This approach applies to any model that follows the message passing scheme. We evaluate ADMP-GNN on the node classification task and observe performance improvements over baseline GNN models.

[840] StoxLSTM: A Stochastic Extended Long Short-Term Memory Network for Time Series Forecasting

Zihao Wang, Yunjie Li, Lingmin Zan, Zheng Gong, Mengtao Zhu

Main category: cs.LG

TL;DR: StoxLSTM enhances xLSTM by incorporating stochastic latent variables into a state space modeling framework, improving temporal pattern capture and forecasting performance on complex time series data.

DetailsMotivation: While xLSTM has shown success in modeling temporal dependencies, there's still potential to improve its representational capacity and forecasting performance, especially on real-world datasets with unknown, intricate, and hierarchical dynamics.

Method: Proposes StoxLSTM that incorporates stochastic latent variables within xLSTM architecture, creating a state space modeling framework with specially designed recurrent blocks to model latent dynamic evolution.

Result: Extensive experiments on benchmark datasets show StoxLSTM consistently outperforms state-of-the-art baselines with better robustness and stronger generalization ability.

Conclusion: StoxLSTM successfully improves upon xLSTM by integrating stochastic modeling, demonstrating superior performance in capturing complex temporal patterns and dependencies across diverse time series applications.

Abstract: The Extended Long Short-Term Memory (xLSTM) network has attracted widespread research interest due to its enhanced capability to model complex temporal dependencies in diverse time series applications. Despite its success, there is still potential to further improve its representational capacity and forecasting performance, particularly on challenging real-world datasets with unknown, intricate, and hierarchical dynamics. In this work, we propose a stochastic xLSTM, termed StoxLSTM, that improves the original architecture into a state space modeling framework by incorporating stochastic latent variables within xLSTM. StoxLSTM models the latent dynamic evolution through specially designed recurrent blocks, enabling it to effectively capture the underlying temporal patterns and dependencies. Extensive experiments on publicly available benchmark datasets from multiple research communities demonstrate that StoxLSTM consistently outperforms state-of-the-art baselines with better robustness and stronger generalization ability.

[841] Preserving Vector Space Properties in Dimensionality Reduction: A Relationship Preserving Loss Framework

Eddi Weinwurm, Alexander Kovalenko

Main category: cs.LG

TL;DR: RPL is a loss function that preserves vector space properties like orthogonality and linear independence during dimensionality reduction by minimizing discrepancies between relationship matrices of high-dimensional data and their low-dimensional embeddings.

DetailsMotivation: Dimensionality reduction techniques often distort critical vector space properties (orthogonality, linear independence) that are essential for tasks like cross-modal retrieval, clustering, and classification.

Method: Propose Relationship Preserving Loss (RPL) that minimizes discrepancies between relationship matrices (Gram or cosine) of original high-dimensional data and their low-dimensional embeddings. Uses neural networks for non-linear projections with error bounds from matrix perturbation theory.

Result: Initial experiments show RPL reduces embedding dimensions while largely retaining performance on downstream tasks, likely due to preservation of key vector space properties.

Conclusion: RPL effectively preserves vector space geometry during dimensionality reduction and has broad applicability beyond DR to cross-domain alignment, transfer learning, knowledge distillation, fairness, graph learning, and federated learning where geometric consistency is crucial.

Abstract: Dimensionality reduction can distort vector space properties such as orthogonality and linear independence, which are critical for tasks including cross-modal retrieval, clustering, and classification. We propose a Relationship Preserving Loss (RPL), a loss function that preserves these properties by minimizing discrepancies between relationship matrices (e.g., Gram or cosine) of high-dimensional data and their low-dimensional embeddings. RPL trains neural networks for non-linear projections and is supported by error bounds derived from matrix perturbation theory. Initial experiments suggest that RPL reduces embedding dimensions while largely retaining performance on downstream tasks, likely due to its preservation of key vector space properties. While we describe here the use of RPL in dimensionality reduction, this loss can also be applied more broadly, for example to cross-domain alignment and transfer learning, knowledge distillation, fairness and invariance, dehubbing, graph and manifold learning, and federated learning, where distributed embeddings must remain geometrically consistent.

[842] Geometric origin of adversarial vulnerability in deep learning

Yixiong Ren, Wenkang Du, Jianhui Zhou, Haiping Huang

Main category: cs.LG

TL;DR: A geometry-aware deep learning framework using layer-wise local training to improve adversarial robustness while maintaining training accuracy, explained through an energy model with Hebbian coupling.

DetailsMotivation: To address the challenge of balancing training accuracy and adversarial robustness in deep learning by developing a framework that promotes better feature space organization.

Method: Layer-wise local training framework that sculpts internal representations to achieve intra-class compactness and inter-class separation, using an energy model with Hebbian coupling.

Result: Achieves manifold smoothness and adversarial robustness against both white and black box attacks, while enabling assimilation of new information with reduced representation interference.

Conclusion: The framework provides insights into the physics of learning and advances alignment between biological and artificial intelligence systems through improved feature space organization.

Abstract: How to balance training accuracy and adversarial robustness has become a challenge since the birth of deep learning. Here, we introduce a geometry-aware deep learning framework that leverages layer-wise local training to sculpt the internal representations of deep neural networks. This framework promotes intra-class compactness and inter-class separation in feature space, leading to manifold smoothness and adversarial robustness against white or black box attacks. The performance can be explained by an energy model with Hebbian coupling between elements of the hidden representation. Our results thus shed light on the physics of learning in the direction of alignment between biological and artificial intelligence systems. Using the current framework, the deep network can assimilate new information into existing knowledge structures while reducing representation interference.

[843] What Expressivity Theory Misses: Message Passing Complexity for GNNs

Niklas Kemper, Tom Wollschläger, Stephan Günnemann

Main category: cs.LG

TL;DR: The paper argues that current focus on GNN expressivity theory is misguided and proposes Message Passing Complexity (MPC) as a better continuous measure that captures practical limitations and better correlates with empirical performance.

DetailsMotivation: Current expressivity theory for GNNs is too binary and idealized, failing to reflect practical capabilities. Higher expressivity is often unnecessary for real-world tasks, and the theory doesn't account for practical limitations like over-squashing.

Method: Proposes Message Passing Complexity (MPC) - a continuous measure that quantifies the difficulty for a GNN architecture to solve tasks through message passing, capturing practical limitations while preserving theoretical impossibility results.

Result: MPC’s theoretical predictions correlate well with empirical performance on fundamental GNN tasks, successfully explaining architectural successes and failures.

Conclusion: MPC provides a more powerful and nuanced framework than expressivity theory for understanding and improving GNN architectures, effectively narrowing the gap between theory and practice.

Abstract: Expressivity theory, characterizing which graphs a GNN can distinguish, has become the predominant framework for analyzing GNNs, with new models striving for higher expressivity. However, we argue that this focus is misguided: First, higher expressivity is not necessary for most real-world tasks as these tasks rarely require expressivity beyond the basic WL test. Second, expressivity theory’s binary characterization and idealized assumptions fail to reflect GNNs’ practical capabilities. To overcome these limitations, we propose Message Passing Complexity (MPC): a continuous measure that quantifies the difficulty for a GNN architecture to solve a given task through message passing. MPC captures practical limitations like over-squashing while preserving the theoretical impossibility results from expressivity theory, effectively narrowing the gap between theory and practice. Through extensive validation on fundamental GNN tasks, we show that MPC’s theoretical predictions correlate with empirical performance, successfully explaining architectural successes and failures. Thereby, MPC advances beyond expressivity theory to provide a more powerful and nuanced framework for understanding and improving GNN architectures.

[844] Multi-Agent Reinforcement Learning for Task Offloading in Wireless Edge Networks

Andrea Fox, Francesco De Pellegrini, Eitan Altman

Main category: cs.LG

TL;DR: Decentralized MARL framework using constrained MDPs with shared constraint vectors for edge computing resource coordination, enabling local decision-making with minimal communication.

DetailsMotivation: Existing multi-agent reinforcement learning methods rely on centralized critics or frequent communication, which fail under limited observability and communication constraints in edge computing systems where agents compete for shared resources.

Method: Proposes a decentralized framework where each agent solves a constrained Markov decision process (CMDP) with coordination through infrequently updated shared constraint vectors that prevent overloading shared server resources, using safe reinforcement learning.

Result: Theoretical guarantees established under mild assumptions, with experimental validation showing improved performance over centralized and independent baselines, particularly in large-scale settings.

Conclusion: The framework enables effective coordination with minimal communication, allowing agents to align with global resource usage objectives while making fast local decisions in edge computing environments.

Abstract: In edge computing systems, autonomous agents must make fast local decisions while competing for shared resources. Existing MARL methods often resume to centralized critics or frequent communication, which fail under limited observability and communication constraints. We propose a decentralized framework in which each agent solves a constrained Markov decision process (CMDP), coordinating implicitly through a shared constraint vector. For the specific case of offloading, e.g., constraints prevent overloading shared server resources. Coordination constraints are updated infrequently and act as a lightweight coordination mechanism. They enable agents to align with global resource usage objectives but require little direct communication. Using safe reinforcement learning, agents learn policies that meet both local and global goals. We establish theoretical guarantees under mild assumptions and validate our approach experimentally, showing improved performance over centralized and independent baselines, especially in large-scale settings.

[845] Iterative In-Context Learning to Enhance LLMs Abstract Reasoning: The Case-Study of Algebraic Tasks

Stefano Fioravanti, Matteo Zavatteri, Roberto Confalonieri, Kamyar Zeinalipour, Paolo Frazzetto, Alessandro Sperduti, Nicolò Navarin

Main category: cs.LG

TL;DR: Iterative example selection strategy improves LLM generalization for compositional reasoning tasks like algebraic expressions with non-standard rules.

DetailsMotivation: LLMs struggle with systematic generalization, especially for compositional reasoning tasks and out-of-distribution examples that require handling non-standard rules.

Method: In-context learning methodology with iterative example selection strategy that incrementally constructs tailored few-shot examples optimized for specific tasks, applied to algebraic expressions with changed operation priorities.

Result: LLMs show limited proficiency in mathematical tasks with non-standard rules, but performance improves with iterative shot selection and explicit reasoning instructions. Simpler few-shot examples work better than complex ones matching test distribution.

Conclusion: The iterative example selection approach enhances LLM generalization capabilities for compositional reasoning tasks, demonstrating that simpler examples can be more effective than complex ones for systematic generalization.

Abstract: LLMs face significant challenges in systematic generalization, particularly when dealing with reasoning tasks requiring compositional rules and handling out-of-distribution examples. To address these challenges, we introduce an in-context learning methodology that improves the generalization capabilities of general purpose LLMs. Our approach employs an iterative example selection strategy, which incrementally constructs a tailored set of few-shot examples optimized to enhance model’s performance on a given task. As a proof of concept, we apply this methodology to the resolution of algebraic expressions involving non-standard simplification rules, according to which the priority of addition and multiplication is changed. Our findings indicate that LLMs exhibit limited proficiency in these mathematical tasks. We further demonstrate that LLMs reasoning benefits from our iterative shot selection prompting strategy integrated with explicit reasoning instructions. Crucially, our experiments reveal that some LLMs achieve better generalization performances when prompted with simpler few-shot examples rather than complex ones following the test data distribution.

[846] Building surrogate models using trajectories of agents trained by Reinforcement Learning

Julen Cestero, Marco Quartulli, Marcello Restelli

Main category: cs.LG

TL;DR: Proposes using RL-trained policies to efficiently sample deterministic simulation environments for surrogate modeling, outperforming traditional sampling methods like Latin-Hypercube and Active Learning.

DetailsMotivation: Addresses sample efficiency challenges in surrogate modeling for computationally expensive simulations with wide state spaces, where current sampling strategies are ineffective.

Method: Uses policies trained by Reinforcement Learning to sample simulated deterministic environments, including random agents, expert agents, and agents trained to explore maximum entropy regions of state transition distribution.

Result: Mixed dataset combining samples from random agents, expert agents, and maximum entropy exploration agents provides best performance across all datasets, enabling meaningful state space representation.

Conclusion: The proposed method improves state-of-the-art sampling efficiency and enables surrogate-aided Reinforcement Learning policy optimization strategies on complex simulators.

Abstract: Sample efficiency in the face of computationally expensive simulations is a common concern in surrogate modeling. Current strategies to minimize the number of samples needed are not as effective in simulated environments with wide state spaces. As a response to this challenge, we propose a novel method to efficiently sample simulated deterministic environments by using policies trained by Reinforcement Learning. We provide an extensive analysis of these surrogate-building strategies with respect to Latin-Hypercube sampling or Active Learning and Kriging, cross-validating performances with all sampled datasets. The analysis shows that a mixed dataset that includes samples acquired by random agents, expert agents, and agents trained to explore the regions of maximum entropy of the state transition distribution provides the best scores through all datasets, which is crucial for a meaningful state space representation. We conclude that the proposed method improves the state-of-the-art and clears the path to enable the application of surrogate-aided Reinforcement Learning policy optimization strategies on complex simulators.

[847] Equivariant U-Shaped Neural Operators for the Cahn-Hilliard Phase-Field Model

Xiao Xue, M. F. P. ten Eikelder, Tianyue Yang, Yiqing Li, Kan He, Shuo Wang, Peter V. Coveney

Main category: cs.LG

TL;DR: E-UNO: an equivariant U-shaped neural operator that accurately predicts phase separation dynamics in binary mixtures by encoding physical symmetries and multiscale behavior, outperforming existing neural operators.

DetailsMotivation: Traditional numerical solvers for Cahn-Hilliard equation are computationally expensive and lack flexibility, while current neural operators fail to capture multiscale behavior and physical symmetries in phase separation problems.

Method: Developed an equivariant U-shaped neural operator (E-UNO) that combines global spectral convolution with multi-resolution architecture, regulates translation equivariance, and learns from short histories of past dynamics.

Result: E-UNO outperforms standard Fourier neural operator and U-shaped neural operator baselines, particularly on fine-scale and high-frequency structures, with better generalization and less training data required.

Conclusion: E-UNO establishes an efficient surrogate model for complex phase-field systems by encoding symmetry and scale hierarchy, yielding physically consistent dynamics across varying conditions.

Abstract: Phase separation in binary mixtures, governed by the Cahn-Hilliard equation, plays a central role in interfacial dynamics across materials science and soft matter. While numerical solvers are accurate, they are often computationally expensive and lack flexibility across varying initial conditions and geometries. Neural operators provide a data-driven alternative by learning solution operators between function spaces, but current architectures often fail to capture multiscale behavior and neglect underlying physical symmetries. Here we show that an equivariant U-shaped neural operator (E-UNO) can learn the evolution of the phase-field variable from short histories of past dynamics, achieving accurate predictions across space and time. The model combines global spectral convolution with a multi-resolution U-shaped architecture and regulates translation equivariance to align with the underlying physics. E-UNO outperforms standard Fourier neural operator and U-shaped neural operator baselines, particularly on fine-scale and high-frequency structures. By encoding symmetry and scale hierarchy, the model generalizes better, requires less training data, and yields physically consistent dynamics. This establishes E-UNO as an efficient surrogate for complex phase-field systems.

[848] Towards Trustworthy Vital Sign Forecasting: Leveraging Uncertainty for Prediction Intervals

Li Rong Wang, Thomas C. Henderson, Yew Soon Ong, Yih Yng Ng, Xiuyi Fan

Main category: cs.LG

TL;DR: Two methods for deriving prediction intervals from Reconstruction Uncertainty Estimate (RUE) to improve uncertainty quantification in vital sign forecasting, with Gaussian copula method working best for low-frequency data and KNN approach for high-frequency data.

DetailsMotivation: Deep learning models for vital sign forecasting lack reliable uncertainty quantification, making it difficult for clinicians to trust model outputs and distinguish meaningful warnings from model noise, which hinders clinical decision-making.

Method: Two approaches: 1) Parametric Gaussian copula method that assumes prediction errors and uncertainty estimates follow Gaussian copula distribution for closed-form PI computation, and 2) Non-parametric KNN approach that empirically estimates conditional error distribution using similar validation instances.

Result: Gaussian copula method consistently outperforms conformal prediction baselines on low-frequency data, while KNN approach performs best on high-frequency data across two large public datasets with minute- and hour-level sampling.

Conclusion: RUE-derived prediction intervals show clinical promise for delivering interpretable, uncertainty-aware vital sign forecasts that can support better clinical decision-making.

Abstract: Vital signs, such as heart rate and blood pressure, are critical indicators of patient health and are widely used in clinical monitoring and decision-making. While deep learning models have shown promise in forecasting these signals, their deployment in healthcare remains limited in part because clinicians must be able to trust and interpret model outputs. Without reliable uncertainty quantification – particularly calibrated prediction intervals (PIs) – it is unclear whether a forecasted abnormality constitutes a meaningful warning or merely reflects model noise, hindering clinical decision-making. To address this, we present two methods for deriving PIs from the Reconstruction Uncertainty Estimate (RUE), an uncertainty measure well-suited to vital-sign forecasting due to its sensitivity to data shifts and support for label-free calibration. Our parametric approach assumes that prediction errors and uncertainty estimates follow a Gaussian copula distribution, enabling closed-form PI computation. Our non-parametric approach, based on k-nearest neighbours (KNN), empirically estimates the conditional error distribution using similar validation instances. We evaluate these methods on two large public datasets with minute- and hour-level sampling, representing high- and low-frequency health signals. Experiments demonstrate that the Gaussian copula method consistently outperforms conformal prediction baselines on low-frequency data, while the KNN approach performs best on high-frequency data. These results underscore the clinical promise of RUE-derived PIs for delivering interpretable, uncertainty-aware vital sign forecasts.

[849] Towards High Data Efficiency in Reinforcement Learning with Verifiable Reward

Xinyu Tang, Zhenduo Zhang, Yurou Liu, Wayne Xin Zhao, Zujie Wen, Zhiqiang Zhang, Jun Zhou

Main category: cs.LG

TL;DR: DEPO is a data-efficient policy optimization pipeline that combines optimized offline and online data selection strategies to reduce training costs while improving reasoning model performance.

DetailsMotivation: Current RLVR methods for large reasoning models require extensive rollout computation and large datasets, leading to high training costs and low data efficiency.

Method: Combines offline curation of high-quality training samples based on diversity, influence, and difficulty, with online sample-level explorability filtering and replay mechanism for under-explored samples.

Result: Outperforms existing methods across five reasoning benchmarks, achieving 1.85x speed-up on AIME24 and 1.66x speed-up on AIME25 using only 20% of training data compared to full-dataset GRPO.

Conclusion: DEPO provides an effective solution for data-efficient reinforcement learning with verifiable rewards, significantly reducing computational costs while maintaining or improving model performance.

Abstract: Recent advances in large reasoning models have leveraged reinforcement learning with verifiable rewards (RLVR) to improve reasoning capabilities. However, scaling these methods typically requires extensive rollout computation and large datasets, leading to high training costs and low data efficiency. To mitigate this issue, we propose DEPO, a Data-Efficient Policy Optimization pipeline that combines optimized strategies for both offline and online data selection. In the offline phase, we curate a high-quality subset of training samples based on diversity, influence, and appropriate difficulty. During online RLVR training, we introduce a sample-level explorability metric to dynamically filter samples with low exploration potential, thereby reducing substantial rollout computational costs. Furthermore, we incorporate a replay mechanism for under-explored samples to ensure adequate training, which enhances the model’s final convergence performance. Experiments across five reasoning benchmarks show that DEPO consistently outperforms existing methods in both offline and online data selection scenarios. Notably, using only 20% of the training data, our approach achieves a 1.85 times speed-up on AIME24 and a 1.66 times speed-up on AIME25 compared to GRPO trained on the full dataset.

[850] Multitask Battery Management with Flexible Pretraining

Hong Lu, Jiali Chen, Jingzhao Zhang, Guannan He, Xuebing Han, Minggao Ouyang

Main category: cs.LG

TL;DR: FMAE is a flexible pretraining framework that learns unified battery representations from heterogeneous data, enabling efficient multi-task battery management with minimal data requirements.

DetailsMotivation: Industrial battery management requires task-specific methods that need extensive data and engineering effort, limiting scalability. Different tasks use diverse data across temporal scales, sensor resolutions, and channels.

Method: Flexible Masked Autoencoder (FMAE) framework that learns with missing battery data channels and captures inter-correlations across data snippets through pretraining.

Result: FMAE outperforms all task-specific methods across 5 battery management tasks with 11 datasets. Uses 50x less inference data for life prediction while maintaining state-of-the-art results. Handles missing information like system voltage with minimal performance impact.

Conclusion: FMAE provides a practical, flexible, and data-efficient solution for multi-task management of dynamical systems, simplifying real-world battery management with unified representations.

Abstract: Industrial-scale battery management involves various types of tasks, such as estimation, prediction, and system-level diagnostics. Each task employs distinct data across temporal scales, sensor resolutions, and data channels. Building task-specific methods requires a great deal of data and engineering effort, which limits the scalability of intelligent battery management. Here we present the Flexible Masked Autoencoder (FMAE), a flexible pretraining framework that can learn with missing battery data channels and capture inter-correlations across data snippets. FMAE learns unified battery representations from heterogeneous data and can be adopted by different tasks with minimal data and engineering efforts. Experimentally, FMAE consistently outperforms all task-specific methods across five battery management tasks with eleven battery datasets. On remaining life prediction tasks, FMAE uses 50 times less inference data while maintaining state-of-the-art results. Moreover, when real-world data lack certain information, such as system voltage, FMAE can still be applied with marginal performance impact, achieving comparable results with the best hand-crafted features. FMAE demonstrates a practical route to a flexible, data-efficient model that simplifies real-world multi-task management of dynamical systems.

[851] Globally aware optimization with resurgence

Wei Bu

Main category: cs.LG

TL;DR: Novel optimization framework using resurgence theory to extract global landscape information from divergent asymptotic series, enabling principled guidance for local optimizers.

DetailsMotivation: Local gradient-based methods lack global information about objective function landscapes, leading to suboptimal convergence and sensitivity to initialization.

Method: Compute statistical mechanical partition function Z(g) for small coupling g, extract asymptotic series coefficients, identify Borel plane singularities that correspond to critical objective function values.

Result: Borel transform singularities provide one-to-one mapping to critical objective function values, offering global guidance for local optimizers with principled learning rate adaptation.

Conclusion: The framework provides theoretically grounded global optimization guidance based on landscape geometry, unlike heuristic adaptive methods.

Abstract: Modern optimization faces a fundamental challenge: local gradient-based methods provide no global information about the objective function $L$ landscape, often leading to suboptimal convergence and sensitivity to initialization. We introduce a novel optimization framework that leverages resurgence theory from complex analysis to extract global structural information from divergent asymptotic series. Our key insight is that the factorially divergent perturbative expansions of parameter space partition functions encode precise information about all critical objective function value in the landscape through their Borel transform singularities. The algorithm works by computing the statistical mechanical partition function $Z(g) = \int e^{-L(\theta)/g} d\theta$ for small coupling $g\ll 1$, extracting its asymptotic series coefficients, and identifying Borel plane singularities that correspond one-to-one with critical objective function values. These target values provide global guidance to local optimizers, enabling principled learning rate adaptation and escape from suboptimal regions. Unlike heuristic adaptive methods, targets are theoretically grounded in the geometry of the optimization landscape.

[852] AT Loss: Advanced Torrential Loss Function for Precipitation Forecasting

Jaeho Choi, Hyeri Kim, Kwang-Ho Kim, Jaesung Lee

Main category: cs.LG

TL;DR: Proposes a new differentiable loss function called AT loss to address limitations of CSI in precipitation forecasting during dry periods, using QUBO formulation and approximation.

DetailsMotivation: Traditional CSI-based optimization becomes ineffective during extended dry periods when precipitation remains below threshold, limiting forecast accuracy.

Method: Introduces a penalty expression reinterpreted as QUBO formulation, then relaxes it into differentiable AT loss function through approximation process.

Result: Superior performance demonstrated through Lipschitz constant analysis, forecast evaluations, consistency experiments, and ablation studies with operational models.

Conclusion: The proposed AT loss function effectively addresses CSI limitations and improves precipitation forecasting performance, especially during dry periods.

Abstract: Accurate precipitation forecasting is becoming increasingly important in the context of climate change. In response, machine learning-based approaches have recently gained attention as an emerging alternative to traditional methods such as numerical weather prediction and climate models. Nonetheless, many recent approaches still rely on off-the-shelf loss functions, and even the more advanced ones merely involve optimization processes based on the critical success index (CSI). The problem, however, is that CSI may become ineffective during extended dry periods when precipitation remains below the threshold, rendering it less than ideal as a criterion for optimization. To address this limitation, we introduce a simple penalty expression and reinterpret it as a quadratic unconstrained binary optimization (QUBO) formulation. Ultimately, the resulting QUBO formulation is relaxed into a differentiable advanced torrential (AT) loss function through an approximation process. The proposed AT loss demonstrates its superiority through the Lipschitz constant, forecast performance evaluations, consistency experiments, and ablation studies with the operational model.

[853] Causal Sensitivity Identification using Generative Learning

Soma Bandyopadhyay, Sudeshna Sarkar

Main category: cs.LG

TL;DR: A novel generative method using CVAE to identify causal impact through interventions and counterfactuals, applied to spatiotemporal trajectory prediction with reduced confounding bias.

DetailsMotivation: To develop a method that can identify causal relationships between features and outcomes in prediction tasks, particularly for spatiotemporal trajectory analysis where understanding causal influences is crucial for accurate recommendations.

Method: Uses Conditional Variational Autoencoder (CVAE) to perform causal impact analysis through interventional and counterfactual perspectives. Identifies causally sensitive features via interventions and evaluates causal effects through counterfactual reasoning.

Result: The method effectively reduces confounding bias and improves predictive performance. Validated on large-scale GeoLife dataset and Asia Bayesian network benchmark, demonstrating ability to identify causal impact and enhance location recommendation accuracy.

Conclusion: The proposed generative causal impact identification method successfully combines intervention and counterfactual analysis using CVAE, providing a robust framework for causal inference in prediction tasks with practical applications in spatiotemporal trajectory analysis.

Abstract: In this work, we propose a novel generative method to identify the causal impact and apply it to prediction tasks. We conduct causal impact analysis using interventional and counterfactual perspectives. First, applying interventions, we identify features that have a causal influence on the predicted outcome, which we refer to as causally sensitive features, and second, applying counterfactuals, we evaluate how changes in the cause affect the effect. Our method exploits the Conditional Variational Autoencoder (CVAE) to identify the causal impact and serve as a generative predictor. We are able to reduce confounding bias by identifying causally sensitive features. We demonstrate the effectiveness of our method by recommending the most likely locations a user will visit next in their spatiotemporal trajectory influenced by the causal relationships among various features. Experiments on the large-scale GeoLife [Zheng et al., 2010] dataset and the benchmark Asia Bayesian network validate the ability of our method to identify causal impact and improve predictive performance.

[854] DPF-CM: A Data Processing Framework with Privacy-Preserving Vector Databases for Chinese Medical LLMs Training and Deployment

Wei Huang, Anda Cheng, Zhao Zhang, Yinggui Wang

Main category: cs.LG

TL;DR: DPF-CM is a comprehensive data processing framework for Chinese medical LLMs that addresses training data optimization and privacy preservation, achieving SOTA performance and 27% privacy leakage reduction.

DetailsMotivation: Current Chinese medical LLM training pipelines focus on methodology optimization but neglect comprehensive data processing, creating gaps in instruction content and privacy protection.

Method: Two core modules: (1) Training data pipeline with chained examples context-learning for instruction generation and ensemble-based filtering for preference data curation; (2) Privacy preservation using PPVD approach with model memory search, high-risk database construction, secure database, and match-and-replace.

Result: Significantly improves model accuracy, achieving state-of-the-art performance among open-source Chinese medical LLMs, and reduces training data privacy leakage by 27%.

Conclusion: DPF-CM provides a holistic framework that effectively addresses both training data optimization and privacy preservation challenges in Chinese medical LLM development.

Abstract: Current open-source training pipelines for Chinese medical language models predominantly emphasize optimizing training methodologies to enhance the performance of large language models (LLMs), yet lack comprehensive exploration into training data processing. To address this gap, we propose DPF-CM, a holistic Data Processing Framework for Chinese Medical LLMs training and deployment. DPF-CM comprises two core modules. The first module is a data processing pipeline tailored for model training. Beyond standard data processing operations, we (1) introduce a chained examples context-learning strategy to generate question-oriented instructions to mitigate the lack of instruction content, and (2) implement an ensemble-based filtering mechanism for preference data curation that averages multiple reward models to suppress noisy samples. The second module focuses on privacy preservation during model deployment. To prevent privacy risks from the inadvertent exposure of training data, we propose a Privacy Preserving Vector Database (PPVD) approach, which involves model memory search, high-risk database construction, secure database construction, and match-and-replace, four key stages to minimize privacy leakage during inference collectively. Experimental results show that DPF-CM significantly improves model accuracy, enabling our trained Chinese medical LLM to achieve state-of-the-art performance among open-source counterparts. Moreover, the framework reduces training data privacy leakage by 27%.

[855] CbLDM: A Diffusion Model for recovering nanostructure from pair distribution function

Jiarui Cao, Zhiyang Zhang, Heming Wang, Jun Xu, Ling Lan, Ran Gu

Main category: cs.LG

TL;DR: CbLDM - a conditional latent diffusion model for nanostructure recovery from PDF data, using Laplacian matrix and improved sampling efficiency

DetailsMotivation: Solving the nanostructure inverse problem to understand relationships between properties and structures of nanomaterials using PDF data

Method: Condition-based Latent Diffusion Model (CbLDM) that uses conditional prior to estimate posterior distribution, reduces sampling steps, and employs Laplacian matrix instead of distance matrix

Result: CbLDM demonstrates significantly higher prediction accuracy than existing models for nanostructure inverse problem

Conclusion: CbLDM effectively solves nanostructure inverse problem and shows potential for other continuous conditional generation tasks

Abstract: Nowadays, the nanostructure inverse problem is an attractive problem that helps researchers to understand the relationship between the properties and the structure of nanomaterials. This article focuses on the problem of using PDF to recover the nanostructure, which this article views as a conditional generation problem. This article propose a deep learning model CbLDM, Condition-based Latent Diffusion Model. Based on the original latent diffusion model, the sampling steps of the diffusion model are reduced and the sample generation efficiency is improved by using the conditional prior to estimate conditional posterior distribution, which is the approximated distribution of p(z|x). In addition, this article uses the Laplacian matrix instead of the distance matrix to recover the nanostructure, which can reduce the reconstruction error. Finally, this article compares CbLDM with existing models which were used to solve the nanostructure inverse problem, and find that CbLDM demonstrates significantly higher prediction accuracy than these models, which reflects the ability of CbLDM to solve the nanostructure inverse problem and the potential to cope with other continuous conditional generation tasks.

[856] Learn to Jump: Adaptive Random Walks for Long-Range Propagation through Graph Hierarchies

Joël Mathys, Federico Errica

Main category: cs.LG

TL;DR: Novel hierarchical graph approach using adaptive random walks with learnable transition probabilities to overcome long-range dependency limitations in message-passing architectures.

DetailsMotivation: Message-passing architectures struggle to model long-range dependencies in node and graph prediction tasks, creating a need for more effective methods to capture distant relationships in graph data.

Method: Proposes hierarchical graph structures with adaptive random walks featuring learnable transition probabilities that decide whether to traverse the original graph or use hierarchical shortcuts.

Result: On synthetic long-range tasks, the approach exceeds theoretical bounds of traditional methods - walks preferring hierarchy achieve same performance as longer walks on original graph.

Conclusion: The method opens a promising direction for efficiently processing large graphs while effectively capturing long-range dependencies through hierarchical structures and adaptive walks.

Abstract: Message-passing architectures struggle to sufficiently model long-range dependencies in node and graph prediction tasks. We propose a novel approach exploiting hierarchical graph structures and adaptive random walks to address this challenge. Our method introduces learnable transition probabilities that decide whether the walk should prefer the original graph or travel across hierarchical shortcuts. On a synthetic long-range task, we demonstrate that our approach can exceed the theoretical bound that constrains traditional approaches operating solely on the original topology. Specifically, walks that prefer the hierarchy achieve the same performance as longer walks on the original graph. These preliminary findings open a promising direction for efficiently processing large graphs while effectively capturing long-range dependencies.

[857] Distillation of a tractable model from the VQ-VAE

Armin Hadžić, Milan Papez, Tomáš Pevný

Main category: cs.LG

TL;DR: VQ-VAE models can be made tractable by distilling them into probabilistic circuits through selection of high-probability latent variables, enabling efficient probabilistic inference while maintaining expressiveness.

DetailsMotivation: Deep generative models with discrete latent spaces like VQ-VAE have excellent generation capabilities but are considered intractable for probabilistic inference due to their large latent space size.

Method: Distill VQ-VAE into a tractable model by selecting a subset of latent variables with high probabilities, framing the distilled model as a probabilistic circuit.

Result: The distilled model preserves VQ-VAE’s expressiveness while providing tractable probabilistic inference, showing competitive performance in density estimation and conditional generation tasks.

Conclusion: This approach challenges the view of VQ-VAE as inherently intractable and demonstrates that efficient probabilistic inference is achievable through strategic latent variable selection.

Abstract: Deep generative models with discrete latent space, such as the Vector-Quantized Variational Autoencoder (VQ-VAE), offer excellent data generation capabilities, but, due to the large size of their latent space, their probabilistic inference is deemed intractable. We demonstrate that the VQ-VAE can be distilled into a tractable model by selecting a subset of latent variables with high probabilities. This simple strategy is particularly efficient, especially if the VQ-VAE underutilizes its latent space, which is, indeed, very often the case. We frame the distilled model as a probabilistic circuit, and show that it preserves expressiveness of the VQ-VAE while providing tractable probabilistic inference. Experiments illustrate competitive performance in density estimation and conditional generation tasks, challenging the view of the VQ-VAE as an inherently intractable model.

[858] Evaluating the stability of model explanations in instance-dependent cost-sensitive credit scoring

Matteo Ballegeer, Matthias Bogaert, Dries F. Benoit

Main category: cs.LG

TL;DR: IDCS classifiers improve cost-efficiency in credit scoring but produce less stable model explanations (SHAP/LIME) compared to traditional models, especially with class imbalance, creating a trade-off between cost optimization and interpretability.

DetailsMotivation: Address the unexplored impact of instance-dependent cost-sensitive classifiers on explanation stability despite increasing regulatory demands for transparency in credit scoring.

Method: Evaluated LIME and SHAP stability on IDCS models using four credit scoring datasets, assessed discriminatory power and cost-efficiency with a novel metric, and investigated feature importance stability under varying class imbalance through controlled resampling.

Result: IDCS classifiers improve cost-efficiency but produce significantly less stable explanations than traditional models, particularly as class imbalance increases.

Conclusion: There’s a critical trade-off between cost optimization and interpretability in credit scoring, highlighting the need to address stability issues in IDCS classifiers to ensure their cost advantages aren’t undermined by unreliable explanations.

Abstract: Instance-dependent cost-sensitive (IDCS) classifiers offer a promising approach to improving cost-efficiency in credit scoring by tailoring loss functions to instance-specific costs. However, the impact of such loss functions on the stability of model explanations remains unexplored in literature, despite increasing regulatory demands for transparency. This study addresses this gap by evaluating the stability of Local Interpretable Model-agnostic Explanations (LIME) and SHapley Additive exPlanations (SHAP) when applied to IDCS models. Using four publicly available credit scoring datasets, we first assess the discriminatory power and cost-efficiency of IDCS classifiers, introducing a novel metric to enhance cross-dataset comparability. We then investigate the stability of SHAP and LIME feature importance rankings under varying degrees of class imbalance through controlled resampling. Our results reveal that while IDCS classifiers improve cost-efficiency, they produce significantly less stable explanations compared to traditional models, particularly as class imbalance increases, highlighting a critical trade-off between cost optimization and interpretability in credit scoring. Amid increasing regulatory scrutiny on explainability, this research underscores the pressing need to address stability issues in IDCS classifiers to ensure that their cost advantages are not undermined by unstable or untrustworthy explanations.

[859] Accelerating PDE Solvers with Equation-Recast Neural Operator Preconditioning

Qiyun Cheng, Md Hossain Sahadath, Huihua Yang, Shaowu Pan, Wei Ji

Main category: cs.LG

TL;DR: MD-PNOP framework accelerates parametric PDE solvers by using neural operators as preconditioners, enabling extrapolation from single-parameter training to diverse configurations without retraining while preserving physical constraints.

DetailsMotivation: Traditional numerical solvers for PDEs have high computational overhead that limits large-scale parametric studies and design optimization. Neural operators struggle with extrapolation beyond training data.

Method: Recasts residual from parameter deviation as additional source term, uses any trained neural operator to refine solutions offline, embeds predictions as improved initial guesses in iterative PDE solvers while enforcing governing equations.

Result: Neural operators trained on single constant parameters successfully accelerate solutions with heterogeneous, sinusoidal, and discontinuous distributions. Achieves ~50% computational time reduction while maintaining full order fidelity across various problem types.

Conclusion: MD-PNOP establishes a new paradigm for accelerating parametric PDE solvers, overcoming neural operator extrapolation limitations while preserving physical constraints and interpretability, making it architecture-agnostic and broadly applicable.

Abstract: The computational overhead of traditional numerical solvers for partial differential equations (PDEs) remains a critical bottleneck for large-scale parametric studies and design optimization. We introduce a Minimal-Data Parametric Neural Operator Preconditioning (MD-PNOP) framework, which establishes a new paradigm for accelerating parametric PDE solvers while strictly preserving physical constraints. The key idea is to recast the residual from parameter deviation as additional source term, where any trained neural operator can be used to refine the solution in an offline fashion. This directly addresses the fundamental extrapolation limitation of neural operators, enabling extrapolative generalization of any neural operator trained at a single parameter setting across a wide range of configurations without any retraining. The neural operator predictions are then embedded into iterative PDE solvers as improved initial guesses, thereby reducing convergence iterations without sacrificing accuracy. Unlike purely data-driven approaches, MD-PNOP guarantees that the governing equations remain fully enforced, eliminating concerns about loss of physics or interpretability. The framework is architecture-agnostic and is demonstrated using both Deep Operator Networks (DeepONet) and Fourier Neural Operators (FNO) for Boltzmann transport equation solvers in neutron transport applications. We demonstrated that neural operators trained on a single set of constant parameters successfully accelerate solutions with heterogeneous, sinusoidal, and discontinuous parameter distributions. Besides, MD-PNOP consistently achieves ~50% reduction in computational time while maintaining full order fidelity for fixed-source, single-group eigenvalue, and multigroup coupled eigenvalue problems.

[860] The Geometry of Nonlinear Reinforcement Learning

Nikola Milosevic, Nico Scherf

Main category: cs.LG

TL;DR: A unified geometric framework that integrates reward maximization, safe exploration, and intrinsic motivation as instances of a single optimization problem on the space of achievable long-term behavior in reinforcement learning.

DetailsMotivation: To provide a unified perspective that connects traditionally separate RL objectives (reward maximization, safe exploration, intrinsic motivation) through a geometric framework, enabling generalization of classical methods to nonlinear utilities and convex constraints.

Method: Develops a geometric framework that views various RL goals as optimization problems on the space of achievable long-term behavior. Generalizes classical methods like policy mirror descent, natural policy gradient, and trust-region algorithms to handle nonlinear utilities and convex constraints.

Result: The framework successfully captures diverse objectives including robustness, safety, exploration, and diversity within a single unified perspective, showing how classical RL algorithms can be extended to these broader problem settings.

Conclusion: This geometric approach provides a powerful unifying framework for multiple RL objectives and highlights important open challenges at the intersection of geometry and deep reinforcement learning that warrant further research.

Abstract: Reward maximization, safe exploration, and intrinsic motivation are often studied as separate objectives in reinforcement learning (RL). We present a unified geometric framework, that views these goals as instances of a single optimization problem on the space of achievable long-term behavior in an environment. Within this framework, classical methods such as policy mirror descent, natural policy gradient, and trust-region algorithms naturally generalize to nonlinear utilities and convex constraints. We illustrate how this perspective captures robustness, safety, exploration, and diversity objectives, and outline open challenges at the interface of geometry and deep RL.

[861] Benchmarking Optimizers for Large Language Model Pretraining

Andrei Semenov, Matteo Pagliardini, Martin Jaggi

Main category: cs.LG

TL;DR: Comprehensive evaluation of recent optimization techniques for LLMs across standardized pretraining scenarios to provide practical guidance and highlight future research directions.

DetailsMotivation: The proliferation of novel optimization methods for LLMs with various claims (faster convergence, hyperparameter independence) makes direct comparisons challenging due to diverse experimental protocols.

Method: Systematic evaluation across standardized LLM pretraining scenarios with careful tuning of each method, varying model size, batch size, and training duration.

Result: Provides guidance on which optimizer is best suited for different scenarios and highlights promising directions for future optimization research.

Conclusion: The study offers practical recommendations for practitioners and researchers, with released code and fully reproducible experiments to support future method development and rigorous benchmarking.

Abstract: The recent development of Large Language Models (LLMs) has been accompanied by an effervescence of novel ideas and methods to better optimize the loss of deep learning models. Claims from those methods are myriad: from faster convergence to removing reliance on certain hyperparameters. However, the diverse experimental protocols used to validate these claims make direct comparisons between methods challenging. This study presents a comprehensive evaluation of recent optimization techniques across standardized LLM pretraining scenarios, systematically varying model size, batch size, and training duration. Through careful tuning of each method, we provide guidance to practitioners on which optimizer is best suited for each scenario. For researchers, our work highlights promising directions for future optimization research. Finally, by releasing our code and making all experiments fully reproducible, we hope our efforts can help the development and rigorous benchmarking of future methods.

[862] Hierarchical Motion Captioning Utilizing External Text Data Source

Clayton Leite, Yu Xiao

Main category: cs.LG

TL;DR: Novel hierarchical approach for motion captioning that uses LLMs to generate detailed descriptions and retrieval from external text sources to improve accuracy, achieving 6-50% performance gains.

DetailsMotivation: Existing motion captioning methods require high-level annotated motion data which is scarce in datasets, and current datasets lack low-level motion descriptions needed for detailed captioning.

Method: Two-step approach: 1) Use large language models to create detailed descriptions from high-level captions, then retrain motion-to-text models; 2) Retrieval-based mechanism that aligns detailed captions with candidate high-level captions from external text sources combined with motion features.

Result: Achieved 6% to 50% improvement in average performance across BLEU-1, BLEU-4, CIDEr, and ROUGE-L metrics compared to state-of-the-art M2T-Interpretable on three datasets (HumanML3D, KIT, and BOTH57M).

Conclusion: The method successfully leverages external text knowledge to enhance motion captioning accuracy, particularly for movements not covered in existing datasets, demonstrating significant performance improvements over current state-of-the-art approaches.

Abstract: This paper introduces a novel approach to enhance existing motion captioning methods, which directly map representations of movement to high-level descriptive captions (e.g., a person doing jumping jacks"). The existing methods require motion data annotated with high-level descriptions (e.g., jumping jacks"). However, such data is rarely available in existing motion-text datasets, which additionally do not include low-level motion descriptions. To address this, we propose a two-step hierarchical approach. First, we employ large language models to create detailed descriptions corresponding to each high-level caption that appears in the motion-text datasets (e.g., jumping while synchronizing arm extensions with the opening and closing of legs" for jumping jacks"). These refined annotations are used to retrain motion-to-text models to produce captions with low-level details. Second, we introduce a pioneering retrieval-based mechanism. It aligns the detailed low-level captions with candidate high-level captions from additional text data sources, and combine them with motion features to fabricate precise high-level captions. Our methodology is distinctive in its ability to harness knowledge from external text sources to greatly increase motion captioning accuracy, especially for movements not covered in existing motion-text datasets. Experiments on three distinct motion-text datasets (HumanML3D, KIT, and BOTH57M) demonstrate that our method achieves an improvement in average performance (across BLEU-1, BLEU-4, CIDEr, and ROUGE-L) ranging from 6% to 50% compared to the state-of-the-art M2T-Interpretable.

[863] Prior-Guided Flow Matching for Target-Aware Molecule Design with Learnable Atom Number

Jingyuan Zhou, Hao Qian, Shikui Tu, Lei Xu

Main category: cs.LG

TL;DR: PAFlow is a novel target-aware molecular generation model that uses flow matching and protein-ligand interaction guidance to generate 3D molecules with high binding affinity to target proteins, addressing issues of unstable probability dynamics and molecule-pocket size mismatch.

DetailsMotivation: Current structure-based drug design models suffer from unstable probability dynamics and mismatch between generated molecule size and protein pocket geometry, leading to inconsistent quality and off-target effects.

Method: PAFlow uses flow matching framework with conditional flow matching for discrete atom types, incorporates protein-ligand interaction predictor for guidance, and includes learnable atom number predictor based on protein pocket information.

Result: Achieves state-of-the-art binding affinity (up to -8.31 Avg. Vina Score) on CrossDocked2020 benchmark while maintaining favorable molecular properties.

Conclusion: PAFlow effectively addresses key challenges in structure-based drug design by providing stable generation, better molecule-pocket size alignment, and improved binding affinity through interaction guidance and atom number prediction.

Abstract: Structure-based drug design (SBDD), aiming to generate 3D molecules with high binding affinity toward target proteins, is a vital approach in novel drug discovery. Although recent generative models have shown great potential, they suffer from unstable probability dynamics and mismatch between generated molecule size and the protein pockets geometry, resulting in inconsistent quality and off-target effects. We propose PAFlow, a novel target-aware molecular generation model featuring prior interaction guidance and a learnable atom number predictor. PAFlow adopts the efficient flow matching framework to model the generation process and constructs a new form of conditional flow matching for discrete atom types. A protein-ligand interaction predictor is incorporated to guide the vector field toward higher-affinity regions during generation, while an atom number predictor based on protein pocket information is designed to better align generated molecule size with target geometry. Extensive experiments on the CrossDocked2020 benchmark show that PAFlow achieves a new state-of-the-art in binding affinity (up to -8.31 Avg. Vina Score), simultaneously maintains favorable molecular properties.

[864] Unsupervised Identification and Replay-based Detection (UIRD) for New Category Anomaly Detection in ECG Signal

Zhangyue Shi, Zekai Wang, Yuxuan Li

Main category: cs.LG

TL;DR: Proposed a pseudo-replay based semi-supervised continual learning framework for ECG anomaly detection that addresses class imbalance and storage limitations by using GAN-based novel pattern detection and pseudo data generation instead of storing historical data.

DetailsMotivation: Address class imbalance issues in ECG signal analysis due to limited samples of certain types, and overcome storage limitations from growing patient data volumes while maintaining accurate anomaly detection performance.

Method: Two-component framework: 1) Unsupervised GAN-based identification for novel pattern detection, 2) Pseudo replay-based learning strategy using generators to learn data distributions and synthesize pseudo data for previous classes when new tasks arise.

Result: Validated on four public ECG datasets, showing promising performance in identifying novel anomalies while maintaining good detection of existing ECG signals.

Conclusion: The proposed framework effectively enhances ECG anomaly detection performance while addressing storage limitations through pseudo-replay continual learning, making it suitable for real-world clinical applications with growing data volumes.

Abstract: In clinical practice, automatic analysis of electrocardiogram (ECG) is widely applied to identify irregular heart rhythms and other electrical anomalies of the heart, enabling timely intervention and potentially improving clinical outcomes. However, due to the limited samples in certain types of ECG signals, the class imbalance issues pose a challenge for ECG-based detection. In addition, as the volume of patient data grows, long-term storage of all historical data becomes increasingly burdensome as training samples to recognize new patterns and classify existing ECG signals accurately. Therefore, to enhance the performance of anomaly detection while addressing storage limitations, we propose a pseudo-replay based semi-supervised continual learning framework, which consists of two components: unsupervised identification and replay-based detection. For unsupervised identification, an unsupervised generative adversarial network (GAN)-based framework is integrated to detect novel patterns. Besides, instead of directly storing all historical data, a pseudo replay-based learning strategy is proposed which utilizes a generator to learn the data distribution for each individual task. When a new task arises, the generator synthesizes pseudo data representative of previous learnt classes, enabling the model to detect both the existed patterns and the newly presented anomalies. The effectiveness of the proposed framework is validated in four public ECG datasets, which leverages supervised classification problems for anomaly detection. The experimental results show that the developed approach is very promising in identifying novel anomalies while maintaining good performance on detecting existing ECG signals.

[865] Prediction, Generation of WWTPs microbiome community structures and Clustering of WWTPs various feature attributes using DE-BP model, SiTime-GAN model and DPNG-EPMC ensemble clustering algorithm with modulation of microbial ecosystem health

Mingzhi Dai, Weiwei Cai, Xiang Feng, Huiqun Yu, Weibin Guo, Miao Guo

Main category: cs.LG

TL;DR: This paper presents a machine learning framework using DE-BP neural networks for predicting microbial composition in wastewater treatment plants, introduces DPNG-EPMC clustering for WWTP analysis, and employs SiTime-GAN for synthetic data generation to understand activated sludge communities.

DetailsMotivation: Microbiome engineering faces significant challenges in achieving desired control and improvements, particularly in engineered ecosystems like wastewater treatment where understanding microbial composition is crucial for optimization.

Method: Used backpropagation neural network optimized through differential evolution (DE-BP) for microbial composition prediction, developed novel DPNG-EPMC clustering algorithm for WWTP analysis, and employed Similar Time Generative Adversarial Networks (SiTime-GAN) for synthetic data generation.

Result: DE-BP model provided superior predictions of microbial composition, DPNG-EPMC effectively analyzed WWTPs across various feature attributes, and SiTime-GAN generated valuable incremental synthetic data for understanding activated sludge communities.

Conclusion: The developed machine learning framework successfully predicts microbial communities, analyzes wastewater treatment plants, and generates synthetic data, contributing to better understanding of factors influencing activated sludge communities and advancing microbiome engineering capabilities.

Abstract: Microbiomes not only underpin Earth’s biogeochemical cycles but also play crucial roles in both engineered and natural ecosystems, such as the soil, wastewater treatment, and the human gut. However, microbiome engineering faces significant obstacles to surmount to deliver the desired improvements in microbiome control. Here, we use the backpropagation neural network (BPNN), optimized through differential evolution (DE-BP), to predict the microbial composition of activated sludge (AS) systems collected from wastewater treatment plants (WWTPs) located worldwide. Furthermore, we introduce a novel clustering algorithm termed Directional Position Nonlinear Emotional Preference Migration Behavior Clustering (DPNG-EPMC). This method is applied to conduct a clustering analysis of WWTPs across various feature attributes. Finally, we employ the Similar Time Generative Adversarial Networks (SiTime-GAN), to synthesize novel microbial compositions and feature attributes data. As a result, we demonstrate that the DE-BP model can provide superior predictions of the microbial composition. Additionally, we show that the DPNG-EPMC can be applied to the analysis of WWTPs under various feature attributes. Finally, we demonstrate that the SiTime-GAN model can generate valuable incremental synthetic data. Our results, obtained through predicting the microbial community and conducting analysis of WWTPs under various feature attributes, develop an understanding of the factors influencing AS communities.

[866] Forward-Only Continual Learning

Jiao Chen, Jiayi He, Fangfang Chen, Zuohong Lv, Jianhua Tang

Main category: cs.LG

TL;DR: FoRo is a forward-only, gradient-free continual learning method that uses prompt tuning with CMA-ES optimization and knowledge encoding via random projection to prevent catastrophic forgetting in pre-trained models without backpropagation.

DetailsMotivation: Address catastrophic forgetting in continual learning with pre-trained models while avoiding computationally intensive backpropagation and gradient-based optimization for resource-constrained environments.

Method: Lightweight prompt tuning strategy with CMA-ES optimization for prompt embeddings, plus knowledge encoding mechanism using nonlinear random projection and recursive least squares for incremental classifier updates without revisiting prior data.

Result: Significantly reduces average forgetting and improves accuracy while reducing memory usage and run time, maintaining high knowledge retention across long task sequences.

Conclusion: FoRo provides an efficient and effective continual learning approach suitable for real-world multimedia applications where both computational efficiency and performance are critical.

Abstract: Catastrophic forgetting remains a central challenge in continual learning (CL) with pre-trained models. While existing approaches typically freeze the backbone and fine-tune a small number of parameters to mitigate forgetting, they still rely on iterative error backpropagation and gradient-based optimization, which can be computationally intensive and less suitable for resource-constrained environments. To address this, we propose FoRo, a forward-only, gradient-free continual learning method. FoRo consists of a lightweight prompt tuning strategy and a novel knowledge encoding mechanism, both designed without modifying the pre-trained model. Specifically, prompt embeddings are inserted at the input layer and optimized using the Covariance Matrix Adaptation Evolution Strategy (CMA-ES), which mitigates distribution shifts and extracts high-quality task representations. Subsequently, task-specific knowledge is encoded into a knowledge encoding matrix via nonlinear random projection and recursive least squares, enabling incremental updates to the classifier without revisiting prior data. Experiments show that FoRo significantly reduces average forgetting and improves accuracy. Thanks to forward-only learning, FoRo reduces memory usage and run time while maintaining high knowledge retention across long task sequences. These results suggest that FoRo could serve as a promising direction for exploring continual learning with pre-trained models, especially in real-world multimedia applications where both efficiency and effectiveness are critical.

[867] Graph Contrastive Learning versus Untrained Baselines: The Role of Dataset Size

Smayan Khanna, Doruk Efe Gökmen, Risi Kondor, Vincenzo Vitelli

Main category: cs.LG

TL;DR: GCL’s performance advantage over untrained baselines depends heavily on dataset size and task difficulty, with untrained models often matching or exceeding GCL on standard datasets, while GCL only pulls ahead on larger molecular datasets.

DetailsMotivation: To critically evaluate whether Graph Contrastive Learning (GCL) actually outperforms simple untrained baselines, given its widespread adoption and strong reported performance in various applications.

Method: Comparative analysis of GCL against untrained Graph Neural Networks, simple multilayer perceptrons, and handcrafted statistics across standard datasets, large molecular datasets (ogbg-molhiv), and synthetic datasets with varying complexity.

Result: GCL’s advantage is dataset-size dependent: untrained baselines rival or exceed GCL on standard datasets; GCL lags at small scales but pulls ahead beyond few thousand graphs on ogbg-molhiv (though gains plateau); performance scales logarithmically with dataset size and gap varies with task complexity.

Conclusion: Dataset size plays a crucial role in GCL performance evaluation, and future work should focus on designing GCL algorithms that avoid performance plateaus while considering dataset scale in benchmarks.

Abstract: Graph Contrastive Learning (GCL) has emerged as a leading paradigm for self- supervised learning on graphs, with strong performance reported on standardized datasets and growing applications ranging from genomics to drug discovery. We ask a basic question: does GCL actually outperform untrained baselines? We find that GCL’s advantage depends strongly on dataset size and task difficulty. On standard datasets, untrained Graph Neural Networks (GNNs), simple multilayer perceptrons, and even handcrafted statistics can rival or exceed GCL. On the large molecular dataset ogbg-molhiv, we observe a crossover: GCL lags at small scales but pulls ahead beyond a few thousand graphs, though this gain eventually plateaus. On synthetic datasets, GCL accuracy approximately scales with the logarithm of the number of graphs and its performance gap (compared with untrained GNNs) varies with respect to task complexity. Moving forward, it is crucial to identify the role of dataset size in benchmarks and applications, as well as to design GCL algorithms that avoid performance plateaus.

[868] Feynman-Kac-Flow: Inference Steering of Conditional Flow Matching to an Energy-Tilted Posterior

Konstantin Mark, Leonard Galustian, Maximilian P. -P. Kovar, Esther Heid

Main category: cs.LG

TL;DR: Feynman-Kac steering for Conditional Flow Matching enables precise control over generated samples by tilting outputs with energy potentials, solving challenging tasks like generating transition states with correct chirality in chemical reactions.

DetailsMotivation: While steering approaches exist for diffusion models, they haven't been extended to Conditional Flow Matching (CFM) despite the need for precise sample steering towards specific requirements in many applications.

Method: Formulate steering as tilting the output with an energy potential and derive Feynman-Kac steering specifically for Conditional Flow Matching approaches.

Result: Successfully evaluated on synthetic tasks including tilted distributions in high-dimensional spaces and demonstrated impact on generating transition states of chemical reactions with correct chirality, solving previously unsolved challenges.

Conclusion: Feynman-Kac steering for CFM provides an effective method for precise sample control, particularly valuable for complex applications like chemical reaction modeling where geometric constraints are critical.

Abstract: Conditional Flow Matching(CFM) represents a fast and high-quality approach to generative modelling, but in many applications it is of interest to steer the generated samples towards precise requirements. While steering approaches like gradient-based guidance, sequential Monte Carlo steering or Feynman-Kac steering are well established for diffusion models, they have not been extended to flow matching approaches yet. In this work, we formulate this requirement as tilting the output with an energy potential. We derive, for the first time, Feynman-Kac steering for CFM. We evaluate our approach on a set of synthetic tasks, including the generation of tilted distributions in a high-dimensional space, which is a particularly challenging case for steering approaches. We then demonstrate the impact of Feynman-Kac steered CFM on the previously unsolved challenge of generated transition states of chemical reactions with the correct chirality, where the reactants or products can have a different handedness, leading to geometric constraints of the viable reaction pathways connecting reactants and products. Code to reproduce this study is avaiable open-source at https://github.com/heid-lab/fkflow.

[869] Model Unmerging: Making Your Models Unmergeable for Secure Model Sharing

Zihao Wang, Enneng Yang, Lu Yin, Shiwei Liu, Li Shen

Main category: cs.LG

TL;DR: MergeLock is a protection mechanism that disrupts model parameters to prevent unauthorized merging of finetuned models while maintaining original functionality.

DetailsMotivation: Growing concerns about safety and rights infringement from unauthorized model merging of publicly available finetuned models, as existing detection methods fail to prevent illegal merging.

Method: Leverages Transformer attention mechanism symmetry by applying randomly sampled invertible matrices to QK and VO branches, keeping output unchanged but pushing model out of shared parameter space.

Result: Degrades merged model performance by over 95% when protected model is involved, across vision and language tasks. Protected models cannot be effectively recovered with low-cost restoration methods.

Conclusion: MergeLock effectively prevents unauthorized model merging while maintaining model functionality, providing robust protection against infringement and sensitive information leakage.

Abstract: Model merging leverages multiple finetuned expert models to construct a multi-task model with low cost, and is gaining increasing attention. However, as a growing number of finetuned models become publicly available, concerns about the safety of model merging have emerged. Unauthorized merging may infringe on developers’ rights and risk leaking sensitive personal information. Most existing methods focus on detecting whether a merged model originates from a specific source model, but fail to effectively prevent illegal merging. In this paper, we propose MergeLock, an active protection mechanism that disrupts model parameters to render them unmergeable, thereby directly preventing unauthorized model merging. Specifically, leveraging the inherent symmetry of the attention mechanism in Transformer-based models, we randomly sample two pairs of invertible matrices and apply them to the Query-Key (QK) and Value-Output (VO) branches. This transformation keeps the model’s output unchanged while pushing it away from the shared parameter space of other finetuned models. Extensive experiments across both vision and language tasks demonstrate that MergeLock can degrade the performance of merged models by over 95% when a protected model is involved in most cases, demonstrating its effectiveness. Moreover, we further demonstrate that merged models protected by MergeLock cannot be effectively recovered using low-cost restoration methods, further enhancing robustness against unauthorized merging. The code is available at https://github.com/hetailang/Merge-Lock.

[870] Direct Profit Estimation Using Uplift Modeling under Clustered Network Interference

Bram van den Akker

Main category: cs.LG

TL;DR: This paper introduces a practical methodology using AddIPW estimator for interference-aware uplift modeling in recommender systems, enabling gradient-based optimization of economic outcomes like incremental profit.

DetailsMotivation: Standard uplift modeling methods fail to account for interference effects where treating one item affects others, violating SUTVA and leading to suboptimal policies in real-world marketplaces.

Method: Uses Additive Inverse Propensity Weighting (AddIPW) estimator as a differentiable learning objective for gradient-based optimization, integrated with response transformation techniques to directly optimize economic outcomes.

Result: Simulations show the approach significantly outperforms interference-naive methods, especially as interference effects grow, and profit-centric strategies yield superior performance in identifying high-impact interventions.

Conclusion: Provides a practical path toward more profitable incentive personalization by bridging the gap between interference-aware estimators and uplift modeling optimization.

Abstract: Uplift modeling is a key technique for promotion optimization in recommender systems, but standard methods typically fail to account for interference, where treating one item affects the outcomes of others. This violation of the Stable Unit Treatment Value Assumption (SUTVA) leads to suboptimal policies in real-world marketplaces. Recent developments in interference-aware estimators such as Additive Inverse Propensity Weighting (AddIPW) have not found their way into the uplift modeling literature yet, and optimising policies using these estimators is not well-established. This paper proposes a practical methodology to bridge this gap. We use the AddIPW estimator as a differentiable learning objective suitable for gradient-based optimization. We demonstrate how this framework can be integrated with proven response transformation techniques to directly optimize for economic outcomes like incremental profit. Through simulations, we show that our approach significantly outperforms interference-naive methods, especially as interference effects grow. Furthermore, we find that adapting profit-centric uplift strategies within our framework can yield superior performance in identifying the highest-impact interventions, offering a practical path toward more profitable incentive personalization.

[871] Learning Longitudinal Stress Dynamics from Irregular Self-Reports via Time Embeddings

Louis Simon, Mohamed Chetouani

Main category: cs.LG

TL;DR: Ema2Vec - a novel time embedding method that improves stress prediction from irregularly spaced ecological momentary assessments by capturing time dependencies, outperforming standard baselines.

DetailsMotivation: Mobile/wearable sensing enables continuous affect monitoring, but irregular timing and missing data in self-reports make human state prediction challenging. Time dependencies in ecological momentary assessments need better modeling.

Method: Developed Ema2Vec time embedding method specifically designed to handle irregularly spaced self-reports and capture time dependencies within EMA sequences for longitudinal stress prediction.

Result: Outperformed standard stress prediction baselines using fixed-size daily windows and models trained directly on longitudinal sequences without time-aware representations.

Conclusion: Time embeddings are crucial for effectively modeling irregularly sampled longitudinal data, with Ema2Vec demonstrating superior performance in stress prediction tasks.

Abstract: The widespread adoption of mobile and wearable sensing technologies has enabled continuous and personalized monitoring of affect, mood disorders, and stress. When combined with ecological self-report questionnaires, these systems offer a powerful opportunity to explore longitudinal modeling of human behaviors. However, challenges arise from missing data and the irregular timing of self-reports, which make challenging the prediction of human states and behaviors. In this study, we investigate the use of time embeddings to capture time dependencies within sequences of Ecological Momentary Assessments (EMA). We introduce a novel time embedding method, Ema2Vec, designed to effectively handle irregularly spaced self-reports, and evaluate it on a new task of longitudinal stress prediction. Our method outperforms standard stress prediction baselines that rely on fixed-size daily windows, as well as models trained directly on longitudinal sequences without time-aware representations. These findings emphasize the importance of incorporating time embeddings when modeling irregularly sampled longitudinal data.

[872] One-Shot Clustering for Federated Learning Under Clustering-Agnostic Assumption

Maciej Krzysztof Zuziak, Roberto Pellungrini, Salvatore Rinzivillo

Main category: cs.LG

TL;DR: One-Shot Clustered Federated Learning (OCFL) is a clustering-agnostic algorithm that automatically detects the optimal moment for clustering in federated learning by computing cosine distance between client gradients and using temperature measures to detect model convergence.

DetailsMotivation: Clustered Federated Learning (CFL) addresses client clustering for personalized models but remains largely unexplored with different assumptions than standard FL. Current approaches lack automated timing detection for clustering.

Method: OCFL uses cosine distance between client gradients and a temperature measure to detect when the federated model starts converging, enabling automatic clustering timing. Tested with various one-shot clustering algorithms on 40+ tasks across 5 benchmark datasets.

Result: The approach demonstrates good performance in automated CFL without hyperparameter tuning. Density-based clustering methods prove highly efficient for differentiating neural network loss surfaces from different distributions. GradCAM local explanations provide insights into personalization-explainability relationships.

Conclusion: OCFL successfully automates clustering timing in federated learning, showing practical feasibility of gradient-based CFL methods and providing valuable insights into the connection between model personalization and local prediction explainability.

Abstract: Federated Learning (FL) is a widespread and well-adopted paradigm of decentralised learning that allows training one model from multiple sources without the need to transfer data between participating clients directly. Since its inception in 2015, it has been divided into numerous subfields that deal with application-specific issues, such as data heterogeneity or resource allocation. One such sub-field, Clustered Federated Learning (CFL), deals with the problem of clustering the population of clients into separate cohorts to deliver personalised models. Although a few remarkable works have been published in this domain, the problem remains largely unexplored, as its basic assumptions and settings differ slightly from those of standard FL. In this work, we present One-Shot Clustered Federated Learning (OCFL), a clustering-agnostic algorithm that can automatically detect the earliest suitable moment for clustering. Our algorithm is based on computing the cosine distance between the gradients of the clients and a temperature measure that detects when the federated model starts to converge. We empirically evaluate our methodology by testing various one-shot clustering algorithms for over forty different tasks on five benchmark datasets. Our experiments showcase the good performance of our approach when used to perform CFL in an automated manner without the need to adjust hyperparameters. We also revisit the practical feasibility of CFL algorithms based on the gradients of the clients, providing firm evidence of the high efficiency of density-based clustering methods when used to differentiate between the loss surfaces of neural networks trained on different distributions. Moreover, by inspecting the feasibility of local explanations generated with the help of GradCAM, we can provide more insights into the relationship between personalisation and the explainability of local predictions.

[873] Entropy-Driven Curriculum for Multi-Task Training in Human Mobility Prediction

Tianye Fang, Xuanshu Luo, Martin Werner

Main category: cs.LG

TL;DR: A unified training framework combining entropy-driven curriculum learning and multi-task learning for human mobility prediction, achieving state-of-the-art performance with faster convergence.

DetailsMotivation: Human mobility data's diverse complexity impedes model training, and exclusively predicting next locations neglects important movement determinants like distances and directions, leading to suboptimal results.

Method: Proposes entropy-driven curriculum learning that quantifies trajectory predictability using Lempel-Ziv compression to organize training from simple to complex trajectories, combined with multi-task learning that simultaneously optimizes location prediction alongside auxiliary distance and direction estimation.

Result: Achieves state-of-the-art performance on GEO-BLEU (0.354) and DTW (26.15) metrics with up to 2.92-fold convergence speed compared to training without curriculum learning, as demonstrated in HuMob Challenge experiments.

Conclusion: The integrated framework of entropy-driven curriculum learning and multi-task learning effectively addresses training inefficiencies and improves human mobility prediction accuracy by considering both location and movement patterns.

Abstract: The increasing availability of big mobility data from ubiquitous portable devices enables human mobility prediction through deep learning approaches. However, the diverse complexity of human mobility data impedes model training, leading to inefficient gradient updates and potential underfitting. Meanwhile, exclusively predicting next locations neglects implicit determinants, including distances and directions, thereby yielding suboptimal prediction results. This paper presents a unified training framework that integrates entropy-driven curriculum and multi-task learning to address these challenges. The proposed entropy-driven curriculum learning strategy quantifies trajectory predictability based on Lempel-Ziv compression and organizes training from simple to complex for faster convergence and enhanced performance. The multi-task training simultaneously optimizes the primary location prediction alongside auxiliary estimation of movement distance and direction for learning realistic mobility patterns, and improve prediction accuracy through complementary supervision signals. Extensive experiments conducted in accordance with the HuMob Challenge demonstrate that our approach achieves state-of-the-art performance on GEO-BLEU (0.354) and DTW (26.15) metrics with up to 2.92-fold convergence speed compared to training without curriculum learning.

[874] Effects of Distributional Biases on Gradient-Based Causal Discovery in the Bivariate Categorical Case

Tim Schwabe, Moritz Lange, Laurenz Wiskott, Maribel Acosta

Main category: cs.LG

TL;DR: Gradient-based causal discovery methods are vulnerable to distributional biases including Marginal Distribution Asymmetry and Marginal Distribution Shift Asymmetry, which can skew causal learning even in synthetic data. The paper demonstrates these biases and shows how eliminating competition between causal factorizations can make models robust.

DetailsMotivation: Gradient-based causal discovery shows great potential but can be susceptible to distributional biases in training data, which may lead to incorrect causal structure learning.

Method: The study uses bivariate categorical setup with Dirichlet priors and employs two simple models that learn marginal or conditional data distributions to examine how these biases affect gradient-based methods.

Result: The research demonstrates that both Marginal Distribution Asymmetry and Marginal Distribution Shift Asymmetry biases can occur even in controlled synthetic data and affect gradient-based causal discovery models. The biases can be controlled by eliminating competition between possible causal factorizations.

Conclusion: Eliminating competition between causal factorizations can make gradient-based causal discovery models robust to the identified distributional biases, improving their reliability in learning causal structures from data.

Abstract: Gradient-based causal discovery shows great potential for deducing causal structure from data in an efficient and scalable way. Those approaches however can be susceptible to distributional biases in the data they are trained on. We identify two such biases: Marginal Distribution Asymmetry, where differences in entropy skew causal learning toward certain factorizations, and Marginal Distribution Shift Asymmetry, where repeated interventions cause faster shifts in some variables than in others. For the bivariate categorical setup with Dirichlet priors, we illustrate how these biases can occur even in controlled synthetic data. To examine their impact on gradient-based methods, we employ two simple models that derive causal factorizations by learning marginal or conditional data distributions - a common strategy in gradient-based causal discovery. We demonstrate how these models can be susceptible to both biases. We additionally show how the biases can be controlled. An empirical evaluation of two related, existing approaches indicates that eliminating competition between possible causal factorizations can make models robust to the presented biases.

[875] Relative Trajectory Balance is equivalent to Trust-PCL

Tristan Deleu, Padideh Nouri, Yoshua Bengio, Doina Precup

Main category: cs.LG

TL;DR: Establishes equivalence between Relative Trajectory Balance (RTB) from GFlowNets and Trust-PCL RL method, showing both achieve comparable performance in KL-regularized fine-tuning of sequential generative models.

DetailsMotivation: To bridge the theoretical gap between GFlowNets' RTB objective and KL-regularized reinforcement learning methods, providing a unified perspective on fine-tuning approaches for generative models.

Method: Theoretical analysis establishing equivalence between Relative Trajectory Balance (GFlowNets) and Trust-PCL (RL method), followed by empirical validation using illustrative examples from prior RTB work.

Result: Demonstrated that RTB and KL-regularized RL methods achieve comparable performance, contrary to previous reports that suggested RTB superiority.

Conclusion: RTB can be situated within the broader theoretical framework of KL-regularized RL, offering alternative perspectives and methods for fine-tuning sequential generative models.

Abstract: Recent progress in generative modeling has highlighted the importance of Reinforcement Learning (RL) for fine-tuning, with KL-regularized methods in particular proving to be highly effective for both autoregressive and diffusion models. Complementing this line of work, the Relative Trajectory Balance (RTB) objective was recently introduced in the context of Generative Flow Networks (GFlowNets) to serve the same role of improving fine-tuning in sequential generative models. Building on prior work linking GFlowNets and maximum-entropy RL, we establish in this paper an equivalence between RTB and Trust-PCL, an off-policy RL method with KL regularization. This equivalence situates RTB within the broader theoretical landscape of KL-regularized RL, and clarifies its relationship to earlier methods. Leveraging this insight, we revisit an illustrative example from the RTB paper and show that KL-regularized RL methods achieve comparable performance, offering an alternative perspective to what was previously reported.

[876] REVELIO – Universal Multimodal Task Load Estimation for Cross-Domain Generalization

Maximilian P. Oppelt, Andreas Foltyn, Nadine R. Lang-Richter, Bjoern M. Eskofier

Main category: cs.LG

TL;DR: This paper introduces a multimodal dataset for cognitive load detection that combines established benchmarks with real-world gaming applications, showing that multimodal approaches outperform unimodal ones but models struggle with cross-domain generalization.

DetailsMotivation: Current task load detection models lack generalizability beyond narrow experimental domains, and there's a gap in evaluating model robustness and transferability in real-world scenarios.

Method: Created a new multimodal dataset extending cognitive load detection benchmarks with real-world gaming using n-back test foundation. Used objective performance, NASA-TLX ratings, and task-level design for annotations. Evaluated state-of-the-art models (xLSTM, ConvNeXt, Transformer) across multiple modalities and domains.

Result: Multimodal approaches consistently outperformed unimodal baselines, with modality and architecture impact varying by application. Models trained on one domain showed reduced performance when transferred to novel applications.

Conclusion: The findings provide robust baselines and insights for developing more generalizable cognitive load detection systems, advancing research in human-computer interaction and adaptive systems, though cross-domain generalization remains challenging.

Abstract: Task load detection is essential for optimizing human performance across diverse applications, yet current models often lack generalizability beyond narrow experimental domains. While prior research has focused on individual tasks and limited modalities, there remains a gap in evaluating model robustness and transferability in real-world scenarios. This paper addresses these limitations by introducing a new multimodal dataset that extends established cognitive load detection benchmarks with a real-world gaming application, using the $n$-back test as a scientific foundation. Task load annotations are derived from objective performance, subjective NASA-TLX ratings, and task-level design, enabling a comprehensive evaluation framework. State-of-the-art end-to-end model, including xLSTM, ConvNeXt, and Transformer architectures are systematically trained and evaluated on multiple modalities and application domains to assess their predictive performance and cross-domain generalization. Results demonstrate that multimodal approaches consistently outperform unimodal baselines, with specific modalities and model architectures showing varying impact depending on the application subset. Importantly, models trained on one domain exhibit reduced performance when transferred to novel applications, underscoring remaining challenges for universal cognitive load estimation. These findings provide robust baselines and actionable insights for developing more generalizable cognitive load detection systems, advancing both research and practical implementation in human-computer interaction and adaptive systems.

[877] Distilled Pretraining: A modern lens of Data, In-Context Learning and Test-Time Scaling

Sachin Goyal, David Lopez-Paz, Kartik Ahuja

Main category: cs.LG

TL;DR: Distillation in LLM pretraining improves test-time scaling but impairs in-context learning, particularly induction heads. Analysis using bigram models reveals underlying mechanisms and provides practical design guidance.

DetailsMotivation: To understand how distillation affects modern LLM capabilities like test-time scaling and in-context learning, which remain underexplored despite distillation's renewed prominence in recent models like Llama-3.2 and Gemma.

Method: Studied pretraining with distillation, analyzed test-time scaling and in-context learning capabilities, and used a bigram model sandbox to isolate the principal factors behind the observed effects.

Result: Distillation yields models with remarkably better test-time scaling but impairs in-context learning capabilities, particularly those modeled via induction heads. The bigram model analysis helped isolate the common underlying factor.

Conclusion: The findings provide insights into the trade-offs of distillation and offer guidance on pretraining design choices to help practitioners balance improved scaling with maintained in-context learning capabilities.

Abstract: In the past year, distillation has seen a renewed prominence in large language model (LLM) pretraining, exemplified by the Llama-3.2 and Gemma model families. While distillation has historically been shown to improve statistical modeling, its effects on new paradigms that are key to modern LLMs, such as test-time scaling and in-context learning, remain underexplored. In this work, we make three main contributions. First, we show that pretraining with distillation yields models that exhibit remarkably better test-time scaling. Second, we observe that this benefit comes with a trade-off: distillation impairs in-context learning capabilities, particularly the one modeled via induction heads. Third, to demystify these findings, we study distilled pretraining in a sandbox of a bigram model, which helps us isolate the common principal factor behind our observations. Finally, using these insights, we shed light on various design choices for pretraining that should help practitioners going forward.

[878] Efficient Transformer-Inspired Variants of Physics-Informed Deep Operator Networks

Zhi-Feng Wei, Wenqian Chen, Panos Stinis

Main category: cs.LG

TL;DR: Transformer-inspired DeepONet variants with bidirectional cross-conditioning between branch and trunk networks achieve improved accuracy and training efficiency compared to existing DeepONet architectures across multiple PDE benchmarks.

DetailsMotivation: To develop more efficient and accurate operator learning frameworks for PDEs that combine the simplicity of vanilla DeepONets with the accuracy of modified DeepONets, while enabling dynamic dependencies through cross-conditioning.

Method: Proposed Transformer-inspired DeepONet variants that introduce bidirectional cross-conditioning - injecting query-point information into branch network and input-function information into trunk network, preserving vanilla DeepONet’s efficiency in a non-intrusive manner.

Result: Experiments on four PDE benchmarks (advection, diffusion-reaction, Burgers’, Korteweg-de Vries equations) show variants match or surpass modified DeepONet accuracy with improved training efficiency. Best variant for each equation aligns with equation’s characteristics, indicating cross-conditioning effectiveness depends on underlying physics.

Conclusion: The proposed cross-conditioning approach provides a flexible framework for operator learning that adapts to different PDE characteristics while maintaining efficiency, with statistical validation confirming robustness across various PDE types.

Abstract: Operator learning has emerged as a promising tool for accelerating the solution of partial differential equations (PDEs). The Deep Operator Networks (DeepONets) represent a pioneering framework in this area: the “vanilla” DeepONet is valued for its simplicity and efficiency, while the modified DeepONet achieves higher accuracy at the cost of increased training time. In this work, we propose a series of Transformer-inspired DeepONet variants that introduce bidirectional cross-conditioning between the branch and trunk networks in DeepONet. Query-point information is injected into the branch network and input-function information into the trunk network, enabling dynamic dependencies while preserving the simplicity and efficiency of the “vanilla” DeepONet in a non-intrusive manner. Experiments on four PDE benchmarks – advection, diffusion-reaction, Burgers’, and Korteweg-de Vries equations – show that for each case, there exists a variant that matches or surpasses the accuracy of the modified DeepONet while offering improved training efficiency. Moreover, the best-performing variant for each equation aligns naturally with the equation’s underlying characteristics, suggesting that the effectiveness of cross-conditioning depends on the characteristics of the equation and its underlying physics. To ensure robustness, we validate the effectiveness of our variants through a range of rigorous statistical analyses, among them the Wilcoxon Two One-Sided Test, Glass’s Delta, and Spearman’s rank correlation.

[879] Reinforcement Learning for Machine Learning Engineering Agents

Sherry Yang, Joy He-Yueya, Percy Liang

Main category: cs.LG

TL;DR: RL-trained smaller models outperform static larger models in ML engineering tasks by addressing variable-duration actions and providing partial credit rewards.

DetailsMotivation: Existing agents using static language models don't improve with experience, while RL-trained weaker models could potentially outperform larger static models.

Method: Proposed duration-aware gradient updates for variable-time actions and environment instrumentation with print statements to provide partial credit rewards using a separate static model.

Result: RL-trained Qwen2.5-3B model outperforms Claude-3.5-Sonnet by 22% average across 12 Kaggle tasks on MLEBench.

Conclusion: RL-based agents with proper reward engineering can surpass static larger models, demonstrating the value of learning from experience in ML engineering tasks.

Abstract: Existing agents for solving tasks such as ML engineering rely on prompting powerful language models. As a result, these agents do not improve with more experience. In this paper, we show that agents backed by weaker models that improve via reinforcement learning (RL) can outperform agents backed by much larger, but static models. We identify two major challenges with RL in this setting. First, actions can take a variable amount of time (e.g., executing code for different solutions), which leads to asynchronous policy gradient updates that favor faster but suboptimal solutions. To tackle variable-duration actions, we propose duration- aware gradient updates in a distributed asynchronous RL framework to amplify high-cost but high-reward actions. Second, using only test split performance as a reward provides limited feedback. A program that is nearly correct is treated the same as one that fails entirely. To address this, we propose environment instrumentation to offer partial credit, distinguishing almost-correct programs from those that fail early (e.g., during data loading). Environment instrumentation uses a separate static language model to insert print statement to an existing program to log the agent’s experimental progress, from which partial credit can be extracted as reward signals for learning. Our experimental results on MLEBench suggest that performing gradient updates on a much smaller model (Qwen2.5-3B) trained with RL outperforms prompting a much larger model (Claude-3.5-Sonnet) with agent scaffolds, by an average of 22% across 12 Kaggle tasks.

[880] Evaluating Cumulative Spectral Gradient as a Complexity Measure

Haji Gul, Abdul Ghani Naim, Ajaz Ahmad Bhat

Main category: cs.LG

TL;DR: CSG metric fails to scale with class count and shows weak correlation with performance metrics in KG link prediction tasks, contrary to original claims.

DetailsMotivation: To rigorously evaluate the Cumulative Spectral Gradient (CSG) metric's behavior on knowledge graph link prediction benchmarks, testing its claimed scalability and correlation with downstream performance.

Method: Conducted experiments on standard KG benchmarks (FB15k-237, WN18RR) using multi-class tail prediction, analyzing CSG sensitivity to parameters M (Monte Carlo samples) and K (nearest neighbors).

Result: CSG is highly sensitive to parameter K and does not scale with number of classes; shows weak/no correlation with established metrics like mean reciprocal rank (MRR); stability and predictive power break down in link prediction settings.

Conclusion: Current CSG complexity measure is unreliable for KG link prediction evaluation, highlighting the need for more robust, classifier-agnostic complexity measures in this domain.

Abstract: Accurate estimation of dataset complexity is crucial for evaluating and comparing link prediction models for knowledge graphs (KGs). The Cumulative Spectral Gradient (CSG) metric derived from probabilistic divergence between classes within a spectral clustering framework was proposed as a dataset complexity measure that (1) naturally scales with the number of classes and (2) correlates strongly with downstream classification performance. In this work, we rigorously assess CSG behavior on standard knowledge graph link prediction benchmarks a multi class tail prediction task, using two key parameters governing its computation, M, the number of Monte Carlo sampled points per class, and K, the number of nearest neighbors in the embedding space. Contrary to the original claims, we find that (1) CSG is highly sensitive to the choice of K and therefore does not inherently scale with the number of target classes, and (2) CSG values exhibit weak or no correlation with established performance metrics such as mean reciprocal rank (MRR). Through experiments on FB15k 237, WN18RR, and other standard datasets, we demonstrate that CSG purported stability and generalization predictive power break down in link prediction settings. Our results highlight the need for more robust, classifier agnostic complexity measures in KG link prediction evaluation.

[881] Robust Anomaly Detection through Multi-Modal Autoencoder Fusion for Small Vehicle Damage Detection

Sara Khan, Mehmed Yüksel, Frank Kirchner

Main category: cs.LG

TL;DR: Novel multi-modal anomaly detection system using IMUs and microphones for real-time vehicle wear and tear detection, achieving 92% ROC-AUC performance.

DetailsMotivation: Current manual inspections are labor-intensive and error-prone, while image-based methods struggle with real-time performance and underbody damage detection due to limited visual access.

Method: Multi-modal autoencoder-based architecture with sensors (IMUs and microphones) integrated into a compact windshield-mounted device. Developed ensemble pooling multi-modal model variants.

Result: Achieved 92% ROC-AUC score, outperforming unimodal and state-of-the-art methods, demonstrating effectiveness for real-time damage detection.

Conclusion: The approach enables real-time vehicle damage detection without resource-intensive sensors and can be extended to automotive safety systems and autonomous vehicle collision detection.

Abstract: Wear and tear detection in fleet and shared vehicle systems is a critical challenge, particularly in rental and car-sharing services, where minor damage, such as dents, scratches, and underbody impacts, often goes unnoticed or is detected too late. Currently, manual inspection methods are the default approach but are labour intensive and prone to human error. In contrast, state-of-the-art image-based methods struggle with real-time performance and are less effective at detecting underbody damage due to limited visual access and poor spatial coverage. This work introduces a novel multi-modal architecture based on anomaly detection to address these issues. Sensors such as IMUs and microphones are integrated into a compact device mounted on the vehicle’s windshield. This approach supports real-time damage detection while avoiding the need for highly resource-intensive sensors. We developed multiple variants of multi-modal autoencoder-based architectures and evaluated them against unimodal and state-of-the-art methods. Our ensemble pooling multi-modal model achieved the highest performance, with a Receiver Operating Characteristic-Area Under Curve (ROC-AUC) of 92%, demonstrating its effectiveness in real-world applications. This approach can also be extended to other applications, such as improving automotive safety - where it can integrate with airbag systems for efficient deployment - and helping autonomous vehicles by complementing other sensors in collision detection.

[882] DynaGuard: A Dynamic Guardrail Model With User-Defined Policies

Monte Hoover, Vatsal Baherwani, Neel Jain, Khalid Saifullah, Joseph Vincent, Chirag Jain, Melissa Kazemi Rad, C. Bayan Bruss, Ashwinee Panda, Tom Goldstein

Main category: cs.LG

TL;DR: Dynamic guardian models that evaluate text based on user-defined policies instead of predefined static categories, offering comparable accuracy to static models while handling free-form policies efficiently.

DetailsMotivation: Standard guardian models like LlamaGuard only detect predefined static harm categories, limiting their usefulness across different application domains that require customized policies.

Method: Proposed dynamic guardian models that can evaluate text based on user-defined policies, with two approaches: fast detection of policy violations and chain-of-thought reasoning that articulates and justifies model outputs.

Result: Dynamic guardian models match static models in detection accuracy for static harm categories while identifying violations of free-form policies with accuracy comparable to frontier reasoning models, but in a fraction of the time.

Conclusion: Dynamic guardian models provide flexible, efficient policy enforcement for diverse application domains, overcoming the limitations of static predefined categories while maintaining high accuracy and speed.

Abstract: Guardian models are used to supervise and moderate the outputs of user-facing chatbots, enforcing guardrails and detecting bad behaviors. Standard guardian models like LlamaGuard detect predefined, static categories of harms. We propose dynamic guardian models that evaluate text based on user-defined policies, making them useful for different application domains that are not addressed by standard guardian models. Our dynamic guardian models can be used for fast detection of policy violations or with chain-of-thought reasoning that articulates and justifies the model outputs. Our dynamic guardian models match static models in detection accuracy for static harm categories while identifying violations of free-form policies with accuracy comparable to frontier reasoning models in a fraction of the time.

[883] Succeed or Learn Slowly: Sample Efficient Off-Policy Reinforcement Learning for Mobile App Control

Georgios Papoudakis, Thomas Coste, Jianye Hao, Jun Wang, Kun Shao

Main category: cs.LG

TL;DR: SoLS is a novel off-policy RL algorithm that improves sample efficiency for fine-tuning foundation models in UI navigation tasks by applying direct updates for positive samples and conservative regularized updates for negative samples, outperforming existing methods by at least 17% with significantly faster inference.

DetailsMotivation: Address challenges in RL using foundation models for multi-turn tasks, particularly sparse reward settings and policy gradient updates that can harm model performance when learning from negative samples.

Method: Succeed or Learn Slowly (SoLS) algorithm with modified off-policy actor-critic approach: direct policy updates for positive samples, conservative regularized updates for negative samples, augmented with Successful Transition Replay (STR) to prioritize learning from successful interactions.

Result: Significantly outperforms existing methods on AndroidWorld benchmark (at least 17% relative increase), requires substantially fewer computational resources than GPT-4o-based methods with 5-60x faster inference.

Conclusion: SoLS effectively addresses RL challenges for foundation models in UI navigation tasks, demonstrating superior performance and efficiency through differentiated update strategies for positive and negative samples.

Abstract: Reinforcement learning (RL) using foundation models for policy approximations in multi-turn tasks remains challenging. We identify two main limitations related to sparse reward settings and policy gradient updates, based on which we formulate a key insight: updates from positive samples with high returns typically do not require policy regularisation, whereas updates from negative samples, reflecting undesirable behaviour, can harm model performance. This paper introduces Succeed or Learn Slowly (SoLS), a novel off-policy RL algorithm evaluated on mobile app control tasks. SoLS improves sample efficiency when fine-tuning foundation models for user interface navigation via a modified off-policy actor-critic approach, applying direct policy updates for positive samples and conservative, regularised updates for negative ones to prevent model degradation. We augment SoLS with Successful Transition Replay (STR), which prioritises learning from successful interactions, further improving sample efficiency. We evaluate SoLS on the AndroidWorld benchmark, where it significantly outperforms existing methods (at least 17% relative increase), including prompt-engineering and RL approaches, while requiring substantially fewer computational resources than GPT-4o-based methods with 5-60x faster inference.

[884] Convolutional Monge Mapping between EEG Datasets to Support Independent Component Labeling

Austin Meek, Carlos H. Mendoza-Cardenas, Austin J. Brockmeier

Main category: cs.LG

TL;DR: Novel extension of CMMN method with two approaches for EEG spectral normalization, enabling space-time separable filters that improve IC classification for brain vs non-brain component recognition.

DetailsMotivation: EEG recordings contain artifacts, noise, and equipment-related spectral differences that impact machine learning performance for automated artifact removal and IC labeling in clinical applications like epilepsy and psychosis monitoring.

Method: Extended CMMN with two source reference spectrum computation approaches: (1) channel-averaged l1-normalized barycenter, and (2) subject-to-subject mapping finding closest spectrum source. Creates space-time separable filters for cross-dataset compatibility.

Result: Significant improvement in recognizing brain versus non-brain independent components in IC classification tasks.

Conclusion: The proposed spectral normalization extension enables better artifact removal through improved IC classification, benefiting clinical EEG analysis by minimizing equipment and context-related spectral differences.

Abstract: EEG recordings contain rich information about neural activity but are subject to artifacts, noise, and superficial differences due to sensors, amplifiers, and filtering. Independent component analysis and automatic labeling of independent components (ICs) enable artifact removal in EEG pipelines. Convolutional Monge Mapping Normalization (CMMN) is a recent tool used to achieve spectral conformity of EEG signals, which was shown to improve deep neural network approaches for sleep staging. Here we propose a novel extension of the CMMN method with two alternative approaches to computing the source reference spectrum the target signals are mapped to: (1) channel-averaged and $l_1$-normalized barycenter, and (2) a subject-to-subject mapping that finds the source subject with the closest spectrum to the target subject. Notably, our extension yields space-time separable filters that can be used to map between datasets with different numbers of EEG channels. We apply these filters in an IC classification task, and show significant improvement in recognizing brain versus non-brain ICs. Clinical relevance - EEG recordings are used in the diagnosis and monitoring of multiple neuropathologies, including epilepsy and psychosis. While EEG analysis can benefit from automating artifact removal through independent component analysis and labeling, differences in recording equipment and context (the presence of noise from electrical wiring and other devices) may impact the performance of machine learning models, but these differences can be minimized by appropriate spectral normalization through filtering.

[885] BM-CL: Bias Mitigation through the lens of Continual Learning

Lucas Mansilla, Rodrigo Echeveste, Camila Gonzalez, Diego H. Milone, Enzo Ferrante

Main category: cs.LG

TL;DR: BM-CL framework uses continual learning principles to mitigate ML biases without the leveling-down effect, improving outcomes for disadvantaged groups while preserving advantaged group performance.

DetailsMotivation: Traditional bias mitigation techniques often cause leveling-down effects where improving outcomes for disadvantaged groups reduces performance for advantaged groups, creating an undesirable trade-off.

Method: Reinterprets bias mitigation as a domain-incremental continual learning problem, drawing inspiration from Learning without Forgetting and Elastic Weight Consolidation techniques to balance fairness objectives incrementally.

Result: Experiments on synthetic and real-world image datasets with diverse bias sources show effective bias mitigation while minimizing loss of original knowledge.

Conclusion: The approach successfully bridges fairness and continual learning fields, providing a pathway for developing equitable and effective ML systems without performance trade-offs.

Abstract: Biases in machine learning pose significant challenges, particularly when models amplify disparities that affect disadvantaged groups. Traditional bias mitigation techniques often lead to a {\itshape leveling-down effect}, whereby improving outcomes of disadvantaged groups comes at the expense of reduced performance for advantaged groups. This study introduces Bias Mitigation through Continual Learning (BM-CL), a novel framework that leverages the principles of continual learning to address this trade-off. We postulate that mitigating bias is conceptually similar to domain-incremental continual learning, where the model must adjust to changing fairness conditions, improving outcomes for disadvantaged groups without forgetting the knowledge that benefits advantaged groups. Drawing inspiration from techniques such as Learning without Forgetting and Elastic Weight Consolidation, we reinterpret bias mitigation as a continual learning problem. This perspective allows models to incrementally balance fairness objectives, enhancing outcomes for disadvantaged groups while preserving performance for advantaged groups. Experiments on synthetic and real-world image datasets, characterized by diverse sources of bias, demonstrate that the proposed framework mitigates biases while minimizing the loss of original knowledge. Our approach bridges the fields of fairness and continual learning, offering a promising pathway for developing machine learning systems that are both equitable and effective.

[886] Communication-Aware Knowledge Distillation for Federated LLM Fine-Tuning over Wireless Networks

Xinlu Zhang, Na Yan, Yang Su, Yansha Deng, Toktam Mahmoodi

Main category: cs.LG

TL;DR: Proposes an efficient federated learning framework for LLMs using adaptive Top-k logit selection and LoRA-enhanced distillation to reduce communication overhead by ~50% while maintaining performance.

DetailsMotivation: Federated learning for LLMs faces high communication overhead and struggles with heterogeneous architectures. Traditional parameter-sharing methods are bandwidth-intensive, and transmitting full logits from LLMs is challenging for bandwidth-limited clients.

Method: 1) Adaptive Top-k logit selection for dynamic sparsification based on real-time communication conditions 2) Adaptive logits aggregation scheme to handle dimensional inconsistency 3) Incorporation of LoRA-adapted hidden-layer projection into distillation loss for richer representation

Result: Experimental results show superior performance compared to baseline methods while effectively reducing communication overhead by approximately 50%.

Conclusion: The proposed framework successfully addresses communication efficiency challenges in federated LLM distillation through adaptive logit selection and aggregation, combined with LoRA-enhanced representation learning.

Abstract: Federated learning (FL) for large language models (LLMs) offers a privacy-preserving scheme, enabling clients to collaboratively fine-tune locally deployed LLMs or smaller language models (SLMs) without exchanging raw data. While parameter-sharing methods in traditional FL models solves number of technical challenges, they still incur high communication overhead and struggle with adapting to heterogeneous model architectures. Federated distillation, a framework for mutual knowledge transfer via shared logits, typically offers lower communication overhead than parameter-sharing methods. However, transmitting logits from LLMs remains challenging for bandwidth-limited clients due to their high dimensionality. In this work, we focus on a federated LLM distillation with efficient communication overhead. To achieve this, we first propose an adaptive Top-k logit selection mechanism, dynamically sparsifying logits according to real-time communication conditions. Then to tackle the dimensional inconsistency introduced by the adaptive sparsification, we design an adaptive logits aggregation scheme, effectively alleviating the artificial and uninformative inputs introduced by conventional zero-padding methods. Finally, to enhance the distillation effect, we incorporate LoRA-adapted hidden-layer projection from LLM into the distillation loss, reducing the communication overhead further while providing richer representation. Experimental results demonstrate that our scheme achieves superior performance compared to baseline methods while effectively reducing communication overhead by approximately 50%.

[887] Toward a Unified Benchmark and Taxonomy of Stochastic Environments

Aryan Amit Barsainyan, Jing Yu Lim, Dianbo Liu

Main category: cs.LG

TL;DR: Introduces STORI benchmark for evaluating RL methods under diverse stochastic conditions and proposes a taxonomy of stochasticity in RL environments.

DetailsMotivation: Current RL benchmarks lack robustness testing for real-world stochastic and partially observable conditions, limiting systematic evaluation of model-based RL approaches.

Method: Created STORI benchmark incorporating diverse stochastic effects and developed a taxonomy to categorize different types of uncertainty in RL environments.

Result: Provides a standardized framework for rigorous assessment of RL methods under varied forms of uncertainty, addressing gaps in current evaluation practices.

Conclusion: STORI benchmark and stochasticity taxonomy enable more systematic evaluation and comparison of RL approaches in realistic, stochastic environments.

Abstract: Reinforcement Learning (RL) agents have achieved strong results on benchmarks such as Atari100k, yet they remain limited in robustness to real-world conditions. Model-Based RL approaches that rely on learned World Models often struggle in environments with true stochasticity and partial observability, despite their theoretical grounding in POMDPs. Current benchmarks rarely capture these challenges, focusing instead on deterministic or overly simplified settings, and the lack of a clear taxonomy of stochasticity further hampers systematic evaluation. To address this gap, we introduce STORI (STOchastic-ataRI), a benchmark that incorporates diverse stochastic effects and enables rigorous assessment of RL methods under varied forms of uncertainty. In addition, we propose a taxonomy of stochasticity in RL environments, providing a unified framework for analyzing and comparing approaches.

[888] A Multi-target Bayesian Transformer Framework for Predicting Cardiovascular Disease Biomarkers during Pandemics

Trusting Inekwe, Emmanuel Agu, Winnie Mkandawire, Andres Colubri

Main category: cs.LG

TL;DR: MBT-CB: Multi-target Bayesian Transformer for predicting CVD biomarkers from EHR data during COVID-19, capturing interdependencies, temporal patterns, and uncertainty.

DetailsMotivation: COVID-19 disrupted CVD care, affecting key biomarkers. Need accurate multi-target prediction models that capture biomarker relationships, temporal patterns, and uncertainty from EHR data.

Method: Proposed MBT-CB - Multi-target Bayesian Transformer with BERT-based framework, Bayesian Variational Inference for uncertainty, embeddings for temporal relationships, and DeepMTR for biomarker inter-relationships.

Result: Outperformed baselines with MAE 0.00887, RMSE 0.0135, MSE 0.00027. Effectively captured uncertainty, biomarker relationships, and temporal dynamics using attention mechanisms.

Conclusion: MBT-CB shows superior performance for CVD biomarker prediction, supporting clinical decision-making during pandemics through improved uncertainty estimation and relationship modeling.

Abstract: The COVID-19 pandemic disrupted healthcare systems worldwide, disproportionately impacting individuals with chronic conditions such as cardiovascular disease (CVD). These disruptions – through delayed care and behavioral changes, affected key CVD biomarkers, including LDL cholesterol (LDL-C), HbA1c, BMI, and systolic blood pressure (SysBP). Accurate modeling of these changes is crucial for predicting disease progression and guiding preventive care. However, prior work has not addressed multi-target prediction of CVD biomarker from Electronic Health Records (EHRs) using machine learning (ML), while jointly capturing biomarker interdependencies, temporal patterns, and predictive uncertainty. In this paper, we propose MBT-CB, a Multi-target Bayesian Transformer (MBT) with pre-trained BERT-based transformer framework to jointly predict LDL-C, HbA1c, BMI and SysBP CVD biomarkers from EHR data. The model leverages Bayesian Variational Inference to estimate uncertainties, embeddings to capture temporal relationships and a DeepMTR model to capture biomarker inter-relationships. We evaluate MBT-CT on retrospective EHR data from 3,390 CVD patient records (304 unique patients) in Central Massachusetts during the Covid-19 pandemic. MBT-CB outperformed a comprehensive set of baselines including other BERT-based ML models, achieving an MAE of 0.00887, RMSE of 0.0135 and MSE of 0.00027, while effectively capturing data and model uncertainty, patient biomarker inter-relationships, and temporal dynamics via its attention and embedding mechanisms. MBT-CB’s superior performance highlights its potential to improve CVD biomarker prediction and support clinical decision-making during pandemics.

[889] When LLM Meets Time Series: Can LLMs Perform Multi-Step Time Series Reasoning and Inference

Wen Ye, Jinbo Liu, Defu Cao, Wei Yang, Yan Liu

Main category: cs.LG

TL;DR: The paper introduces TSAIA Benchmark, a comprehensive evaluation framework for assessing LLMs as time-series AI assistants across 33 real-world tasks requiring complex temporal reasoning.

DetailsMotivation: To address the underexplored capability of LLMs in performing complex reasoning over temporal data and establish a rigorous benchmark for evaluating their performance in time series analysis tasks.

Method: Created TSAIA Benchmark by surveying over 20 academic publications to identify 33 real-world task formulations, with a dynamic question generator supporting continuous expansion. Adopted task-specific success criteria and tailored inference-quality metrics for heterogeneous tasks.

Result: Evaluation of eight state-of-the-art LLMs revealed limitations in their ability to assemble complex time series analysis workflows, highlighting the need for specialized domain adaptation methodologies.

Conclusion: Current LLMs struggle with complex temporal reasoning tasks, demonstrating the necessity for specialized approaches and the value of the TSAIA Benchmark as an extensible evaluation framework for future research in time series AI assistance.

Abstract: The rapid advancement of Large Language Models (LLMs) has sparked growing interest in their application to time series analysis tasks. However, their ability to perform complex reasoning over temporal data in real-world application domains remains underexplored. To move toward this goal, a first step is to establish a rigorous benchmark dataset for evaluation. In this work, we introduce the TSAIA Benchmark, a first attempt to evaluate LLMs as time-series AI assistants. To ensure both scientific rigor and practical relevance, we surveyed over 20 academic publications and identified 33 real-world task formulations. The benchmark encompasses a broad spectrum of challenges, ranging from constraint-aware forecasting to anomaly detection with threshold calibration: tasks that require compositional reasoning and multi-step time series analysis. The question generator is designed to be dynamic and extensible, supporting continuous expansion as new datasets or task types are introduced. Given the heterogeneous nature of the tasks, we adopt task-specific success criteria and tailored inference-quality metrics to ensure meaningful evaluation for each task. We apply this benchmark to assess eight state-of-the-art LLMs under a unified evaluation protocol. Our analysis reveals limitations in current models’ ability to assemble complex time series analysis workflows, underscoring the need for specialized methodologies for domain-specific adaptation. Our benchmark is available at https://huggingface.co/datasets/Melady/TSAIA, and the code is available at https://github.com/USC-Melady/TSAIA.

[890] Goal-Conditioned Reinforcement Learning for Data-Driven Maritime Navigation

Vaishnav Vaidheeswaran, Dilith Jayakody, Samruddhi Mulay, Anand Lo, Md Mahbub Alam, Gabriel Spadon

Main category: cs.LG

TL;DR: RL-based vessel routing using AIS data and wind fields to optimize fuel efficiency, travel time, and route diversity across multiple origin-destination pairs.

DetailsMotivation: Existing vessel routing methods lack generalization across multiple routes and don't leverage large-scale traffic data from AIS systems.

Method: Reinforcement learning with Proximal Policy Optimization, recurrent networks, invalid-action masking, and exploration strategies. Uses AIS-derived traffic graphs with ERA5 wind fields in continuous observation space with multi-discrete actions.

Result: Action masking significantly improves policy performance, and combining penalty feedback with positive shaping rewards yields additional gains.

Conclusion: The proposed RL framework effectively handles complex maritime routing challenges and demonstrates practical improvements through action masking and reward shaping techniques.

Abstract: Routing vessels through narrow and dynamic waterways is challenging due to changing environmental conditions and operational constraints. Existing vessel-routing studies typically fail to generalize across multiple origin-destination pairs and do not exploit large-scale, data-driven traffic graphs. In this paper, we propose a reinforcement learning solution for big maritime data that can learn to find a route across multiple origin-destination pairs while adapting to different hexagonal grid resolutions. Agents learn to select direction and speed under continuous observations in a multi-discrete action space. A reward function balances fuel efficiency, travel time, wind resistance, and route diversity, using an Automatic Identification System (AIS)-derived traffic graph with ERA5 wind fields. The approach is demonstrated in the Gulf of St. Lawrence, one of the largest estuaries in the world. We evaluate configurations that combine Proximal Policy Optimization with recurrent networks, invalid-action masking, and exploration strategies. Our experiments demonstrate that action masking yields a clear improvement in policy performance and that supplementing penalty-only feedback with positive shaping rewards produces additional gains.

[891] Optimizing In-Context Learning for Efficient Full Conformal Prediction

Weicao Deng, Sangwoo Park, Min Li, Osvaldo Simeone

Main category: cs.LG

TL;DR: E-ICL+FCP is a new conformal prediction framework that combines in-context learning with a CP-aware loss to achieve efficient full conformal prediction without expensive retraining, outperforming existing methods in efficiency-coverage trade-offs.

DetailsMotivation: Existing conformal prediction methods face complementary limitations - Split CP suffers from data inefficiency due to dataset partitioning, while Full CP has prohibitive retraining complexity. Recent meta-learning/ICL approaches don't specifically optimize for CP, leading to large prediction sets.

Method: Enhanced ICL-based Full CP (E-ICL+FCP) uses a permutation-invariant Transformer-based ICL model trained with a CP-aware loss. It simulates multiple retrained models required by FCP without actual retraining, preserving coverage while reducing computational overhead.

Result: Experiments on synthetic and real tasks show E-ICL+FCP achieves superior efficiency-coverage trade-offs compared to existing SCP and FCP baselines, maintaining coverage while significantly improving computational efficiency.

Conclusion: E-ICL+FCP provides an efficient framework for full conformal prediction that overcomes the data inefficiency of split CP and computational complexity of traditional FCP, making reliable uncertainty quantification more practical for real-world applications.

Abstract: Reliable uncertainty quantification is critical for trustworthy AI. Conformal Prediction (CP) provides prediction sets with distribution-free coverage guarantees, but its two main variants face complementary limitations. Split CP (SCP) suffers from data inefficiency due to dataset partitioning, while full CP (FCP) improves data efficiency at the cost of prohibitive retraining complexity. Recent approaches based on meta-learning or in-context learning (ICL) partially mitigate these drawbacks. However, they rely on training procedures not specifically tailored to CP, which may yield large prediction sets. We introduce an efficient FCP framework, termed enhanced ICL-based FCP (E-ICL+FCP), which employs a permutation-invariant Transformer-based ICL model trained with a CP-aware loss. By simulating the multiple retrained models required by FCP without actual retraining, E-ICL+FCP preserves coverage while markedly reducing both inefficiency and computational overhead. Experiments on synthetic and real tasks demonstrate that E-ICL+FCP attains superior efficiency-coverage trade-offs compared to existing SCP and FCP baselines.

[892] GradES: Significantly Faster Training in Transformers with Gradient-Based Early Stopping

Qifu Wen, Xi Zeng, Zihan Zhou, Shuaijun Liu, Mehdi Hosseinzadeh, Reza Rawassizadeh

Main category: cs.LG

TL;DR: GradES is a gradient-based early stopping method that freezes individual transformer components when their gradients converge, eliminating validation passes and speeding up training 1.57-7.22x while improving accuracy by 1.2%.

DetailsMotivation: Traditional early stopping requires costly validation inference on large transformers. Different transformer components converge at varying rates during fine-tuning, suggesting inefficient simultaneous parameter updates.

Method: Tracks gradient magnitudes in attention projections and FFN matrices during backpropagation. Freezes individual matrices when gradients fall below threshold τ, allowing slow-converging components to continue learning.

Result: Achieves 1.57-7.22x training speedup by eliminating validation passes. Improves generalization with 1.2% higher average accuracy by preventing overfitting through component-wise early stopping.

Conclusion: GradES provides efficient component-level early stopping that accelerates transformer training while enhancing performance, making it superior to traditional global validation-based approaches.

Abstract: Early stopping monitors global validation loss and halts all parameter updates simultaneously, which is computationally costly for large transformers due to the extended time required for validation inference. We propose GradES, a novel gradient-based early stopping approach that operates within transformer components (attention projections and Feed-Forward layer matrices). We found that different components converge at varying rates during fine-tuning. GradES tracks the magnitude of gradients in backpropagation for these matrices during training. When a projection matrix’s gradients fall below a convergence threshold $\tau$, we exclude that projection matrix from further updates individually, eliminating costly validation passes while allowing slow converging matrices to continue learning. By strategically freezing parameters when their gradients converge, GradES speeds up training time by 1.57–7.22$\times$ while simultaneously enhancing generalization through early prevention of overfitting, resulting in 1.2% higher average accuracy.

[893] Preserving Bilinear Weight Spectra with a Signed and Shrunk Quadratic Activation Function

Jason Abohwo, Thomas Mosen

Main category: cs.LG

TL;DR: SQS activation function enables GLUs to learn interpretable features from weights without performance trade-offs, achieving competitive results with interpretability benefits.

DetailsMotivation: Need for reliable ML model understanding through weight-based interpretability rather than activation analysis, addressing drawbacks of existing methods like reduced performance and data inefficiency.

Method: Introduces Signed Quadratic Shrink (SQS) activation function specifically designed for Gated Linear Units (GLUs) to enable direct weight-based feature interpretation.

Result: SQS achieves performance competitive with state-of-the-art activation functions while providing weight-based interpretability capabilities.

Conclusion: SQS successfully bridges the gap between model performance and interpretability, allowing direct feature analysis from network weights without sacrificing computational efficiency.

Abstract: Understanding the inner workings of machine learning models is critical for ensuring their reliability and robustness. Whilst many techniques in mechanistic interpretability focus on activation driven analyses, being able to derive meaningful features directly from the weights of a neural network would provide greater guarantees and more computational efficiency. Existing techniques for analyzing model features through weights suffer from drawbacks such as reduced performance and data inefficiency. In this paper, we introduce Signed Quadratic Shrink (SQS), an activation function designed to allow Gated Linear Units (GLUs) to learn interpretable features without these drawbacks. Our experimental results show that SQS achieves performance competitive with state-of-the-art activation functions whilst enabling weight-based interpretability

[894] Semi-on-Demand Transit Feeders with Shared Autonomous Vehicles and Reinforcement-Learning-Based Zonal Dispatching Control

Max T. M. Ng, Roman Engelhardt, Florian Dandl, Hani S. Mahmassani, Klaus Bogenberger

Main category: cs.LG

TL;DR: Semi-on-demand transit feeder service using shared autonomous vehicles with RL-based zonal dispatching control, combining fixed-route efficiency with demand-responsive flexibility for better accessibility in low-density areas.

DetailsMotivation: To improve accessibility in lower-density areas by combining cost-effectiveness of fixed-route transit with adaptability of demand-responsive transport, addressing first-mile-last-mile problems in multimodal transit systems.

Method: Deep reinforcement learning model using Proximal Policy Optimization algorithm to dynamically assign vehicles to subdivided flexible-route zones based on real-time demand fluctuations. Agent-based simulations on a real-world bus route in Munich, Germany.

Result: The semi-on-demand service with dynamic zonal control serves 16% more passengers at 13% higher generalized costs compared to traditional fixed-route service. RL control specifically contributes 2.4% more passengers at 1.4% higher costs.

Conclusion: The study demonstrates the potential of integrating shared autonomous vehicle feeders and machine learning techniques into public transit, providing groundwork for innovations in addressing first-mile-last-mile problems in multimodal transit systems.

Abstract: This paper develops a semi-on-demand transit feeder service using shared autonomous vehicles (SAVs) and zonal dispatching control based on reinforcement learning (RL). This service combines the cost-effectiveness of fixed-route transit with the adaptability of demand-responsive transport to improve accessibility in lower-density areas. Departing from the terminus, SAVs first make scheduled fixed stops, then offer on-demand pick-ups and drop-offs in a pre-determined flexible-route area. Our deep RL model dynamically assigns vehicles to subdivided flexible-route zones in response to real-time demand fluctuations and operations, using a policy gradient algorithm - Proximal Policy Optimization. The methodology is demonstrated through agent-based simulations on a real-world bus route in Munich, Germany. Results show that after efficient training of the RL model, the semi-on-demand service with dynamic zonal control serves 16% more passengers at 13% higher generalized costs on average compared to traditional fixed-route service. The efficiency gain brought by RL control brings 2.4% more passengers at 1.4% higher costs. This study not only showcases the potential of integrating SAV feeders and machine learning techniques into public transit, but also sets the groundwork for further innovations in addressing first-mile-last-mile problems in multimodal transit systems.

[895] Deep Reinforcement Learning for Real-Time Drone Routing in Post-Disaster Road Assessment Without Domain Knowledge

Huatian Gong, Jiuh-Biing Sheu, Zheng Wang, Xiaoguang Yang, Ran Yan

Main category: cs.LG

TL;DR: Proposes an attention-based encoder-decoder model using deep reinforcement learning for real-time drone routing in post-disaster road damage assessment, achieving superior solution quality and faster computation than traditional methods.

DetailsMotivation: Traditional optimization methods for post-disaster road damage assessment are computationally slow and require domain expertise, making them unsuitable for time-sensitive emergency response scenarios where rapid decision-making is critical for saving lives.

Method: Develops an attention-based encoder-decoder model with deep reinforcement learning, network transformation from link-based to node-based routing, synthetic road network generation for training data, and policy optimization with multiple optima (POMO) with multi-task learning capabilities.

Result: Outperforms commercial solvers by 16-69% in solution quality, achieves real-time inference (1-2 seconds vs 100-2000 seconds), and demonstrates strong generalization across varying problem scales, drone numbers, and time constraints on both synthetic and real-world networks.

Conclusion: The proposed method effectively balances computational efficiency with solution quality, making it particularly suitable for time-critical disaster response applications where rapid drone routing decisions are essential for emergency operations.

Abstract: Rapid post-disaster road damage assessment is critical for effective emergency response, yet traditional optimization methods suffer from excessive computational time and require domain knowledge for algorithm design, making them unsuitable for time-sensitive disaster scenarios. This study proposes an attention-based encoder-decoder model (AEDM) for real-time drone routing decision in post-disaster road damage assessment. The method employs deep reinforcement learning to determine high-quality drone assessment routes without requiring algorithmic design knowledge. A network transformation method is developed to convert link-based routing problems into equivalent node-based formulations, while a synthetic road network generation technique addresses the scarcity of large-scale training datasets. The model is trained using policy optimization with multiple optima (POMO) with multi-task learning capabilities to handle diverse parameter combinations. Experimental results demonstrate two key strengths of AEDM: it outperforms commercial solvers by 16–69% in solution quality and achieves real-time inference (1–2 seconds) versus 100–2,000 seconds for traditional methods. The model exhibits strong generalization across varying problem scales, drone numbers, and time constraints, consistently outperforming baseline methods on unseen parameter distributions and real-world road networks. The proposed method effectively balances computational efficiency with solution quality, making it particularly suitable for time-critical disaster response applications where rapid decision-making is essential for saving lives.

[896] Predicting NCAP Safety Ratings: An Analysis of Vehicle Characteristics and ADAS Features Using Machine Learning

Raunak Kunwar, Aera Kim LeBoulluec

Main category: cs.LG

TL;DR: Machine learning analysis of NCAP safety ratings shows that while traditional vehicle characteristics (curb weight, model year) are most predictive of 5-star ratings, ADAS features also contribute meaningfully to achieving top safety scores.

DetailsMotivation: To understand how Advanced Driver-Assistance Systems (ADAS) interact with traditional vehicle attributes in predicting the highest NCAP safety ratings, as vehicle safety assessment evolves to include both passive and active safety technologies.

Method: Used a dataset of 5,128 vehicle variants (2011-2025) and compared four ML models (logistic regression, random forest, gradient boosting, SVC) with 5-fold cross-validation. Optimized best models (random forest and gradient boost) with RandomizedSearchCV and analyzed feature importance.

Result: Random Forest model achieved 89.18% accuracy and 0.9586 ROC AUC. Traditional vehicle characteristics (curb weight and model year) contributed over 55% of predictive capability, but ADAS features also provided meaningful predictive contributions.

Conclusion: Machine learning can effectively analyze NCAP data, revealing that both established vehicle parameters and modern ADAS features are important for achieving top safety ratings, with traditional characteristics remaining dominant but ADAS adding predictive value.

Abstract: Vehicle safety assessment is crucial for consumer information and regulatory oversight. The New Car Assessment Program (NCAP) assigns standardized safety ratings, which traditionally emphasize passive safety measures but now include active safety technologies such as Advanced Driver-Assistance Systems (ADAS). It is crucial to understand how these various systems interact empirically. This study explores whether particular ADAS features like Forward Collision Warning, Lane Departure Warning, Crash Imminent Braking, and Blind Spot Detection, together with established vehicle attributes (e.g., Curb Weight, Model Year, Vehicle Type, Drive Train), can reliably predict a vehicle’s likelihood of earning the highest (5-star) overall NCAP rating. Using a publicly available dataset derived from NCAP reports that contain approximately 5,128 vehicle variants spanning model years 2011-2025, we compared four different machine learning models: logistic regression, random forest, gradient boosting, and support vector classifier (SVC) using a 5-fold stratified cross-validation approach. The two best-performing algorithms (random forest and gradient boost) were hyperparameter optimized using RandomizedSearchCV. Analysis of feature importance showed that basic vehicle characteristics, specifically curb weight and model year, dominated predictive capability, contributing more than 55% of the feature relevance of the Random Forest model. However, the inclusion of ADAS features also provided meaningful predictive contributions. The optimized Random Forest model achieved robust results on a held-out test set, with an accuracy of 89.18% and a ROC AUC of 0.9586. This research reveals the use of machine learning to analyze large-scale NCAP data and highlights the combined predictive importance of both established vehicle parameters and modern ADAS features to achieve top safety ratings.

[897] VISP: Volatility Informed Stochastic Projection for Adaptive Regularization

Tanvir Islam

Main category: cs.LG

TL;DR: VISP is an adaptive regularization method that uses gradient volatility to dynamically inject stochastic noise in neural networks, improving generalization performance over fixed-noise approaches.

DetailsMotivation: Conventional regularization methods apply uniform noise or fixed dropout rates, which may not optimally address overfitting. VISP aims to leverage gradient volatility information to selectively regularize unstable representations while preserving stable ones.

Method: VISP dynamically computes volatility from gradient statistics and uses it to scale a stochastic projection matrix. This allows selective regularization of inputs and hidden nodes with higher gradient volatility.

Result: Extensive experiments on MNIST, CIFAR-10, and SVHN show VISP consistently improves generalization over baseline models and fixed-noise alternatives. Analyses reveal it stabilizes internal network dynamics and fosters more robust feature representations.

Conclusion: VISP provides an effective adaptive regularization approach that leverages gradient volatility to guide stochastic noise injection, leading to better generalization and more stable network dynamics compared to conventional methods.

Abstract: We propose VISP: Volatility Informed Stochastic Projection, an adaptive regularization method that leverages gradient volatility to guide stochastic noise injection in deep neural networks. Unlike conventional techniques that apply uniform noise or fixed dropout rates, VISP dynamically computes volatility from gradient statistics and uses it to scale a stochastic projection matrix. This mechanism selectively regularizes inputs and hidden nodes that exhibit higher gradient volatility while preserving stable representations, thereby mitigating overfitting. Extensive experiments on MNIST, CIFAR-10, and SVHN demonstrate that VISP consistently improves generalization performance over baseline models and fixed-noise alternatives. In addition, detailed analyses of the evolution of volatility, the spectral properties of the projection matrix, and activation distributions reveal that VISP not only stabilizes the internal dynamics of the network but also fosters a more robust feature representation.

[898] Causal representation learning from network data

Jifan Zhang, Michelle M. Li, Elena Zheleva

Main category: cs.LG

TL;DR: GraCE-VAE is a framework for causal disentanglement in non-i.i.d. settings using structured network data, combining VAE with graph neural networks to recover latent causal graphs and intervention effects.

DetailsMotivation: Existing causal disentanglement methods assume i.i.d. data, but many real-world scenarios involve structured context like network data that should be leveraged for better causal inference.

Method: GraCE-VAE integrates discrepancy-based variational autoencoders with graph neural networks to jointly recover the true latent causal graph and intervention effects in non-i.i.d. settings.

Result: Theoretical identifiability results from i.i.d. data hold in this setup, and empirical evaluation on three genetic perturbation datasets shows improved performance over state-of-the-art baselines.

Conclusion: Leveraging structured context through GraCE-VAE significantly improves causal disentanglement performance in non-i.i.d. settings compared to traditional i.i.d. approaches.

Abstract: Causal disentanglement from soft interventions is identifiable under the assumptions of linear interventional faithfulness and availability of both observational and interventional data. Previous research has looked into this problem from the perspective of i.i.d. data. Here, we develop a framework, GraCE-VAE, for non-i.i.d. settings, in which structured context in the form of network data is available. GraCE-VAE integrates discrepancy-based variational autoencoders with graph neural networks to jointly recover the true latent causal graph and intervention effects. We show that the theoretical results of identifiability from i.i.d. data hold in our setup. We also empirically evaluate GraCE-VAE against state-of-the-art baselines on three genetic perturbation datasets to demonstrate the impact of leveraging structured context for causal disentanglement.

[899] A Continuous Encoding-Based Representation for Efficient Multi-Fidelity Multi-Objective Neural Architecture Search

Zhao Wei, Chin Chun Ooi, Yew-Soon Ong

Main category: cs.LG

TL;DR: Proposes a multi-fidelity multi-objective NAS algorithm with adaptive Co-Kriging and clustering-based sampling to reduce computational cost, achieving state-of-the-art results on multiple benchmarks.

DetailsMotivation: Neural architecture search (NAS) is computationally expensive, especially for multiple conflicting objectives, requiring more efficient methods to reduce computational budget.

Method: Uses adaptive Co-Kriging-assisted multi-fidelity multi-objective NAS with clustering-based local infill sampling and novel continuous encoding for cell connections in U-Net backbones.

Result: Outperforms previous state-of-the-art methods on numerical benchmarks, 2D Darcy flow regression, and biomedical image segmentation; successfully applied to wind velocity regression with good prediction and lower complexity.

Conclusion: The NAS algorithm efficiently reduces computational costs while discovering optimal architectures, independently identifying established U-Net design principles like information flow between cells.

Abstract: Neural architecture search (NAS) is an attractive approach to automate the design of optimized architectures but is constrained by high computational budget, especially when optimizing for multiple, important conflicting objectives. To address this, an adaptive Co-Kriging-assisted multi-fidelity multi-objective NAS algorithm is proposed to further reduce the computational cost of NAS by incorporating a clustering-based local multi-fidelity infill sampling strategy, enabling efficient exploration of the search space for faster convergence. This algorithm is further accelerated by the use of a novel continuous encoding method to represent the connections of nodes in each cell within a generalized cell-based U-Net backbone, thereby decreasing the search dimension (number of variables). Results indicate that the proposed NAS algorithm outperforms previously published state-of-the-art methods under limited computational budget on three numerical benchmarks, a 2D Darcy flow regression problem and a CHASE_DB1 biomedical image segmentation problem. The proposed method is subsequently used to create a wind velocity regression model with application in urban modelling, with the found model able to achieve good prediction with less computational complexity. Further analysis revealed that the NAS algorithm independently identified principles undergirding superior U-Net architectures in other literature, such as the importance of allowing each cell to incorporate information from prior cells.

[900] Knowledge distillation as a pathway toward next-generation intelligent ecohydrological modeling systems

Long Jiang, Yang Yang, Ting Fong May Chui, Morgan Thornwell, Hoshin Vijai Gupta

Main category: cs.LG

TL;DR: A three-phase framework integrating process-based models with machine learning through knowledge distillation for intelligent ecohydrological modeling

DetailsMotivation: To overcome limitations of traditional process-based models (structural rigidity, high computational costs) and machine learning methods (lack of interpretability and transferability) in ecohydrological simulations

Method: Three-phase knowledge distillation: Phase I (behavioral distillation) uses surrogate learning for model simplification; Phase II (structural distillation) reformulates process equations as GNN components; Phase III (cognitive distillation) embeds expert reasoning using Eyes-Brain-Hands-Mouth architecture

Result: Successfully demonstrated in Samish watershed, reproducing process-based model outputs, improving predictive accuracy, and supporting scenario-based decision-making

Conclusion: Provides a scalable and transferable pathway for next-generation intelligent ecohydrological modeling systems with potential extension to other process-based domains

Abstract: Simulating ecohydrological processes is essential for understanding complex environmental systems and guiding sustainable management amid accelerating climate change and human pressures. Process-based models provide physical realism but can suffer from structural rigidity, high computational costs, and complex calibration, while machine learning (ML) methods are efficient and flexible yet often lack interpretability and transferability. We propose a unified three-phase framework that integrates process-based models with ML and progressively embeds them into artificial intelligence (AI) through knowledge distillation. Phase I, behavioral distillation, enhances process models via surrogate learning and model simplification to capture key dynamics at lower computational cost. Phase II, structural distillation, reformulates process equations as modular components within a graph neural network (GNN), enabling multiscale representation and seamless integration with ML models. Phase III, cognitive distillation, embeds expert reasoning and adaptive decision-making into intelligent modeling agents using the Eyes-Brain-Hands-Mouth architecture. Demonstrations for the Samish watershed highlight the framework’s applicability to ecohydrological modeling, showing that it can reproduce process-based model outputs, improve predictive accuracy, and support scenario-based decision-making. The framework offers a scalable and transferable pathway toward next-generation intelligent ecohydrological modeling systems, with the potential extension to other process-based domains.

[901] Semantic and episodic memories in a predictive coding model of the neocortex

Lucie Fontaine, Frédéric Alexandre

Main category: cs.LG

TL;DR: Predictive coding models can exhibit episodic memory capabilities but only when trained on small datasets, suggesting episodic memory arises from semantic learning in limited contexts, supporting the need for hippocampus-like sparse representations.

DetailsMotivation: To challenge the traditional Complementary Learning Systems theory by exploring whether predictive coding models of the neocortex can demonstrate hippocampus-like episodic memory capabilities, and to understand the relationship between semantic and episodic memory systems.

Method: Developed a predictive coding neural network model of the neocortex and tested its episodic memory capabilities by training it on varying numbers of examples to assess recall specificity and generalization performance.

Result: The predictive coding model could recall specific individual examples but only when trained on a small number of examples. The model overfitted to these examples and showed poor generalization. Models trained on many examples lost their episodic recall capabilities.

Conclusion: Episodic memory can arise from semantic learning in the neocortex but only for a limited number of examples, supporting the biological necessity of the hippocampus’s sparse, pattern-separated representations for true episodic memory capabilities.

Abstract: Complementary Learning Systems theory holds that intelligent agents need two learning systems. Semantic memory is encoded in the neocortex with dense, overlapping representations and acquires structured knowledge. Episodic memory is encoded in the hippocampus with sparse, pattern-separated representations and quickly learns the specifics of individual experiences. Recently, this duality between semantic and episodic memories has been challenged by predictive coding, a biologically plausible neural network model of the neocortex which was shown to have hippocampus-like abilities on auto-associative memory tasks. These results raise the question of the episodic capabilities of the neocortex and their relation to semantic memory. In this paper, we present such a predictive coding model of the neocortex and explore its episodic capabilities. We show that this kind of model can indeed recall the specifics of individual examples but only if it is trained on a small number of examples. The model is overfitted to these exemples and does not generalize well, suggesting that episodic memory can arise from semantic learning. Indeed, a model trained with many more examples loses its recall capabilities. This work suggests that individual examples can be encoded gradually in the neocortex using dense, overlapping representations but only in a limited number, motivating the need for sparse, pattern-separated representations as found in the hippocampus.

[902] ACA-Net: Future Graph Learning for Logistical Demand-Supply Forecasting

Jiacheng Shi, Haibin Wei, Jiang Wang, Xiaowei Xu, Longzhi Du, Taixu Jiang

Main category: cs.LG

TL;DR: Proposes a novel spatiotemporal learning model using only two graphs (ongoing and global) for logistical demand-supply forecasting in food delivery platforms, outperforming traditional long-series methods.

DetailsMotivation: Current spatial-temporal methods struggle to effectively capture future order distribution information due to strong randomness and time-series-insensitive nature in online delivery platforms, while maintaining efficiency.

Method: Innovative graph learning network framework (ACA-Net) using adaptive future graph learning and cross attention mechanism with ongoing and global graphs to extract future order distribution information.

Result: Achieves superior performance compared to traditional spatial-temporal long-series methods, significantly enhancing forecasting performance and learning robust future graphs.

Conclusion: The proposed method effectively improves logistical demand-supply pressure forecasting outcomes and has been validated in real-world production environments.

Abstract: Logistical demand-supply forecasting that evaluates the alignment between projected supply and anticipated demand, is essential for the efficiency and quality of on-demand food delivery platforms and serves as a key indicator for scheduling decisions. Future order distribution information, which reflects the distribution of orders in on-demand food delivery, is crucial for the performance of logistical demand-supply forecasting. Current studies utilize spatial-temporal analysis methods to model future order distribution information from serious time slices. However, learning future order distribution in online delivery platform is a time-series-insensitive problem with strong randomness. These approaches often struggle to effectively capture this information while remaining efficient. This paper proposes an innovative spatiotemporal learning model that utilizes only two graphs (ongoing and global) to learn future order distribution information, achieving superior performance compared to traditional spatial-temporal long-series methods. The main contributions are as follows: (1) The introduction of ongoing and global graphs in logistical demand-supply pressure forecasting compared to traditional long time series significantly enhances forecasting performance. (2) An innovative graph learning network framework using adaptive future graph learning and innovative cross attention mechanism (ACA-Net) is proposed to extract future order distribution information, effectively learning a robust future graph that substantially improves logistical demand-supply pressure forecasting outcomes. (3) The effectiveness of the proposed method is validated in real-world production environments.

[903] Bouncy particle sampler with infinite exchanging parallel tempering

Yohei Saito, Shun Kimura, Koujin Takeda

Main category: cs.LG

TL;DR: Proposes combining parallel tempering with bouncy particle sampler (BPS) to accelerate convergence for multimodal posterior distributions, with infinite temperature exchange rate.

DetailsMotivation: Bayesian inference requires approximating posterior distributions, but existing sampling methods like HMC and MCMC have limitations. BPS offers easier parameter tuning than HMC but needs acceleration for multimodal distributions.

Method: Introduces parallel tempering (PT) to bouncy particle sampler (BPS) with infinite inverse temperature exchange rate to enhance convergence speed for multimodal distributions.

Result: Numerical simulations demonstrate the effectiveness of the proposed PT-BPS approach for handling multimodal distributions.

Conclusion: The combination of parallel tempering with bouncy particle sampler provides an effective method for accelerating convergence to posterior distributions, particularly beneficial for complex multimodal problems.

Abstract: Bayesian inference is useful to obtain a predictive distribution with a small generalization error. However, since posterior distributions are rarely evaluated analytically, we employ the variational Bayesian inference or sampling method to approximate posterior distributions. When we obtain samples from a posterior distribution, Hamiltonian Monte Carlo (HMC) has been widely used for the continuous variable part and Markov chain Monte Carlo (MCMC) for the discrete variable part. Another sampling method, the bouncy particle sampler (BPS), has been proposed, which combines uniform linear motion and stochastic reflection to perform sampling. BPS was reported to have the advantage of being easier to set simulation parameters than HMC. To accelerate the convergence to a posterior distribution, we introduced parallel tempering (PT) to BPS, and then proposed an algorithm when the inverse temperature exchange rate is set to infinity. We performed numerical simulations and demonstrated its effectiveness for multimodal distribution.

[904] Second-Order Tensorial Partial Differential Equations on Graphs

Aref Einizade, Fragkiskos D. Malliaros, Jhony H. Giraldo

Main category: cs.LG

TL;DR: Proposes second-order tensorial partial differential equations on graphs (So-TPDEGs) to overcome limitations of first-order methods in handling multi-domain data with complex structures.

DetailsMotivation: Existing methods are limited to discrete graph filtering or first-order derivatives, which dampen high-frequency signals and slow information propagation, making them ineffective for capturing complex, multi-scale, and heterophilic structures in multi-domain data.

Method: Introduces second-order TPDEGs and leverages separability of cosine kernels in Cartesian product graphs for efficient spectral decomposition while preserving high-frequency information.

Result: Provides rigorous theoretical analyses of stability under graph perturbations and over-smoothing behavior regarding spectral properties.

Conclusion: Establishes a robust theoretical foundation for advancing continuous graph learning across multiple practical domains with improved handling of complex structures and high-frequency information.

Abstract: Processing data that lies on multiple interacting (product) graphs is increasingly important in practical applications, yet existing methods are mostly restricted to discrete graph filtering. Tensorial partial differential equations on graphs (TPDEGs) offer a principled framework for modeling such multidomain data in a continuous setting. However, current continuous approaches are limited to first-order derivatives, which tend to dampen high-frequency signals and slow down information propagation. This makes these TPDEGs-based approaches less effective for capturing complex, multi-scale, and heterophilic structures. In this paper, we introduce second-order TPDEGs (So-TPDEGs) and propose the first theoretically grounded framework for second-order continuous product graph neural networks. Our approach leverages the separability of cosine kernels in Cartesian product graphs to implement efficient spectral decomposition, while naturally preserving high-frequency information. We provide rigorous theoretical analyses of stability under graph perturbations and over-smoothing behavior regarding spectral properties. Our theoretical results establish a robust foundation for advancing continuous graph learning across multiple practical domains.

[905] Genetic Programming with Model Driven Dimension Repair for Learning Interpretable Appointment Scheduling Rules

Huan Zhang, Yang Wang, Ya-Hui Jia, Yi Mei

Main category: cs.LG

TL;DR: A dimensionally aware genetic programming algorithm with dimension repair is developed to evolve high-quality, interpretable appointment rules that outperform manual designs and existing methods.

DetailsMotivation: Direct application of genetic programming to appointment scheduling rules leads to rules that are difficult to interpret and trust due to lack of dimensional consistency, which is crucial for aligning with users' domain knowledge.

Method: Developed a dimensionally aware GP algorithm with dimension repair procedure formulated as a mixed-integer linear programming model, allowing exploration of diverse rule structures while maintaining dimensional consistency.

Result: The method evolved high-quality appointment rules that significantly outperform manually designed rules and state-of-the-art dimensionally aware GP methods in both objective performance and dimensional consistency.

Conclusion: The dimension repair approach enables effective evolution of interpretable and high-performing appointment scheduling rules, providing insights for designing more effective healthcare scheduling systems.

Abstract: Appointment scheduling is a great challenge in healthcare operations management. Appointment rules (AR) provide medical practitioners with a simple yet effective tool to determine patient appointment times. Genetic programming (GP) can be used to evolve ARs. However, directly applying GP to design ARs may lead to rules that are difficult for end-users to interpret and trust. A key reason is that GP is unaware of the dimensional consistency, which ensures that the evolved rules align with users’ domain knowledge and intuitive understanding. In this paper, we develop a new dimensionally aware GP algorithm with dimension repair to evolve ARs with dimensional consistency and high performance. A key innovation of our method is the dimension repair procedure, which optimizes the dimensional consistency of an expression tree while minimizing structural changes and ensuring that its output dimension meets the problem’s requirements. We formulate the task as a mixed-integer linear programming model that can be efficiently solved using common mathematical programming methods. With the support of the dimension repair procedure, our method can explore a wider range of AR structures by temporarily breaking the dimensional consistency of individuals, and then restoring it without altering their overall structure, thereby identifying individuals with greater potential advantages. We evaluated the proposed method in a comprehensive set of simulated clinics. The experimental results demonstrate that our approach managed to evolve high-quality ARs that significantly outperform not only the manually designed ARs but also existing state-of-the-art dimensionally aware GP methods in terms of both objective values and dimensional consistency. In addition, we analyzed the semantics of the evolved ARs, providing insight into the design of more effective and interpretable ARs.

[906] Fantastic Pretraining Optimizers and Where to Find Them

Kaiyue Wen, David Hall, Tengyu Ma, Percy Liang

Main category: cs.LG

TL;DR: AdamW remains dominant despite claims of 1.4-2x speedups from alternative optimizers. Rigorous evaluation shows actual speedups are smaller (1.1x for 1.2B models) and decrease with model size, with matrix-based optimizers performing best but scaling poorly.

DetailsMotivation: To address methodological shortcomings in optimizer comparisons, including unequal hyperparameter tuning and misleading evaluation setups that have hindered fair comparisons and practical adoption of alternative optimizers.

Method: Systematic study of ten deep learning optimizers across four model scales (0.1B-1.2B parameters) and data-to-model ratios (1-8x Chinchilla optimum), with rigorous hyperparameter tuning and end-of-training evaluations.

Result: Matrix-based optimizers (Muon, Soap) perform best but speedup decreases with model size (1.4x for 0.1B to 1.1x for 1.2B models). Intermediate checkpoint comparisons are misleading due to learning rate decay effects. Optimal hyperparameters are optimizer-specific.

Conclusion: AdamW’s dominance is justified as alternative optimizers offer diminishing returns with scale. Fair comparisons require rigorous tuning and end-of-training evaluation across multiple scales and data ratios.

Abstract: AdamW has long been the dominant optimizer in language model pretraining, despite numerous claims that alternative optimizers offer 1.4 to 2x speedup. We posit that two methodological shortcomings have obscured fair comparisons and hindered practical adoption: (i) unequal hyperparameter tuning and (ii) limited or misleading evaluation setups. To address these two issues, we conduct a systematic study of ten deep learning optimizers across four model scales (0.1B-1.2B parameters) and data-to-model ratios (1-8x the Chinchilla optimum). We find that fair and informative comparisons require rigorous hyperparameter tuning and evaluations across a range of model scales and data-to-model ratios, performed at the end of training. First, optimal hyperparameters for one optimizer may be suboptimal for another, making blind hyperparameter transfer unfair. Second, the actual speedup of many proposed optimizers over well-tuned baselines is lower than claimed and decreases with model size to only 1.1x for 1.2B parameter models. Thirdly, comparing intermediate checkpoints before reaching the target training budgets can be misleading, as rankings between two optimizers can flip during training due to learning rate decay. Through our thorough investigation, we find that all the fastest optimizers such as Muon and Soap, use matrices as preconditioners – multiplying gradients with matrices rather than entry-wise scalars. However, the speedup of matrix-based optimizers is inversely proportional to model scale, decreasing from 1.4x over AdamW for 0.1B parameter models to merely 1.1x for 1.2B parameter models.

[907] Privacy-Utility Trade-off in Data Publication: A Bilevel Optimization Framework with Curvature-Guided Perturbation

Yi Yin, Guangquan Zhang, Hua Zuo, Jie Lu

Main category: cs.LG

TL;DR: Novel bilevel optimization framework for private dataset publication that balances data utility (upper-level) and privacy protection (lower-level) using geometric curvature analysis to mitigate membership inference attacks while maintaining data quality.

DetailsMotivation: Existing privacy-preserving techniques like data perturbation and synthetic data generation often degrade data accuracy, specificity, and diversity, creating a need for better balance between privacy preservation and data utility.

Method: Bilevel optimization framework with upper-level focusing on data utility (using discriminator to ensure high-quality sample generation) and lower-level focusing on privacy (using local extrinsic curvature on data manifold to measure individual vulnerability to MIA and perturb samples toward low-curvature regions).

Result: Extensive experiments show the method enhances resistance to membership inference attacks in downstream tasks while surpassing existing methods in sample quality and diversity.

Conclusion: The proposed framework achieves synergistic balance between privacy and utility through alternating optimization, providing effective protection against MIA while maintaining high data quality for downstream applications.

Abstract: Machine learning models require datasets for effective training, but directly sharing raw data poses significant privacy risk such as membership inference attacks (MIA). To mitigate the risk, privacy-preserving techniques such as data perturbation, generalization, and synthetic data generation are commonly utilized. However, these methods often degrade data accuracy, specificity, and diversity, limiting the performance of downstream tasks and thus reducing data utility. Therefore, striking an optimal balance between privacy preservation and data utility remains a critical challenge. To address this issue, we introduce a novel bilevel optimization framework for the publication of private datasets, where the upper-level task focuses on data utility and the lower-level task focuses on data privacy. In the upper-level task, a discriminator guides the generation process to ensure that perturbed latent variables are mapped to high-quality samples, maintaining fidelity for downstream tasks. In the lower-level task, our framework employs local extrinsic curvature on the data manifold as a quantitative measure of individual vulnerability to MIA, providing a geometric foundation for targeted privacy protection. By perturbing samples toward low-curvature regions, our method effectively suppresses distinctive feature combinations that are vulnerable to MIA. Through alternating optimization of both objectives, we achieve a synergistic balance between privacy and utility. Extensive experimental evaluations demonstrate that our method not only enhances resistance to MIA in downstream tasks but also surpasses existing methods in terms of sample quality and diversity.

[908] LUCIE-3D: A three-dimensional climate emulator for forced responses

Haiwen Guan, Troy Arcomano, Ashesh Chattopadhyay, Romit Maulik

Main category: cs.LG

TL;DR: LUCIE-3D is a lightweight 3D climate emulator that captures atmospheric vertical structure using SFNO backbone, trained on ERA5 data, and successfully reproduces climate patterns and dynamics while maintaining computational efficiency.

DetailsMotivation: To develop a computationally efficient 3D climate model that captures vertical atmospheric structure, responds to climate forcings like CO2, and maintains long-term stability for rapid climate experimentation.

Method: Built on LUCIE-2D framework using Spherical Fourier Neural Operator (SFNO) backbone, trained on 30 years of ERA5 reanalysis data across 8 vertical levels, incorporating CO2 forcing and optional sea surface temperature inputs.

Result: Successfully reproduces climatological means, variability, long-term climate change signals (surface warming, stratospheric cooling), captures key dynamical processes (Kelvin waves, MJO, annular modes), and shows credible extreme event statistics with efficient training (<5 hours on 4 GPUs).

Conclusion: LUCIE-3D provides a stable, physically consistent, and accessible tool for rapid climate experimentation, ablation studies, and coupled climate dynamics exploration, with potential applications in paleoclimate research and Earth system emulation.

Abstract: We introduce LUCIE-3D, a lightweight three-dimensional climate emulator designed to capture the vertical structure of the atmosphere, respond to climate change forcings, and maintain computational efficiency with long-term stability. Building on the original LUCIE-2D framework, LUCIE-3D employs a Spherical Fourier Neural Operator (SFNO) backbone and is trained on 30 years of ERA5 reanalysis data spanning eight vertical {\sigma}-levels. The model incorporates atmospheric CO2 as a forcing variable and optionally integrates prescribed sea surface temperature (SST) to simulate coupled ocean–atmosphere dynamics. Results demonstrate that LUCIE-3D successfully reproduces climatological means, variability, and long-term climate change signals, including surface warming and stratospheric cooling under increasing CO2 concentrations. The model further captures key dynamical processes such as equatorial Kelvin waves, the Madden–Julian Oscillation, and annular modes, while showing credible behavior in the statistics of extreme events. Despite requiring longer training than its 2D predecessor, LUCIE-3D remains efficient, training in under five hours on four GPUs. Its combination of stability, physical consistency, and accessibility makes it a valuable tool for rapid experimentation, ablation studies, and the exploration of coupled climate dynamics, with potential applications extending to paleoclimate research and future Earth system emulation.

[909] Data-Dependent Smoothing for Protein Discovery with Walk-Jump Sampling

Srinivas Anumasa, Barath Chandran. C, Tingting Chen, Dianbo Liu

Main category: cs.LG

TL;DR: A Data-Dependent Smoothing Walk-Jump framework that uses kernel density estimation to determine noise scales per data point, improving diffusion model performance on sparse, heterogeneous protein data.

DetailsMotivation: Existing diffusion models ignore the uneven sparsity and data-dependent variability in complex domains like protein data, where samples are unevenly distributed with dense clusters and sparse regions.

Method: Uses kernel density estimation (KDE) as preprocessing to estimate noise scale σ for each data point, then trains a score model with these data-dependent σ values to incorporate local data geometry into denoising.

Result: Empirical evaluations show consistent improvements across multiple metrics, demonstrating better generative modeling performance in sparse, high-dimensional settings.

Conclusion: Data-aware sigma prediction is crucial for effective generative modeling in sparse, high-dimensional domains like protein data, as it accounts for heterogeneous data distributions.

Abstract: Diffusion models have emerged as a powerful class of generative models by learning to iteratively reverse the noising process. Their ability to generate high-quality samples has extended beyond high-dimensional image data to other complex domains such as proteins, where data distributions are typically sparse and unevenly spread. Importantly, the sparsity itself is uneven. Empirically, we observed that while a small fraction of samples lie in dense clusters, the majority occupy regions of varying sparsity across the data space. Existing approaches largely ignore this data-dependent variability. In this work, we introduce a Data-Dependent Smoothing Walk-Jump framework that employs kernel density estimation (KDE) as a preprocessing step to estimate the noise scale $\sigma$ for each data point, followed by training a score model with these data-dependent $\sigma$ values. By incorporating local data geometry into the denoising process, our method accounts for the heterogeneous distribution of protein data. Empirical evaluations demonstrate that our approach yields consistent improvements across multiple metrics, highlighting the importance of data-aware sigma prediction for generative modeling in sparse, high-dimensional settings.

[910] Abex-rat: Synergizing Abstractive Augmentation and Adversarial Training for Classification of Occupational Accident Reports

Jian Chen, Jinbao Tian, Yunqi Xu, Zhou Li

Main category: cs.LG

TL;DR: ABEX-RAT is a novel framework that combines generative data augmentation and adversarial training to address class imbalance in occupational accident report classification, achieving state-of-the-art performance with 90.32% macro-F1 score.

DetailsMotivation: Severe class imbalance in occupational accident datasets compromises analytical model performance, especially for rare but severe incident types, hindering reliable automated classification systems.

Method: Two-step approach: 1) ABEX pipeline uses LLM to distill core incident semantics and generative model to create diverse synthetic samples for underrepresented classes; 2) Lightweight classifier trained with random adversarial training (RAT) that stochastically applies perturbations for enhanced generalization.

Result: Achieved new SOTA performance on OSHA dataset with 90.32% macro-F1 score, significantly outperforming previous SOTA and fine-tuned large model baselines.

Conclusion: The synergistic strategy of generative data augmentation with robust adversarial training provides a highly effective and efficient alternative to brute-force fine-tuning for specialized, imbalanced classification tasks.

Abstract: The automatic classification of occupational accident reports is a critical research area for enhancing workplace safety and enabling large-scale risk analysis. However, the severe class imbalance inherent in these real-world datasets often compromises the performance of analytical models, particularly for rare but severe incident types, hindering the development of reliable automated systems. To address this challenge, we propose ABEX-RAT, a novel and efficient framework that synergizes generative data augmentation with robust adversarial training. Our approach first employs a twostep abstractive-expansive (ABEX) pipeline, which leverages a large language model to distill core incident semantics and then uses a generative model to create diverse, highquality synthetic samples for underrepresented classes. Subsequently, a lightweight classifier is trained on the augmented data using a computationally efficient random adversarial training (RAT) protocol, which stochastically applies perturbations to enhance model generalization and robustness without significant overhead. Experimental results on the public OSHA dataset demonstrate that our method achieves new state-of-the-art performance, reaching a macro-F1 score of 90.32% and significantly outperforming previous SOTA and fine-tuned large model baselines. Our work validates that this synergistic strategy is a highly effective and efficient alternative to brute-force fine-tuning for specialized, imbalanced classification tasks. The code is publicly available at:https://github.com/nxcc-lab/ABEX-RAT.

[911] HiGraph: A Large-Scale Hierarchical Graph Dataset for Malware Analysis

Han Chen, Hanchen Wang, Hongmei Chen, Ying Zhang, Lu Qin, Wenjie Zhang

Main category: cs.LG

TL;DR: HiGraph is the largest public hierarchical graph dataset for malware analysis, containing 200M CFGs nested within 595K FCGs, enabling better modeling of software structure semantics for robust malware detection.

DetailsMotivation: Existing graph-based malware analysis methods oversimplify programs into single-level graphs, failing to capture the hierarchical relationship between high-level functional interactions and low-level instruction logic, limiting detection robustness against code obfuscation and malware evolution.

Method: Created a two-level hierarchical graph representation with Control Flow Graphs (CFGs) nested within Function Call Graphs (FCGs) to preserve structural semantics essential for malware analysis.

Result: The dataset comprises over 200M CFGs nested within 595K FCGs, and large-scale analysis reveals distinct structural properties between benign and malicious software, establishing it as a foundational benchmark.

Conclusion: HiGraph bridges the gap in hierarchical graph datasets for malware analysis, providing a comprehensive resource that enables building more robust detectors resilient to code obfuscation and malware evolution, with public availability for community use.

Abstract: The advancement of graph-based malware analysis is critically limited by the absence of large-scale datasets that capture the inherent hierarchical structure of software. Existing methods often oversimplify programs into single level graphs, failing to model the crucial semantic relationship between high-level functional interactions and low-level instruction logic. To bridge this gap, we introduce \dataset, the largest public hierarchical graph dataset for malware analysis, comprising over \textbf{200M} Control Flow Graphs (CFGs) nested within \textbf{595K} Function Call Graphs (FCGs). This two-level representation preserves structural semantics essential for building robust detectors resilient to code obfuscation and malware evolution. We demonstrate HiGraph’s utility through a large-scale analysis that reveals distinct structural properties of benign and malicious software, establishing it as a foundational benchmark for the community. The dataset and tools are publicly available at https://higraph.org.

[912] Towards Comprehensive Information-theoretic Multi-view Learning

Long Shi, Yunshan Ye, Wenjie Wang, Tao Lei, Yu Zhao, Gang Kou, Badong Chen

Main category: cs.LG

TL;DR: CIML is a new multi-view learning framework that challenges the traditional multi-view redundancy assumption by leveraging both common and unique information from different views using information theory principles.

DetailsMotivation: Traditional multi-view methods assume that only common information between views is sufficient for downstream tasks, ignoring the potential predictive value of unique information in each view. This paper aims to develop a more comprehensive approach that utilizes both types of information.

Method: The CIML framework uses Gacs-Korner common information to extract shared features and Information Bottleneck to compress task-relevant representations. For unique information, it employs IB to compress unique representations while minimizing mutual information between unique and common representations and among different unique representations.

Result: The authors theoretically prove that the learned joint representation is predictively sufficient for downstream tasks. Extensive experiments show superior performance over state-of-the-art methods.

Conclusion: CIML successfully demonstrates that both common and unique information across views can be effectively leveraged for multi-view learning, outperforming traditional methods that rely solely on common information.

Abstract: Information theory has inspired numerous advancements in multi-view learning. Most multi-view methods incorporating information-theoretic principles rely an assumption called multi-view redundancy which states that common information between views is necessary and sufficient for down-stream tasks. This assumption emphasizes the importance of common information for prediction, but inherently ignores the potential of unique information in each view that could be predictive to the task. In this paper, we propose a comprehensive information-theoretic multi-view learning framework named CIML, which discards the assumption of multi-view redundancy. Specifically, CIML considers the potential predictive capabilities of both common and unique information based on information theory. First, the common representation learning maximizes Gacs-Korner common information to extract shared features and then compresses this information to learn task-relevant representations based on the Information Bottleneck (IB). For unique representation learning, IB is employed to achieve the most compressed unique representation for each view while simultaneously minimizing the mutual information between unique and common representations, as well as among different unique representations. Importantly, we theoretically prove that the learned joint representation is predictively sufficient for the downstream task. Extensive experimental results have demonstrated the superiority of our model over several state-of-art methods. The code is released on CIML.

[913] DivMerge: A divergence-based model merging method for multi-tasking

Touayouch Brahim, Fosse Loïc, Damnati Géraldine, Lecorvé Gwénolé

Main category: cs.LG

TL;DR: A robust model merging method that uses Jensen-Shannon divergence to combine multiple task-specific models into one without labeled data, maintaining performance across all tasks even as task count increases.

DetailsMotivation: Address task interference in multi-task learning when merging multiple fine-tuned models, which worsens with increasing number of tasks, without requiring additional labeled data.

Method: Leverages Jensen-Shannon divergence to guide the model merging process and automatically balance task importance during combination of task-specific models.

Result: The approach remains robust as the number of tasks grows and consistently outperforms prior work in model merging via task arithmetic.

Conclusion: Proposed method successfully merges multiple task-specific models into a single model while maintaining strong performance across all tasks without additional labeled data requirements.

Abstract: Multi-task learning (MTL) is often achieved by merging datasets before fine-tuning, but the growing availability of fine-tuned models has led to new approaches such as model merging via task arithmetic. A major challenge in this setting is task interference, which worsens as the number of tasks increases. We propose a method that merges models trained on different tasks into a single model, maintaining strong performance across all tasks. Our approach leverages Jensen-Shannon divergence to guide the merging process without requiring additional labelled data, and automatically balances task importance. Unlike existing methods, our approach remains robust as the number of tasks grows and consistently outperforms prior work.

[914] Differentiable Expectation-Maximisation and Applications to Gaussian Mixture Model Optimal Transport

Samuel Boïté, Eloi Tanguy, Julie Delon, Agnès Desolneux, Rémi Flamary

Main category: cs.LG

TL;DR: Making EM algorithm differentiable for integration into modern learning pipelines, with applications in Mixture Wasserstein distance computation for various ML tasks.

DetailsMotivation: EM algorithm is widely used but treated as non-differentiable black box, preventing its integration into end-to-end gradient-based learning systems.

Method: Present and compare several differentiation strategies for EM, from full automatic differentiation to approximate methods, and apply differentiable EM to compute Mixture Wasserstein distance between GMMs.

Result: Developed differentiable EM approach that enables MW₂ to be used as differentiable loss, with applications in barycentre computation, colour/style transfer, image generation, and texture synthesis.

Conclusion: The proposed differentiable EM approach is versatile and effective across different settings, with theoretical stability justification and practical applications in various machine learning tasks.

Abstract: The Expectation-Maximisation (EM) algorithm is a central tool in statistics and machine learning, widely used for latent-variable models such as Gaussian Mixture Models (GMMs). Despite its ubiquity, EM is typically treated as a non-differentiable black box, preventing its integration into modern learning pipelines where end-to-end gradient propagation is essential. In this work, we present and compare several differentiation strategies for EM, from full automatic differentiation to approximate methods, assessing their accuracy and computational efficiency. As a key application, we leverage this differentiable EM in the computation of the Mixture Wasserstein distance $\mathrm{MW}_2$ between GMMs, allowing $\mathrm{MW}_2$ to be used as a differentiable loss in imaging and machine learning tasks. To complement our practical use of $\mathrm{MW}_2$, we contribute a novel stability result which provides theoretical justification for the use of $\mathrm{MW}_2$ with EM, and also introduce a novel unbalanced variant of $\mathrm{MW}_2$. Numerical experiments on barycentre computation, colour and style transfer, image generation, and texture synthesis illustrate the versatility and effectiveness of the proposed approach in different settings.

[915] Conditional-$t^3$VAE: Equitable Latent Space Allocation for Fair Generation

Aymene Mohammed Bouayed, Samuel Deslauriers-Gauthier, Adrian Iaccovelli, David Naccache

Main category: cs.LG

TL;DR: Conditional-t³VAE improves generative fairness on imbalanced datasets by enforcing equitable latent space allocation across classes using per-class Student’s t-distribution priors, outperforming previous methods in severe class imbalance scenarios.

DetailsMotivation: Standard VAEs with global priors mirror training set class frequency in latent space, underrepresenting tail classes and reducing generative fairness on imbalanced datasets. Even t³VAE with heavy-tailed priors still allocates latent volume proportionally to class frequency.

Method: Proposes Conditional-t³VAE with per-class Student’s t joint priors over latent and output variables to prevent majority class dominance. Uses closed-form objective derived from γ-power divergence and equal-weight latent mixture of Student’s t-distributions for class-balanced generation.

Result: Achieves lower FID scores than t³VAE and Gaussian-based VAE baselines on SVHN-LT, CIFAR100-LT, and CelebA, particularly under severe class imbalance. Outperforms conditional Gaussian VAE in per-class F1 evaluations across all highly imbalanced settings.

Conclusion: While Gaussian-based models remain competitive under mild imbalance (ρ ≲ 3), Conditional-t³VAE substantially improves generative fairness and diversity in extreme imbalance regimes by ensuring equitable latent space allocation.

Abstract: Variational Autoencoders (VAEs) with global priors mirror the training set’s class frequency in latent space, underrepresenting tail classes and reducing generative fairness on imbalanced datasets. While $t^3$VAE improves robustness via heavy-tailed Student’s t-distribution priors, it still allocates latent volume proportionally to the class frequency.In this work, we address this issue by explicitly enforcing equitable latent space allocation across classes. To this end, we propose Conditional-$t^3$VAE, which defines a per-class \mbox{Student’s t} joint prior over latent and output variables, preventing dominance by majority classes. Our model is optimized using a closed-form objective derived from the $\gamma$-power divergence. Moreover, for class-balanced generation, we derive an equal-weight latent mixture of Student’s t-distributions. On SVHN-LT, CIFAR100-LT, and CelebA, Conditional-$t^3$VAE consistently achieves lower FID scores than both $t^3$VAE and Gaussian-based VAE baselines, particularly under severe class imbalance. In per-class F1 evaluations, Conditional-$t^3$VAE also outperforms the conditional Gaussian VAE across all highly imbalanced settings. While Gaussian-based models remain competitive under mild imbalance ratio ($\rho \lesssim 3$), our approach substantially improves generative fairness and diversity in more extreme regimes.

[916] Threshold-Based Optimal Arm Selection in Monotonic Bandits: Regret Lower Bounds and Algorithms

Chanakya Varude, Jay Chaudhary, Siddharth Kaushik, Prasanna Chaporkar

Main category: cs.LG

TL;DR: This paper introduces threshold-based multi-armed bandit problems where arms are selected based on their relation to a prescribed threshold τ, rather than simply choosing the highest-reward arm.

DetailsMotivation: Motivated by real-world applications in communication networks (CQI allocation), clinical dosing, energy management, and recommendation systems where decisions need to be made relative to specific thresholds rather than absolute maxima.

Method: The authors study various threshold-based variants (first above τ, k-th arm above/below, closest to τ) under monotonic arm mean structure. They derive asymptotic regret lower bounds and propose algorithms with optimality validated through Monte Carlo simulations.

Result: The research shows that regret lower bounds depend only on arms adjacent to the threshold τ, and the proposed algorithms demonstrate optimal performance in simulations.

Conclusion: This work extends classical bandit theory by incorporating threshold constraints, providing efficient decision-making frameworks for applications where relative positioning to thresholds is more important than absolute reward maximization.

Abstract: In multi-armed bandit problems, the typical goal is to identify the arm with the highest reward. This paper explores a threshold-based bandit problem, aiming to select an arm based on its relation to a prescribed threshold (\tau ). We study variants where the optimal arm is the first above (\tau), the (k^{th}) arm above or below it, or the closest to it, under a monotonic structure of arm means. We derive asymptotic regret lower bounds, showing dependence only on arms adjacent to (\tau). Motivated by applications in communication networks (CQI allocation), clinical dosing, energy management, recommendation systems, and more. We propose algorithms with optimality validated through Monte Carlo simulations. Our work extends classical bandit theory with threshold constraints for efficient decision-making.

[917] Scale, Don’t Fine-tune: Guiding Multimodal LLMs for Efficient Visual Place Recognition at Test-Time

Jintao Cheng, Weibin Li, Jiehao Luo, Xiaoyu Tang, Zhijian He, Jin Wu, Yao Zou, Wei Zhang

Main category: cs.LG

TL;DR: A zero-shot Visual Place Recognition framework using Test-Time Scaling with MLLMs that achieves 210x computational efficiency gains and superior cross-domain performance without fine-tuning.

DetailsMotivation: Current VPR approaches using VFMs and MLLMs suffer from high computational overhead and limited cross-domain transferability when fine-tuned, requiring a more efficient solution.

Method: Proposes a zero-shot framework with Test-Time Scaling (TTS) that leverages MLLMs’ vision-language alignment through Guidance-based methods for direct similarity scoring, using structured prompts for JSON outputs and Uncertainty-Aware Self-Consistency for real-time adaptation.

Result: Achieves significant improvements in cross-domain VPR performance with up to 210x computational efficiency gains compared to existing methods.

Conclusion: The proposed TTS framework provides an efficient, training-free solution for VPR that enables real-time adaptation and superior generalization across diverse environments without additional computational costs.

Abstract: Visual Place Recognition (VPR) has evolved from handcrafted descriptors to deep learning approaches, yet significant challenges remain. Current approaches, including Vision Foundation Models (VFMs) and Multimodal Large Language Models (MLLMs), enhance semantic understanding but suffer from high computational overhead and limited cross-domain transferability when fine-tuned. To address these limitations, we propose a novel zero-shot framework employing Test-Time Scaling (TTS) that leverages MLLMs’ vision-language alignment capabilities through Guidance-based methods for direct similarity scoring. Our approach eliminates two-stage processing by employing structured prompts that generate length-controllable JSON outputs. The TTS framework with Uncertainty-Aware Self-Consistency (UASC) enables real-time adaptation without additional training costs, achieving superior generalization across diverse environments. Experimental results demonstrate significant improvements in cross-domain VPR performance with up to 210$\times$ computational efficiency gains.

[918] Online Identification of IT Systems through Active Causal Learning

Kim Hammar, Rolf Stadler

Main category: cs.LG

TL;DR: First principled method for online, data-driven identification of IT system causal models using active causal learning with Gaussian process regression and rollout-based interventions.

DetailsMotivation: Traditional manual causal modeling by experts is challenging with modern complex IT systems. Need automated methods for predicting control effects, optimizing operations, diagnosing failures, and detecting intrusions.

Method: Active causal learning using Gaussian process regression based on system measurements, collected through rollout-based intervention policy. Iteratively estimates causal functions capturing dependencies among system variables.

Result: Method is proven optimal in Bayesian sense and produces effective interventions. Experimental validation shows accurate causal model identification with low interference to system operations.

Conclusion: This approach enables automated, data-driven causal modeling of IT systems, addressing the complexity and dynamism challenges of modern systems while maintaining operational efficiency.

Abstract: Identifying a causal model of an IT system is fundamental to many branches of systems engineering and operation. Such a model can be used to predict the effects of control actions, optimize operations, diagnose failures, detect intrusions, etc., which is central to achieving the longstanding goal of automating network and system management tasks. Traditionally, causal models have been designed and maintained by domain experts. This, however, proves increasingly challenging with the growing complexity and dynamism of modern IT systems. In this paper, we present the first principled method for online, data-driven identification of an IT system in the form of a causal model. The method, which we call active causal learning, estimates causal functions that capture the dependencies among system variables in an iterative fashion using Gaussian process regression based on system measurements, which are collected through a rollout-based intervention policy. We prove that this method is optimal in the Bayesian sense and that it produces effective interventions. Experimental validation on a testbed shows that our method enables accurate identification of a causal system model while inducing low interference with system operations.

[919] Baichuan-M2: Scaling Medical Capability with Large Verifier System

Baichuan-M2 Team, :, Chengfeng Dou, Chong Liu, Fan Yang, Fei Li, Jiyuan Jia, Mingyang Chen, Qiang Ju, Shuai Wang, Shunya Dang, Tianpeng Li, Xiangrong Zeng, Yijie Zhou, Chenzheng Zhu, Da Pan, Fei Deng, Guangwei Ai, Guosheng Dong, Hongda Zhang, Jinyang Tai, Jixiang Hong, Kai Lu, Linzhuang Sun, Peidong Guo, Qian Ma, Rihui Xin, Shihui Yang, Shusen Zhang, Yichuan Mo, Zheng Liang, Zhishou Zhang, Hengfu Cui, Zuyi Zhu, Xiaochuan Wang

Main category: cs.LG

TL;DR: A dynamic verification framework for medical LLMs that uses interactive reinforcement learning with patient simulation and clinical rubrics, resulting in Baichuan-M2 model that outperforms most models on HealthBench benchmarks.

DetailsMotivation: Address the gap between medical LLM performance on static benchmarks (like USMLE) and real-world clinical utility, as traditional exams fail to capture the dynamic, interactive nature of medical consultations.

Method: Developed a dynamic verification framework with Patient Simulator (using de-identified medical records) and Clinical Rubrics Generator. Created Baichuan-M2, a 32B-parameter model trained with multi-stage reinforcement learning using improved Group Relative Policy Optimization (GRPO) algorithm.

Result: Baichuan-M2 outperformed all open-source models and most advanced closed-source counterparts on HealthBench, achieving score above 32 on HealthBench Hard benchmark (previously only exceeded by GPT-5).

Conclusion: Robust dynamic verifier systems are essential for aligning LLM capabilities with practical clinical applications, establishing a new Pareto front in performance-parameter trade-off for medical AI deployment.

Abstract: As large language models (LLMs) advance in conversational and reasoning capabilities, their practical application in healthcare has become a critical research focus. However, there is a notable gap between the performance of medical LLMs on static benchmarks such as USMLE and their utility in real-world clinical decision-making. This discrepancy arises because traditional exams fail to capture the dynamic, interactive nature of medical consultations. To address this challenge, we introduce a novel dynamic verification framework that moves beyond static answer verifier, establishing a large-scale, high-fidelity interactive reinforcement learning system. Our framework comprises two key components: a Patient Simulator that creates realistic clinical environments using de-identified medical records, and a Clinical Rubrics Generator that dynamically produces multi-dimensional evaluation metrics. Building on this foundation, we develop Baichuan-M2, a 32B-parameter medical augmented reasoning model trained through a multi-stage reinforcement learning strategy with an improved Group Relative Policy Optimization (GRPO) algorithm. Evaluated on HealthBench, Baichuan-M2 outperforms all other open-source models and most advanced closed-source counterparts, achieving a score above 32 on the challenging HealthBench Hard benchmark-previously exceeded only by GPT-5. Our work demonstrates that robust dynamic verifier system is essential for aligning LLM capabilities with practical clinical applications, establishing a new Pareto front in the performance-parameter trade-off for medical AI deployment.

[920] Simulating classification models to evaluate Predict-Then-Optimize methods

Pieter Smet

Main category: cs.LG

TL;DR: This paper analyzes how prediction errors from machine learning models affect solution quality in Predict-Then-Optimize frameworks, introduces a new algorithm for simulating multiclass classifier predictions, and demonstrates that the relationship between prediction accuracy and optimization performance is non-trivial.

DetailsMotivation: To experimentally analyze how prediction error impacts solution quality in complex constrained optimization problems without needing to train real machine learning models, addressing a gap in literature where this relationship is often assumed but not properly validated.

Method: Developed a new algorithm for simulating predictions of multiclass classifiers (complementing existing binary classification simulation), conducted computational studies to evaluate algorithm performance, and applied these simulation algorithms to assess a Predict-Then-Optimize approach for machine scheduling problems.

Result: Classifier performance can be simulated with reasonable accuracy (though some variability exists), and experiments revealed that the relationship between prediction error and solution optimality is non-trivial - more accurate predictions do not always guarantee solutions closer to the actual optimum.

Conclusion: The findings highlight important considerations for designing and evaluating decision-making systems based on machine learning predictions, showing that the assumed direct relationship between prediction accuracy and optimization performance requires careful validation in complex optimization contexts.

Abstract: Uncertainty in optimization is often represented as stochastic parameters in the optimization model. In Predict-Then-Optimize approaches, predictions of a machine learning model are used as values for such parameters, effectively transforming the stochastic optimization problem into a deterministic one. This two-stage framework is built on the assumption that more accurate predictions result in solutions that are closer to the actual optimal solution. However, providing evidence for this assumption in the context of complex, constrained optimization problems is challenging and often overlooked in the literature. Simulating predictions of machine learning models offers a way to (experimentally) analyze how prediction error impacts solution quality without the need to train real models. Complementing an algorithm from the literature for simulating binary classification, we introduce a new algorithm for simulating predictions of multiclass classifiers. We conduct a computational study to evaluate the performance of these algorithms, and show that classifier performance can be simulated with reasonable accuracy, although some variability is observed. Additionally, we apply these algorithms to assess the performance of a Predict-Then-Optimize algorithm for a machine scheduling problem. The experiments demonstrate that the relationship between prediction error and how close solutions are to the actual optimum is non-trivial, highlighting important considerations for the design and evaluation of decision-making systems based on machine learning predictions.

[921] ST-Hyper: Learning High-Order Dependencies Across Multiple Spatial-Temporal Scales for Multivariate Time Series Forecasting

Binqing Wu, Jianlong Huang, Zongjiang Shang, Ling Chen

Main category: cs.LG

TL;DR: ST-Hyper is a novel deep learning method for multivariate time series forecasting that uses adaptive hypergraph modeling to capture high-order dependencies across multiple spatial-temporal scales, achieving state-of-the-art performance.

DetailsMotivation: Existing methods fail to model dependencies across multiple spatial-temporal scales (ST-scales) that jointly consider spatial and temporal scopes in multivariate time series forecasting.

Method: Proposes ST-Hyper with Spatial-Temporal Pyramid Modeling (STPM) module to extract multi-scale features, Adaptive Hypergraph Modeling (AHM) module to learn sparse hypergraphs for high-order dependencies, and tri-phase hypergraph propagation to capture spatial-temporal dynamics.

Result: Achieves state-of-the-art performance on six real-world MTS datasets, with average MAE reduction of 3.8% for long-term and 6.8% for short-term forecasting compared to best baselines.

Conclusion: ST-Hyper effectively models high-order dependencies across multiple spatial-temporal scales through adaptive hypergraph modeling, demonstrating superior forecasting performance for both short-term and long-term predictions.

Abstract: In multivariate time series (MTS) forecasting, many deep learning based methods have been proposed for modeling dependencies at multiple spatial (inter-variate) or temporal (intra-variate) scales. However, existing methods may fail to model dependencies across multiple spatial-temporal scales (ST-scales, i.e., scales that jointly consider spatial and temporal scopes). In this work, we propose ST-Hyper to model the high-order dependencies across multiple ST-scales through adaptive hypergraph modeling. Specifically, we introduce a Spatial-Temporal Pyramid Modeling (STPM) module to extract features at multiple ST-scales. Furthermore, we introduce an Adaptive Hypergraph Modeling (AHM) module that learns a sparse hypergraph to capture robust high-order dependencies among features. In addition, we interact with these features through tri-phase hypergraph propagation, which can comprehensively capture multi-scale spatial-temporal dynamics. Experimental results on six real-world MTS datasets demonstrate that ST-Hyper achieves the state-of-the-art performance, outperforming the best baselines with an average MAE reduction of 3.8% and 6.8% for long-term and short-term forecasting, respectively.

[922] DaCe AD: Unifying High-Performance Automatic Differentiation for Machine Learning and Scientific Computing

Afif Boudaoud, Alexandru Calotoiu, Marcin Copik, Torsten Hoefler

Main category: cs.LG

TL;DR: DaCe AD is a novel automatic differentiation engine that requires no code modifications and uses an ILP-based algorithm to optimize memory-performance tradeoffs, outperforming JAX by 92x on HPC benchmarks.

DetailsMotivation: Existing AD frameworks have limitations including limited language support, code modification requirements, poor performance on scientific computing codes, and inefficient memory management for forward-pass data, forcing scientists to manually compute gradients.

Method: DaCe AD uses a novel ILP-based algorithm to optimize the trade-off between storing and recomputing data to achieve maximum performance within given memory constraints, without requiring any code modifications.

Result: The system outperforms JAX (state-of-the-art Python AD framework) by more than 92 times on average when applied to NPBench, a suite of HPC benchmarks with diverse scientific computing patterns.

Conclusion: DaCe AD provides a general, efficient automatic differentiation solution that eliminates the need for code modifications while achieving significant performance improvements for scientific computing applications.

Abstract: Automatic differentiation (AD) is a set of techniques that systematically applies the chain rule to compute the gradients of functions without requiring human intervention. Although the fundamentals of this technology were established decades ago, it is experiencing a renaissance as it plays a key role in efficiently computing gradients for backpropagation in machine learning algorithms. AD is also crucial for many applications in scientific computing domains, particularly emerging techniques that integrate machine learning models within scientific simulations and schemes. Existing AD frameworks have four main limitations: limited support of programming languages, requiring code modifications for AD compatibility, limited performance on scientific computing codes, and a naive store-all solution for forward-pass data required for gradient calculations. These limitations force domain scientists to manually compute the gradients for large problems. This work presents DaCe AD, a general, efficient automatic differentiation engine that requires no code modifications. DaCe AD uses a novel ILP-based algorithm to optimize the trade-off between storing and recomputing to achieve maximum performance within a given memory constraint. We showcase the generality of our method by applying it to NPBench, a suite of HPC benchmarks with diverse scientific computing patterns, where we outperform JAX, a Python framework with state-of-the-art general AD capabilities, by more than 92 times on average without requiring any code changes.

[923] Calibration through the Lens of Indistinguishability

Parikshit Gopalan, Lunjia Hu

Main category: cs.LG

TL;DR: Survey paper on calibration in probabilistic predictions, exploring definitions, measurement approaches, and implications for decision-making through the lens of indistinguishability between predicted and real worlds.

DetailsMotivation: Address the fundamental question of how to interpret predicted probabilities and evaluate predictors when only discrete outcomes are observed, given the widespread use of probabilistic predictions in machine learning.

Method: Survey and analysis of recent work on calibration error definitions and measurement approaches, presenting a unifying viewpoint of calibration as indistinguishability between hypothesized and real worlds.

Result: Various calibration measures quantify the extent to which predicted probabilities can be distinguished from actual outcomes by different classes of statistical tests and distinguishers.

Conclusion: Calibration serves as a form of indistinguishability framework that helps evaluate how well probabilistic predictions align with real-world outcomes, providing guidance for downstream decision-making applications.

Abstract: Calibration is a classical notion from the forecasting literature which aims to address the question: how should predicted probabilities be interpreted? In a world where we only get to observe (discrete) outcomes, how should we evaluate a predictor that hypothesizes (continuous) probabilities over possible outcomes? The study of calibration has seen a surge of recent interest, given the ubiquity of probabilistic predictions in machine learning. This survey describes recent work on the foundational questions of how to define and measure calibration error, and what these measures mean for downstream decision makers who wish to use the predictions to make decisions. A unifying viewpoint that emerges is that of calibration as a form of indistinguishability, between the world hypothesized by the predictor and the real world (governed by nature or the Bayes optimal predictor). In this view, various calibration measures quantify the extent to which the two worlds can be told apart by certain classes of distinguishers or statistical measures.

[924] AdaSwitch: An Adaptive Switching Meta-Algorithm for Learning-Augmented Bounded-Influence Problems

Xi Chen, Yuze Chen, Yuan Zhou

Main category: cs.LG

TL;DR: AdaSwitch meta-algorithm for online decision-making with sequence-based predictions that provides near-optimal performance with accurate predictions while maintaining competitive-ratio guarantees with inaccurate predictions.

DetailsMotivation: To address multi-period online decision-making problems where predictions from ML models are available but their accuracy is not guaranteed, requiring algorithms that can leverage predictions when accurate while maintaining robustness when predictions are poor.

Method: Introduces a bounded-influence framework where past decisions have limited impact on future optimal rewards, and proposes the AdaSwitch meta-algorithm that adaptively switches between prediction-based and classical competitive algorithms.

Result: The framework and AdaSwitch algorithm achieve performance close to offline benchmarks when predictions are accurate, while preserving classical competitive-ratio guarantees under highly inaccurate predictions across diverse applications.

Conclusion: The bounded-influence framework and AdaSwitch meta-algorithm provide a flexible and broadly applicable approach to learning-augmented online decision-making that balances the benefits of predictions with robustness guarantees.

Abstract: We study a class of multi-period online decision-making problems with sequence-based predictions, which may be generated by machine learning models but whose accuracy is not guaranteed. In each period, the decision-maker observes the realized request and must take an irrevocable action that yields a reward or incurs a cost, without knowledge of future arrivals. We introduce a bounded-influence framework, in which past decisions and requests exert only limited impact on the future optimal reward. Within this framework, we propose the AdaSwitch meta-algorithm, which exploits predictions to attain performance close to the offline benchmark when predictions are accurate, while preserving classical competitive-ratio guarantees under highly inaccurate predictions. Our framework and meta-algorithm apply to diverse settings, including lead-time quotation in processing systems, the $k$-server problem, and online allocation of reusable resources. These applications illustrate the flexibility and broad applicability of our approach to learning-augmented online decision-making.

[925] Extrapolated Markov Chain Oversampling Method for Imbalanced Text Classification

Aleksi Avela, Pauliina Ilmonen

Main category: cs.LG

TL;DR: A novel Markov chain-based text oversampling method that addresses imbalanced text classification by expanding the minority class feature space using transition probabilities from both minority and majority classes.

DetailsMotivation: Text classification often faces imbalanced data where classes are unevenly distributed. Existing oversampling methods don't adequately handle the unique challenges of text data, particularly the expansion of feature space when sample size increases.

Method: Developed a Markov chain-based oversampling approach that estimates transition probabilities from both minority and majority classes, allowing the minority feature space to expand during synthetic data generation.

Result: The method produces highly competitive results against prominent oversampling methods, especially effective when dealing with severe class imbalance in real data examples.

Conclusion: The proposed Markov chain-based text oversampling method successfully addresses the distinctive difficulties of imbalanced text data and demonstrates superior performance in handling severe class imbalances.

Abstract: Text classification is the task of automatically assigning text documents correct labels from a predefined set of categories. In real-life (text) classification tasks, observations and misclassification costs are often unevenly distributed between the classes - known as the problem of imbalanced data. Synthetic oversampling is a popular approach to imbalanced classification. The idea is to generate synthetic observations in the minority class to balance the classes in the training set. Many general-purpose oversampling methods can be applied to text data; however, imbalanced text data poses a number of distinctive difficulties that stem from the unique nature of text compared to other domains. One such factor is that when the sample size of text increases, the sample vocabulary (i.e., feature space) is likely to grow as well. We introduce a novel Markov chain based text oversampling method. The transition probabilities are estimated from the minority class but also partly from the majority class, thus allowing the minority feature space to expand in oversampling. We evaluate our approach against prominent oversampling methods and show that our approach is able to produce highly competitive results against the other methods in several real data examples, especially when the imbalance is severe.

[926] RDIT: Residual-based Diffusion Implicit Models for Probabilistic Time Series Forecasting

Chih-Yu Lai, Yu-Chien Ning, Duane S. Boning

Main category: cs.LG

TL;DR: RDIT combines point estimation with residual-based conditional diffusion using bidirectional Mamba network to achieve state-of-the-art probabilistic time series forecasting with improved uncertainty quantification.

DetailsMotivation: Existing probabilistic time series forecasting methods have suboptimal distribution modeling and suffer from training-evaluation metric mismatch. Surprisingly, simple Gaussian augmentation of point estimators can achieve strong performance.

Method: Proposes RDIT framework that combines point estimation with residual-based conditional diffusion using bidirectional Mamba network. Theoretically derives optimal standard deviation for minimizing CRPS and develops distribution matching algorithms.

Result: Evaluations on eight multivariate datasets show RDIT achieves lower CRPS, faster inference, and improved coverage compared to strong baselines across varied forecasting horizons.

Conclusion: RDIT provides an effective plug-and-play framework for probabilistic time series forecasting that bridges the gap between point estimation and uncertainty quantification through residual-based diffusion modeling.

Abstract: Probabilistic Time Series Forecasting (PTSF) plays a critical role in domains requiring accurate and uncertainty-aware predictions for decision-making. However, existing methods offer suboptimal distribution modeling and suffer from a mismatch between training and evaluation metrics. Surprisingly, we found that augmenting a strong point estimator with a zero-mean Gaussian, whose standard deviation matches its training error, can yield state-of-the-art performance in PTSF. In this work, we propose RDIT, a plug-and-play framework that combines point estimation and residual-based conditional diffusion with a bidirectional Mamba network. We theoretically prove that the Continuous Ranked Probability Score (CRPS) can be minimized by adjusting to an optimal standard deviation and then derive algorithms to achieve distribution matching. Evaluations on eight multivariate datasets across varied forecasting horizons demonstrate that RDIT achieves lower CRPS, rapid inference, and improved coverage compared to strong baselines.

[927] Scaffolding Collaborative Learning in STEM: A Two-Year Evaluation of a Tool-Integrated Project-Based Methodology

Caterina Fuster-Barcelo, Gonzalo R. Rios-Munoz, Arrate Munoz-Barrutia

Main category: cs.LG

TL;DR: Integration of digital tools (Google Colab, Weights & Biases) and structured peer evaluation improved assessment fairness and student engagement in a Biomedical Image Processing course.

DetailsMotivation: To enhance student engagement, transparency, and fairness in evaluation for STEM education through digital collaborative tools and structured assessment methods.

Method: Redesigned Biomedical Image Processing course using real-time programming with Google Colab, experiment tracking via Weights & Biases, and rubric-guided peer assessment over two academic years, compared with pre-intervention cohort.

Result: Increased grade dispersion and higher entropy in final project scores indicating improved differentiation and fairness. Survey results showed greater student engagement with subject matter and learning process.

Conclusion: Digital tool-supported collaboration combined with structured evaluation mechanisms can enhance both learning outcomes and equity in STEM education.

Abstract: This study examines the integration of digital collaborative tools and structured peer evaluation in the Machine Learning for Health master’s program, through the redesign of a Biomedical Image Processing course over two academic years. The pedagogical framework combines real-time programming with Google Colab, experiment tracking and reporting via Weights & Biases, and rubric-guided peer assessment to foster student engagement, transparency, and fair evaluation. Compared to a pre-intervention cohort, the two implementation years showed increased grade dispersion and higher entropy in final project scores, suggesting improved differentiation and fairness in assessment. The survey results further indicate greater student engagement with the subject and their own learning process. These findings highlight the potential of integrating tool-supported collaboration and structured evaluation mechanisms to enhance both learning outcomes and equity in STEM education.

[928] Gaming and Cooperation in Federated Learning: What Can Happen and How to Monitor It

Dongseok Kim, Wonjun Jeong, Gisung Oh

Main category: cs.LG

TL;DR: This paper presents a strategic framework for analyzing Federated Learning as a system with rules and incentives, introducing indices to quantify behavioral incentives and performance loss, and providing practical guidelines to prevent metric gaming while maintaining cooperation.

DetailsMotivation: Federated Learning success depends on participant behaviors that occur out of sight, requiring analysis beyond mere optimization to understand strategic interactions, rules, and incentives that drive genuine performance improvement versus metric manipulation.

Method: Developed an analytical framework with two indices quantifying behavioral incentives and collective performance loss. Introduced thresholds, auto-switch rules, early warning signals, and a practical algorithm for audit resource allocation with performance guarantees. Conducted simulations across diverse environments for validation.

Result: Simulations consistently validated framework predictions across diverse environments. The approach provides design principles and operational guidelines that reduce metric gaming incentives while sustaining stable cooperation, with full reproducibility procedures released.

Conclusion: The framework enables clear identification of genuine performance improvement behaviors versus metric-targeting actions. Combining periodic recalibration, randomization, and connectivity-based alarms allows robust application in real-world operations, lowering gaming incentives while expanding cooperation.

Abstract: The success of Federated Learning depends on the actions that participants take out of sight. We model Federated Learning not as a mere optimization task but as a strategic system entangled with rules and incentives. From this perspective, we present an analytical framework that makes it possible to clearly identify where behaviors that genuinely improve performance diverge from those that merely target metrics. We introduce two indices that respectively quantify behavioral incentives and collective performance loss, and we use them as the basis for consistently interpreting the impact of operational choices such as rule design, the level of information disclosure, evaluation methods, and aggregator switching. We further summarize thresholds, auto-switch rules, and early warning signals into a checklist that can be applied directly in practice, and we provide both a practical algorithm for allocating limited audit resources and a performance guarantee. Simulations conducted across diverse environments consistently validate the patterns predicted by our framework, and we release all procedures for full reproducibility. While our approach operates most strongly under several assumptions, combining periodic recalibration, randomization, and connectivity-based alarms enables robust application under the variability of real-world operations. We present both design principles and operational guidelines that lower the incentives for metric gaming while sustaining and expanding stable cooperation.

[929] Fisher information flow in artificial neural networks

Maximilian Weimar, Lukas M. Rachbauer, Ilya Starshynov, Daniele Faccio, Linara Adilova, Dorian Bouchet, Stefan Rotter

Main category: cs.LG

TL;DR: A method to track Fisher information flow through neural networks during parameter estimation, showing optimal performance corresponds to maximal information transmission and providing a model-free training stopping criterion.

DetailsMotivation: As neural networks become integral to measurement systems, understanding how they process parameter-relevant information internally is crucial for optimal estimation performance.

Method: Presented a method to monitor Fisher information flow through ANN layers from input to output during parameter estimation tasks.

Result: Optimal estimation performance corresponds to maximal Fisher information transmission, and training beyond this point causes information loss due to overfitting.

Conclusion: The approach provides a model-free stopping criterion for network training without needing validation data, demonstrated effective in realistic imaging experiments.

Abstract: The estimation of continuous parameters from measured data plays a central role in many fields of physics. A key tool in understanding and improving such estimation processes is the concept of Fisher information, which quantifies how information about unknown parameters propagates through a physical system and determines the ultimate limits of precision. With Artificial Neural Networks (ANNs) gradually becoming an integral part of many measurement systems, it is essential to understand how they process and transmit parameter-relevant information internally. Here, we present a method to monitor the flow of Fisher information through an ANN performing a parameter estimation task, tracking it from the input to the output layer. We show that optimal estimation performance corresponds to the maximal transmission of Fisher information, and that training beyond this point results in information loss due to overfitting. This provides a model-free stopping criterion for network training-eliminating the need for a separate validation dataset. To demonstrate the practical relevance of our approach, we apply it to a network trained on data from an imaging experiment, highlighting its effectiveness in a realistic physical setting.

[930] Cache Management for Mixture-of-Experts LLMs – extended version

Spyros Angelopoulos, Loris Marchal, Adrien Obrecht, Bertrand Simon

Main category: cs.LG

TL;DR: New paging algorithm for efficient expert management in Mixture-of-Experts LLMs that outperforms standard LRU caching policies.

DetailsMotivation: Memory management challenge in large language models with billions of parameters, particularly for Mixture-of-Experts architectures where frequently used experts need efficient caching in fast memory rather than slower secondary storage.

Method: Proposed a layer-based extension of LRU algorithm specifically tailored for expert management optimization in LLMs, with theoretical analysis of competitive ratios for deterministic and randomized algorithms.

Result: Extensive simulations on synthetic datasets and actual MoE usage traces show the proposed algorithm outperforms classic paging policies like standard LRU.

Conclusion: The layer-based LRU extension provides superior expert caching performance for Mixture-of-Experts LLMs, addressing the critical memory management challenge with good theoretical guarantees and practical effectiveness.

Abstract: Large language models (LLMs) have demonstrated remarkable capabilities across a variety of tasks. One of the main challenges towards the successful deployment of LLMs is memory management, since they typically involve billions of parameters. To this end, architectures based on Mixture-of-Experts have been proposed, which aim to reduce the size of the parameters that are activated when producing a token. This raises the equally critical issue of efficiently managing the limited cache of the system, in that frequently used experts should be stored in the fast cache rather than in the slower secondary memory. In this work, we introduce and study a new paging problem that models expert management optimization. Our formulation captures both the layered architecture of LLMs and the requirement that experts are cached efficiently. We first present lower bounds on the competitive ratio of both deterministic and randomized algorithms, which show that under mild assumptions, LRU-like policies have good theoretical competitive performance. We then propose a layer-based extension of LRU that is tailored to the problem at hand. Extensive simulations on both synthetic datasets and actual traces of MoE usage show that our algorithm outperforms policies for the classic paging problem, such as the standard LRU.

[931] Learnable Loss Geometries with Mirror Descent for Scalable and Convergent Meta-Learning

Yilang Zhang, Bingcong Li, Georgios B. Giannakis

Main category: cs.LG

TL;DR: The paper proposes a nonlinear mirror map approach using neural networks to improve meta-learning adaptation by better capturing complex loss geometries, achieving faster convergence with fewer adaptation steps.

DetailsMotivation: Meta-learning faces challenges in efficiently adapting prior knowledge to new tasks with limited data. Simple linear preconditioning methods are insufficient for complex loss geometries, requiring more versatile distance metrics.

Method: The method learns a versatile distance-generating function using an expressive neural network to induce a nonlinear mirror map that captures diverse loss geometries, enabling more effective optimization.

Result: The approach achieves O(ε⁻²) convergence rate (matching standard methods) and demonstrates superior performance on few-shot learning datasets with significantly reduced adaptation steps.

Conclusion: The nonlinear mirror map approach effectively handles complex loss geometries in meta-learning, enabling rapid per-task convergence and making it suitable for large-scale meta-learning models.

Abstract: Utilizing task-invariant knowledge acquired from related tasks as prior information, meta-learning offers a principled approach to learning a new task with limited data records. Sample-efficient adaptation of this prior information is a major challenge facing meta-learning, and plays an important role because it facilitates training the sought task-specific model with just a few optimization steps. Past works deal with this challenge through preconditioning that speeds up convergence of the per-task training. Though effective in representing locally quadratic loss curvatures, simple linear preconditioning can be hardly potent with complex loss geometries. Instead of relying on a quadratic distance metric, the present contribution copes with complex loss metrics by learning a versatile distance-generating function, which induces a nonlinear mirror map to effectively capture and optimize a wide range of loss geometries. With suitable parameterization, this generating function is effected by an expressive neural network that is provably a valid distance. Analytical results establish convergence of not only the proposed method, but also all meta-learning approaches based on preconditioning. To attain gradient norm less than $\epsilon$, the convergence rate of $\mathcal{O}(\epsilon^{-2})$ is on par with standard gradient-based meta-learning methods. Numerical tests on few-shot learning datasets demonstrate the superior empirical performance of the novel algorithm, as well as its rapid per-task convergence, which markedly reduces the number of adaptation steps, hence also accommodating large-scale meta-learning models.

[932] VASSO: Variance Suppression for Sharpness-Aware Minimization

Bingcong Li, Yilang Zhang, Georgios B. Giannakis

Main category: cs.LG

TL;DR: VASSO improves SAM by stabilizing adversarial perturbations through variance suppression to prevent over-friendly adversaries and enhance generalization.

DetailsMotivation: SAM suffers from over-friendly adversaries that limit generalization performance, so there's a need to stabilize these adversaries for better flat minima discovery.

Method: Proposes VASSO (variance suppression) to provably stabilize adversaries in SAM, preventing friendliness and improving generalization across vision and language tasks.

Result: Improved generalization validated on extensive vision and language tasks, with better generalization-computation tradeoff when combined with efficient SAM variants.

Conclusion: VASSO successfully addresses SAM’s limitation with over-friendly adversaries through variance suppression, leading to enhanced generalization performance across multiple domains.

Abstract: Sharpness-aware minimization (SAM) has well-documented merits in enhancing generalization of deep neural network models. Accounting for sharpness in the loss function geometry, where neighborhoods of flat minima' heighten generalization ability, SAM seeks flat valleys’ by minimizing the maximum loss provoked by an adversarial perturbation within the neighborhood. Although critical to account for sharpness of the loss function, in practice SAM suffers from over-friendly adversaries,' which can curtail the outmost level of generalization. To avoid such friendliness,’ the present contribution fosters stabilization of adversaries through variance suppression (VASSO). VASSO offers a general approach to provably stabilize adversaries. In particular, when integrating VASSO with SAM, improved generalizability is numerically validated on extensive vision and language tasks. Once applied on top of a computationally efficient SAM variant, VASSO offers a desirable generalization-computation tradeoff.

[933] Generative Sequential Notification Optimization via Multi-Objective Decision Transformers

Borja Ocejo, Ruofan Wang, Ke Liu, Rohit K. Patra, Haotian Shen, David Liu, Yiwen Yuan, Gokulraj Mohanasundaram, Fedor Borisyuk, Prakruthi Prabhakar

Main category: cs.LG

TL;DR: Decision Transformer framework outperforms Conservative Q-Learning for notification optimization, achieving +0.72% session increase with better relevance and reduced user fatigue.

DetailsMotivation: Offline RL methods like CQL face practical challenges in notification systems including instability, sensitivity to distribution shifts, limited reproducibility, and explainability difficulties in high-dimensional settings.

Method: Decision Transformer-based framework that reframes policy learning as return-conditioned supervised learning, featuring multi-reward design, quantile regression for return-to-go conditioning, and circular buffer-based sequence processing for real-time inference.

Result: Extensive offline and online experiments show improved notification utility and overall session activity while minimizing user fatigue. DT-based approach achieved +0.72% increase in sessions compared to multi-objective CQL-based agent.

Conclusion: The Decision Transformer framework provides a more robust, scalable, and flexible solution for notification optimization, demonstrating practical advantages over traditional offline RL methods in production environments.

Abstract: Notifications are an important communication channel for delivering timely and relevant information. Optimizing their delivery involves addressing complex sequential decision-making challenges under constraints such as message utility and user fatigue. Offline reinforcement learning (RL) methods, such as Conservative Q-Learning (CQL), have been applied to this problem but face practical challenges at scale, including instability, sensitivity to distribution shifts, limited reproducibility, and difficulties with explainability in high-dimensional recommendation settings. We present a Decision Transformer (DT) based framework that reframes policy learning as return-conditioned supervised learning, improving robustness, scalability, and modeling flexibility. Our contributions include a real-world comparison with CQL, a multi-reward design suitable for non-episodic tasks, a quantile regression approach to return-to-go conditioning, and a production-ready system with circular buffer-based sequence processing for near-real-time inference. Extensive offline and online experiments in a deployed notification system show that our approach improves notification utility and overall session activity while minimizing user fatigue. Compared to a multi-objective CQL-based agent, the DT-based approach achieved a +0.72% increase in sessions for notification decision-making at LinkedIn by making notification recommendation more relevant.

[934] Exploring Variational Graph Autoencoders for Distribution Grid Data Generation

Syed Zain Abbas, Ehimare Okoyomon

Main category: cs.LG

TL;DR: VGAEs show promise for synthetic power grid generation but have limitations - simple decoders fail, GCN-based approaches work well on simpler grids but struggle with complex ones, producing disconnected components and repeated patterns.

DetailsMotivation: Address the lack of public power system data for machine learning research in energy networks by generating synthetic distribution grids.

Method: Used variational graph autoencoders (VGAEs) with four decoder variants, evaluated on two open-source datasets (ENGAGE and DINGO) using structural and spectral metrics.

Result: Simple decoders failed to capture realistic topologies. GCN-based approaches achieved strong fidelity on ENGAGE but struggled with DINGO, producing artifacts like disconnected components and repeated motifs.

Conclusion: VGAEs show both promise and limitations for grid synthesis, highlighting the need for more expressive generative models and robust evaluation. Models and analysis released as open source.

Abstract: To address the lack of public power system data for machine learning research in energy networks, we investigate the use of variational graph autoencoders (VGAEs) for synthetic distribution grid generation. Using two open-source datasets, ENGAGE and DINGO, we evaluate four decoder variants and compare generated networks against the original grids using structural and spectral metrics. Results indicate that simple decoders fail to capture realistic topologies, while GCN-based approaches achieve strong fidelity on ENGAGE but struggle on the more complex DINGO dataset, producing artifacts such as disconnected components and repeated motifs. These findings highlight both the promise and limitations of VGAEs for grid synthesis, underscoring the need for more expressive generative models and robust evaluation. We release our models and analysis as open source to support benchmarking and accelerate progress in ML-driven power system research.

[935] SimpleTIR: End-to-End Reinforcement Learning for Multi-Turn Tool-Integrated Reasoning

Zhenghai Xue, Longtao Zheng, Qian Liu, Yingru Li, Xiaosen Zheng, Zejun Ma, Bo An

Main category: cs.LG

TL;DR: SimpleTIR is a plug-and-play algorithm that stabilizes multi-turn Tool-Integrated Reasoning training by filtering out void turns that cause gradient norm explosions, achieving state-of-the-art performance on math reasoning benchmarks.

DetailsMotivation: Extending Tool-Integrated Reasoning to multi-turn scenarios using Reinforcement Learning often suffers from training instability and performance collapse due to distributional drift from external tool feedback.

Method: SimpleTIR identifies and filters out trajectories containing void turns (turns yielding neither code blocks nor final answers) to block harmful high-magnitude gradients and stabilize learning dynamics.

Result: Achieves SOTA performance on math reasoning benchmarks, elevating AIME24 score from 22.1 to 50.5 starting from Qwen2.5-7B base model, while enabling discovery of diverse reasoning patterns like self-correction.

Conclusion: SimpleTIR effectively addresses training instability in multi-turn TIR by removing problematic trajectories, enabling stable RL training and improved reasoning capabilities without supervised fine-tuning constraints.

Abstract: Large Language Models (LLMs) can significantly improve their reasoning capabilities by interacting with external tools, a paradigm known as Tool-Integrated Reasoning (TIR). However, extending TIR to multi-turn scenarios using Reinforcement Learning (RL) is often hindered by training instability and performance collapse. We identify that such instability is primarily caused by a distributional drift from external tool feedback, leading to the generation of low-probability tokens. This issue compounds over successive turns, causing catastrophic gradient norm explosions that derail the training process. To address this challenge, we introduce SimpleTIR , a plug-and-play algorithm that stabilizes multi-turn TIR training. Its core strategy is to identify and filter out trajectories containing void turns, i.e., turns that yield neither a code block nor a final answer. By removing these problematic trajectories from the policy update, SimpleTIR effectively blocks the harmful, high-magnitude gradients, thus stabilizing the learning dynamics. Extensive experiments show that SimpleTIR achieves state-of-the-art performance on challenging math reasoning benchmarks, notably elevating the AIME24 score from a text-only baseline of 22.1 to 50.5 when starting from the Qwen2.5-7B base model. Furthermore, by avoiding the constraints of supervised fine-tuning, SimpleTIR encourages the model to discover diverse and sophisticated reasoning patterns, such as self-correction and cross-validation.

[936] HydroGAT: Distributed Heterogeneous Graph Attention Transformer for Spatiotemporal Flood Prediction

Aishwarya Sarkar, Autrin Hakimi, Xiaoqiong Chen, Hai Huang, Chaoqun Lu, Ibrahim Demir, Ali Jannesari

Main category: cs.LG

TL;DR: HydroGAT is a spatiotemporal graph neural network that uses high-resolution basin graphs with land and river pixels as nodes to capture hydrological flow patterns, achieving superior flood forecasting accuracy with interpretable attention mechanisms and scalable distributed training.

DetailsMotivation: Traditional flood forecasting methods ignore river network topology or collapse spatial resolution, while existing GNN approaches fail to capture spatiotemporal interactions simultaneously. There's a need for high-resolution modeling that preserves hydrological structure while efficiently handling large-scale data.

Method: Creates a heterogeneous basin graph where every land and river pixel is a node connected by hydrological flow directions. Uses HydroGAT spatiotemporal network that learns local temporal importance and influential upstream locations through adaptive attention mechanisms.

Result: Achieves NSE up to 0.97, KGE up to 0.96, and PBIAS within ±5% in hourly discharge prediction across two Midwestern US basins. Shows 15x speedup with distributed training on 64 NVIDIA A100 GPUs. Provides interpretable attention maps revealing sparse intercatchment influences.

Conclusion: HydroGAT successfully integrates high-resolution spatial topology with temporal dynamics for accurate flood forecasting, offering both performance improvements and interpretability while demonstrating scalable distributed training capabilities for large basin-scale modeling.

Abstract: Accurate flood forecasting remains a challenge for water-resource management, as it demands modeling of local, time-varying runoff drivers (e.g., rainfall-induced peaks, baseflow trends) and complex spatial interactions across a river network. Traditional data-driven approaches, such as convolutional networks and sequence-based models, ignore topological information about the region. Graph Neural Networks (GNNs) propagate information exactly along the river network, which is ideal for learning hydrological routing. However, state-of-the-art GNN-based flood prediction models collapse pixels to coarse catchment polygons as the cost of training explodes with graph size and higher resolution. Furthermore, most existing methods treat spatial and temporal dependencies separately, either applying GNNs solely on spatial graphs or transformers purely on temporal sequences, thus failing to simultaneously capture spatiotemporal interactions critical for accurate flood prediction. We introduce a heterogenous basin graph where every land and river pixel is a node connected by physical hydrological flow directions and inter-catchment relationships. We propose HydroGAT, a spatiotemporal network that adaptively learns local temporal importance and the most influential upstream locations. Evaluated in two Midwestern US basins and across five baseline architectures, our model achieves higher NSE (up to 0.97), improved KGE (up to 0.96), and low bias (PBIAS within $\pm$5%) in hourly discharge prediction, while offering interpretable attention maps that reveal sparse, structured intercatchment influences. To support high-resolution basin-scale training, we develop a distributed data-parallel pipeline that scales efficiently up to 64 NVIDIA A100 GPUs on NERSC Perlmutter supercomputer, demonstrating up to 15x speedup across machines. Our code is available at https://github.com/swapp-lab/HydroGAT.

[937] RNN Generalization to Omega-Regular Languages

Charles Pert, Dalal Alrajeh, Alessandra Russo

Main category: cs.LG

TL;DR: RNNs can effectively generalize to ω-regular languages from LTL formulas, achieving high accuracy on sequences 8x longer than training data, with 92.6% of tasks showing perfect or near-perfect generalization.

DetailsMotivation: Büchi automata face scalability issues with complex system behaviors, and neural networks are increasingly used to address these challenges, but their generalization capabilities need investigation.

Method: Train RNNs on ultimately periodic ω-word sequences to replicate target BA behavior, then evaluate generalization on out-of-distribution sequences using LTL formulas corresponding to deterministic automata with 3 to 100+ states.

Result: RNNs achieve high accuracy on target ω-regular languages when tested on sequences up to 8 times longer than training examples, with 92.6% of tasks achieving perfect or near-perfect generalization.

Conclusion: Neural approaches are feasible for learning complex ω-regular languages and show potential as components in neurosymbolic verification methods.

Abstract: B"uchi automata (BAs) recognize $\omega$-regular languages defined by formal specifications like linear temporal logic (LTL) and are commonly used in the verification of reactive systems. However, BAs face scalability challenges when handling and manipulating complex system behaviors. As neural networks are increasingly used to address these scalability challenges in areas like model checking, investigating their ability to generalize beyond training data becomes necessary. This work presents the first study investigating whether recurrent neural networks (RNNs) can generalize to $\omega$-regular languages derived from LTL formulas. We train RNNs on ultimately periodic $\omega$-word sequences to replicate target BA behavior and evaluate how well they generalize to out-of-distribution sequences. Through experiments on LTL formulas corresponding to deterministic automata of varying structural complexity, from 3 to over 100 states, we show that RNNs achieve high accuracy on their target $\omega$-regular languages when evaluated on sequences up to $8 \times$ longer than training examples, with $92.6%$ of tasks achieving perfect or near-perfect generalization. These results establish the feasibility of neural approaches for learning complex $\omega$-regular languages, suggesting their potential as components in neurosymbolic verification methods.

[938] Surrogate Benchmarks for Model Merging Optimization

Rio Akizuki, Yuya Kudo, Nozomu Yoshinari, Yoichi Hirose, Toshiyuki Nishimoto, Kento Uchida, Shinichi Shirakawa

Main category: cs.LG

TL;DR: Surrogate benchmarks for efficient hyperparameter optimization in model merging, enabling low-cost algorithm development and performance comparison.

DetailsMotivation: Model merging hyperparameter tuning is computationally expensive, especially for large language models, creating a need for cost-effective optimization methods.

Method: Developed surrogate benchmarks by defining search spaces and collecting data samples to construct predictive models that estimate merged model performance from hyperparameters.

Result: The benchmarks accurately predict merged model performance and effectively simulate optimization algorithm behaviors.

Conclusion: Surrogate benchmarks provide a practical solution for developing and comparing hyperparameter optimization algorithms in model merging at reduced computational cost.

Abstract: Model merging techniques aim to integrate the abilities of multiple models into a single model. Most model merging techniques have hyperparameters, and their setting affects the performance of the merged model. Because several existing works show that tuning hyperparameters in model merging can enhance the merging outcome, developing hyperparameter optimization algorithms for model merging is a promising direction. However, its optimization process is computationally expensive, particularly in merging LLMs. In this work, we develop surrogate benchmarks for optimization of the merging hyperparameters to realize algorithm development and performance comparison at low cost. We define two search spaces and collect data samples to construct surrogate models to predict the performance of a merged model from a hyperparameter. We demonstrate that our benchmarks can predict the performance of merged models well and simulate optimization algorithm behaviors.

[939] MoPEQ: Mixture of Mixed Precision Quantized Experts

Krishna Teja Chitty-Venkata, Jie Ye, Murali Emani

Main category: cs.LG

TL;DR: MoPEQ is a post-training quantization method that assigns optimal bit widths to each expert in MoE-based VLMs using Hessian trace approximation for sensitivity analysis, achieving competitive accuracy with significant memory reduction.

DetailsMotivation: MoE-based large language and vision models have high computational and memory demands, making deployment challenging. Mixed precision quantization can help reduce these requirements while maintaining performance.

Method: Proposes MoPEQ algorithm that analyzes each expert’s sensitivity using Hessian trace approximation instead of activation frequency, clusters similar experts, and assigns optimal bit widths per expert (2-4 bits).

Result: Experimental results on VLMEvalKit benchmark with Deepseek-VL2 and MolmoE models show competitive accuracy with substantial memory footprint improvements compared to uniform-precision baselines.

Conclusion: The per-expert granularity approach with Hessian-based sensitivity analysis effectively balances accuracy and model size reduction for MoE-based VLMs, providing thorough understanding of mixed precision quantization.

Abstract: Large Language and Vision Models using a Mixture-of-Experts (MoE) architecture pose significant challenges for deployment due to their computational and memory demands. Mixed Precision Quantization assigns different precisions to different layers of an LLM/VLM based on layer sensitivity and importance within the model. In this work, we propose a Post Training Quantization algorithm, MoPEQ, that assigns optimal bit width to each expert. Our method balances accuracy and model size by analyzing each expert’s sensitivity using Hessian trace approximation instead of relying on the activation frequency of the expert. This per-expert granularity approach clusters similar experts to maintain model performance while reducing memory requirements. The experimental results on VLMEvalKit benchmark datasets using State-of-the-art VLMs Deepseek-VL2 -tiny, -small, -base, and MolmoE models demonstrate that our mixed precision quantized MoEs achieve competitive accuracy with substantial improvements in memory footprint compared to uniform-precision baseline methods. We perform a comprehensive study to analyze the impact of expert activation frequency and sensitivity using Hessian trace approximation at both layer-wise and model-wide expert precision allocation of 2, 3, and 4 bits to provide a thorough understanding of mixed precision quantization of VLM-MoEs.

[940] Is RL fine-tuning harder than regression? A PDE learning approach for diffusion models

Wenlong Mou

Main category: cs.LG

TL;DR: A new algorithm class for learning optimal control policies in diffusion process fine-tuning using value function approximation and HJB equations, achieving faster statistical rates than generic RL through supervised regression.

DetailsMotivation: To develop efficient methods for fine-tuning diffusion processes with provable statistical guarantees, addressing the challenge that generic reinforcement learning approaches have slower convergence rates for this specific problem.

Method: Developed algorithms by solving variational inequality problems based on Hamilton-Jacobi-Bellman (HJB) equations, using general value function approximation to learn optimal control policies.

Result: Proved sharp statistical rates for both learned value function and control policy, with rates depending on function class complexity and approximation errors. Showed fine-tuning can be achieved via supervised regression with faster statistical guarantees compared to generic RL.

Conclusion: The approach demonstrates that diffusion process fine-tuning can be efficiently solved using supervised regression methods with superior statistical rate guarantees, providing a more effective alternative to general reinforcement learning techniques for this specific problem domain.

Abstract: We study the problem of learning the optimal control policy for fine-tuning a given diffusion process, using general value function approximation. We develop a new class of algorithms by solving a variational inequality problem based on the Hamilton-Jacobi-Bellman (HJB) equations. We prove sharp statistical rates for the learned value function and control policy, depending on the complexity and approximation errors of the function class. In contrast to generic reinforcement learning problems, our approach shows that fine-tuning can be achieved via supervised regression, with faster statistical rate guarantees.

[941] Federated learning over physical channels: adaptive algorithms with near-optimal guarantees

Rui Zhang, Wenlong Mou

Main category: cs.LG

TL;DR: Proposes adaptive federated SGD algorithms for over-the-air communication that handle channel noise and hardware constraints, with theoretical convergence guarantees and practical effectiveness demonstrated through simulations.

DetailsMotivation: To reduce communication costs in federated learning by transmitting information over physical channels while addressing channel noise and hardware limitations.

Method: Developed a new class of adaptive federated stochastic gradient descent algorithms designed for implementation over physical channels, accounting for both channel noise and hardware constraints.

Result: Established theoretical convergence guarantees showing adaptive convergence rates to stochastic gradient noise level, and demonstrated practical effectiveness through simulation studies with deep learning models.

Conclusion: The proposed adaptive federated SGD algorithms successfully enable efficient over-the-air communication in federated learning while maintaining convergence performance despite channel noise and hardware constraints.

Abstract: In federated learning, communication cost can be significantly reduced by transmitting the information over the air through physical channels. In this paper, we propose a new class of adaptive federated stochastic gradient descent (SGD) algorithms that can be implemented over physical channels, taking into account both channel noise and hardware constraints. We establish theoretical guarantees for the proposed algorithms, demonstrating convergence rates that are adaptive to the stochastic gradient noise level. We also demonstrate the practical effectiveness of our algorithms through simulation studies with deep learning models.

[942] Understanding sparse autoencoder scaling in the presence of feature manifolds

Eric J. Michaud, Liv Gorton, Tom McGrath

Main category: cs.LG

TL;DR: SAE scaling laws show distinct regimes where feature manifolds can cause pathological behavior - learning fewer features than available latents.

DetailsMotivation: To understand how sparse autoencoders scale with the number of latents and how multi-dimensional feature manifolds influence scaling behavior.

Method: Adapted a capacity-allocation model from neural scaling literature to analyze SAE scaling, particularly focusing on feature manifolds’ effects.

Result: The model recovers distinct scaling regimes, with one pathological regime where feature manifolds cause SAEs to learn significantly fewer features than the number of latents available.

Conclusion: Feature manifolds can have pathological effects on SAE scaling, potentially limiting their effectiveness in real-world applications, requiring further investigation into whether SAEs operate in this regime in practice.

Abstract: Sparse autoencoders (SAEs) model the activations of a neural network as linear combinations of sparsely occurring directions of variation (latents). The ability of SAEs to reconstruct activations follows scaling laws w.r.t. the number of latents. In this work, we adapt a capacity-allocation model from the neural scaling literature (Brill, 2024) to understand SAE scaling, and in particular, to understand how “feature manifolds” (multi-dimensional features) influence scaling behavior. Consistent with prior work, the model recovers distinct scaling regimes. Notably, in one regime, feature manifolds have the pathological effect of causing SAEs to learn far fewer features in data than there are latents in the SAE. We provide some preliminary discussion on whether or not SAEs are in this pathological regime in the wild.

[943] Deep Tensor Network

Yifan Zhang

Main category: cs.LG

TL;DR: Deep Tensor Network replaces quadratic-complexity dot-product attention with efficient tensor-based operators that enable O(d²) per-token computation while capturing higher-order dependencies.

DetailsMotivation: Address the quadratic complexity bottleneck of Transformer's dot-product attention that limits foundation models from handling unbounded context lengths.

Method: Introduces Deep Tensor Network framework unifying tensor algebra with neural networks. Core operators: Tensor Attention (data-dependent polynomial kernels for token-mixing) and Tensor Interaction (adaptive channel-mixing), both using second-order summaries that avoid n×n matrices.

Result: Achieves causality-preserving streaming implementation with O(d²) per-token updates and O(d²) state, matching State Space Models efficiency while maintaining attention-like formulation.

Conclusion: Provides a principled new class of building blocks for next-generation sequence models that bridge scalable computation with rich expressive interaction modeling.

Abstract: The quadratic complexity of dot-product attention introduced in Transformer remains a fundamental bottleneck impeding the progress of foundation models toward unbounded context lengths. Addressing this challenge, we introduce the Deep Tensor Network, a new architectural framework that fundamentally reformulates attention by unifying the expressive power of tensor algebra with neural network design. Our approach moves beyond both conventional dot-product attention and subsequent linear-time approximations to capture higher-order statistical dependencies. We introduce two core operators derived from this framework: \emph{Tensor Attention}, which models complex token-mixing via data-dependent polynomial kernels, and Tensor Interaction, a novel mechanism for adaptive channel-mixing. We demonstrate that these operators are powered by second-order summaries that entirely bypass the formation of $n \times n$ matrices, enabling a causality-preserving streaming implementation with $O(d^2)$ per-token updates and $O(d^2)$ state. This efficiency rivals that of modern State Space Models while retaining an attention-like formulation. The Deep Tensor Network thus provides a principled and powerful new class of building blocks for next-generation sequence models, bridging the gap between scalable computation and rich, expressive interaction modeling.

[944] Towards Incremental Learning in Large Language Models: A Critical Review

Mladjan Jovanovic, Peter Voss

Main category: cs.LG

TL;DR: Comprehensive review of incremental learning approaches in Large Language Models, covering continual learning, meta-learning, parameter-efficient learning, and mixture-of-experts paradigms, highlighting current limitations and future challenges.

DetailsMotivation: Incremental learning enables AI systems to adapt and generalize to new tasks over time, which is critical for real-world applications where data changes frequently or is limited.

Method: Synthesizes state-of-the-art incremental learning paradigms through comprehensive analysis of existing research, examining specific achievements and critical factors from related topics.

Result: Key finding reveals that most approaches don’t update the core model and none support real-time incremental updates, while identifying current problems and research gaps.

Conclusion: The review provides a consolidated understanding of incremental learning for LLMs, offering insights for designing and developing more adaptive learning systems, while highlighting areas needing future research.

Abstract: Incremental learning is the ability of systems to acquire knowledge over time, enabling their adaptation and generalization to novel tasks. It is a critical ability for intelligent, real-world systems, especially when data changes frequently or is limited. This review provides a comprehensive analysis of incremental learning in Large Language Models. It synthesizes the state-of-the-art incremental learning paradigms, including continual learning, meta-learning, parameter-efficient learning, and mixture-of-experts learning. We demonstrate their utility for incremental learning by describing specific achievements from these related topics and their critical factors. An important finding is that many of these approaches do not update the core model, and none of them update incrementally in real-time. The paper highlights current problems and challenges for future research in the field. By consolidating the latest relevant research developments, this review offers a comprehensive understanding of incremental learning and its implications for designing and developing LLM-based learning systems.

[945] Towards Efficient Risk-Sensitive Policy Gradient: An Iteration Complexity Analysis

Rui Liu, Anish Gupta, Erfaun Noorani, Pratap Tokekar

Main category: cs.LG

TL;DR: Risk-sensitive REINFORCE achieves O(ε⁻²) iteration complexity for ε-approximate first-order stationary points and converges faster than risk-neutral counterparts in multiple environments.

DetailsMotivation: Traditional RL frameworks face challenges in iteration efficiency and safety. Risk-sensitive policy gradient methods offer safe policies but their iteration complexity remains underexplored.

Method: Rigorous iteration complexity analysis of risk-sensitive REINFORCE algorithm with exponential utility function, comparing with risk-neutral counterparts across CartPole, MiniGrid, and Robot Navigation environments.

Result: Established O(ε⁻²) iteration complexity for ε-FOSP. Empirical evaluation shows risk-sensitive REINFORCE converges and stabilizes faster than risk-neutral version in all tested environments.

Conclusion: Risk-sensitive policy gradient methods not only provide safety benefits but also demonstrate superior convergence efficiency compared to risk-neutral approaches, making them promising for practical RL applications.

Abstract: Reinforcement Learning (RL) has shown exceptional performance across various applications, enabling autonomous agents to learn optimal policies through interaction with their environments. However, traditional RL frameworks often face challenges in terms of iteration efficiency and safety. Risk-sensitive policy gradient methods, which incorporate both expected return and risk measures, have been explored for their ability to yield safe policies, yet their iteration complexity remains largely underexplored. In this work, we conduct a rigorous iteration complexity analysis for the risk-sensitive policy gradient method, focusing on the REINFORCE algorithm with an exponential utility function. We establish an iteration complexity of $\mathcal{O}(\epsilon^{-2})$ to reach an $\epsilon$-approximate first-order stationary point (FOSP). Furthermore, we investigate whether risk-sensitive algorithms can achieve better iteration complexity compared to their risk-neutral counterparts. Our analysis indicates that risk-sensitive REINFORCE can potentially converge faster. To validate our analysis, we empirically evaluate the learning performance and convergence efficiency of the risk-neutral and risk-sensitive REINFORCE algorithms in multiple environments: CartPole, MiniGrid, and Robot Navigation. Empirical results confirm that risk-sensitive cases can converge and stabilize faster compared to their risk-neutral counterparts. More details can be found on our website https://anonymous.4open.science/w/riskrl.

[946] Explaining Length Bias in LLM-Based Preference Evaluations

Zhengyu Hu, Linxin Song, Jieyu Zhang, Zheyuan Xiao, Tianfu Wang, Zhengyu Chen, Nicholas Jing Yuan, Jianxun Lian, Kaize Ding, Hui Xiong

Main category: cs.LG

TL;DR: The paper identifies and addresses LLM evaluation bias favoring longer responses by decomposing win rate into length-independent desirability and length-dependent information mass, proposing AdapAlpaca to align response lengths for fair comparisons.

DetailsMotivation: Large language models used as judges in preference comparisons show significant bias towards longer responses, which undermines the reliability of evaluation metrics.

Method: Decompose preference evaluation metrics into desirability (length-independent quality) and information mass (length-dependent content). Propose AdapAlpaca, a method that aligns reference and test model response lengths under equivalent intervals to ensure fair comparisons.

Result: Empirical experiments demonstrate that response length impacts evaluations primarily through information mass. The proposed decomposition successfully isolates length effects from content quality assessment.

Conclusion: AdapAlpaca provides a simple yet effective adjustment to win rate measurement that enables reliable evaluation of response quality without being confounded by response length bias.

Abstract: The use of large language models (LLMs) as judges, particularly in preference comparisons, has become widespread, but this reveals a notable bias towards longer responses, undermining the reliability of such evaluations. To better understand such bias, we propose to decompose the preference evaluation metric, specifically the win rate, into two key components: desirability and information mass, where the former is length-independent and related to trustworthiness such as correctness, toxicity, and consistency, and the latter is length-dependent and represents the amount of information in the response. We empirically demonstrated the decomposition through controlled experiments and found that response length impacts evaluations by influencing information mass. To derive a reliable evaluation metric that assesses content quality without being confounded by response length, we propose AdapAlpaca, a simple yet effective adjustment to win rate measurement. Specifically, AdapAlpaca ensures a fair comparison of response quality by aligning the lengths of reference and test model responses under equivalent length intervals.

[947] Learning to (Learn at Test Time): RNNs with Expressive Hidden States

Yu Sun, Xinhao Li, Karan Dalal, Jiarui Xu, Arjun Vikram, Genghan Zhang, Yann Dubois, Xinlei Chen, Xiaolong Wang, Sanmi Koyejo, Tatsunori Hashimoto, Carlos Guestrin

Main category: cs.LG

TL;DR: TTT layers combine linear complexity of RNNs with expressive hidden states by making the hidden state a machine learning model updated through self-supervised learning during test time.

DetailsMotivation: Self-attention has quadratic complexity while RNNs have limited expressive power in long contexts. There's a need for sequence modeling layers with both linear complexity and expressive hidden states.

Method: Propose Test-Time Training (TTT) layers where the hidden state is a machine learning model (linear model or MLP) and the update rule is a step of self-supervised learning. Two instantiations: TTT-Linear and TTT-MLP.

Result: TTT-Linear and TTT-MLP can keep reducing perplexity with more tokens (similar to Transformer), while Mamba plateaus after 16k context. TTT-MLP shows larger potential in long context despite memory I/O challenges.

Conclusion: TTT layers present a promising direction for long-context modeling with linear complexity, with TTT-MLP showing particular potential for future research despite current memory limitations.

Abstract: Self-attention performs well in long context but has quadratic complexity. Existing RNN layers have linear complexity, but their performance in long context is limited by the expressive power of their hidden states. We present a practical framework for instantiating sequence modeling layers with linear complexity and expressive hidden states. The key idea is to make the hidden state a machine learning model itself, and the update rule a step of self-supervised learning. Since the hidden state is updated by training even on test sequences, our layers are called Test-Time Training (TTT) layers. We consider two instantiations: TTT-Linear and TTT-MLP, whose hidden state is a linear model and a two-layer MLP respectively. We evaluate our instantiations at the scale of 125M to 1.3B parameters, comparing with a strong Transformer and Mamba, a modern RNN. Similar to Transformer, TTT-Linear and TTT-MLP can keep reducing perplexity by conditioning on more tokens, while Mamba cannot after 16k context. TTT-MLP still faces challenges in memory I/O, but shows larger potential in long context, pointing to a promising direction for future research.

[948] Learn while Unlearn: An Iterative Unlearning Framework for Generative Language Models

Haoyu Tang, Ye Liu, Xi Zhao, Xukai Liu, Yanghai Zhang, Kai Zhang, Xiaofang Zhou, Enhong Chen

Main category: cs.LG

TL;DR: ICU framework enables machine learning models to selectively forget sensitive data while maintaining performance, addressing privacy concerns without requiring original training data.

DetailsMotivation: Privacy regulations like GDPR require models to forget specific data, but existing unlearning methods need original training data and degrade model performance.

Method: Iterative Contrastive Unlearning (ICU) with three modules: Knowledge Unlearning Induction, Contrastive Learning Enhancement, and Iterative Unlearning Refinement.

Result: ICU effectively removes sensitive information while preserving model expressive capabilities and overall performance.

Conclusion: ICU provides a practical solution for privacy-conscious machine learning by enabling selective forgetting without original data access or performance degradation.

Abstract: Recent advances in machine learning, particularly in Natural Language Processing (NLP), have produced powerful models trained on vast datasets. However, these models risk leaking sensitive information, raising privacy concerns. In response, regulatory measures such as the European Union’s General Data Protection Regulation (GDPR) have driven increasing interest in Machine Unlearning techniques, which enable models to selectively forget specific data entries. Early unlearning approaches primarily relied on pre-processing methods, while more recent research has shifted towards training-based solutions. Despite their effectiveness, a key limitation persists: most methods require access to original training data, which is often unavailable. Additionally, directly applying unlearning techniques bears the cost of undermining the model’s expressive capabilities. To address these challenges, we introduce the Iterative Contrastive Unlearning (ICU) framework, which consists of three core components: A Knowledge Unlearning Induction module designed to target specific knowledge for removal using an unlearning loss; A Contrastive Learning Enhancement module to preserve the model’s expressive capabilities against the pure unlearning goal; And an Iterative Unlearning Refinement module that dynamically adjusts the unlearning process through ongoing evaluation and updates. Experimental results demonstrate the efficacy of our ICU method in unlearning sensitive information while maintaining the model’s overall performance, offering a promising solution for privacy-conscious machine learning applications.

[949] Space-aware Socioeconomic Indicator Inference with Heterogeneous Graphs

Xingchen Zou, Jiani Huang, Xixuan Hao, Yuhao Yang, Haomin Wen, Yibo Yan, Chao Huang, Chao Chen, Yuxuan Liang

Main category: cs.LG

TL;DR: GeoHG is a novel space-aware method that uses heterogeneous graphs to infer socioeconomic indicators from limited regional samples, outperforming traditional spatial interpolation methods.

DetailsMotivation: Regional socioeconomic indicators are costly to acquire but essential for urban management. Current methods rely on spatial continuity assumptions that don't capture complex regional variations.

Method: Uses a heterogeneous graph-based structure to represent geospace for non-continuous inference of socioeconomic indicators from limited samples.

Result: Achieves R² score exceeding 0.8 under extreme data scarcity conditions (95% masked ratio), demonstrating superior performance compared to existing methods.

Conclusion: GeoHG provides an effective space-aware approach for socioeconomic indicator inference that handles complex spatial variations and works well with very limited data.

Abstract: Regional socioeconomic indicators are critical across various domains, yet their acquisition can be costly. Inferring global socioeconomic indicators from a limited number of regional samples is essential for enhancing management and sustainability in urban areas and human settlements. Current inference methods typically rely on spatial interpolation based on the assumption of spatial continuity, which does not adequately address the complex variations present within regional spaces. In this paper, we present GeoHG, the first space-aware socioeconomic indicator inference method that utilizes a heterogeneous graph-based structure to represent geospace for non-continuous inference. Extensive experiments demonstrate the effectiveness of GeoHG in comparison to existing methods, achieving an $R^2$ score exceeding 0.8 under extreme data scarcity with a masked ratio of 95%.

[950] Leveraging Offline Data in Linear Latent Contextual Bandits

Chinmaya Kausik, Kevin Tan, Ambuj Tewari

Main category: cs.LG

TL;DR: Error: OutputParser failed

DetailsMotivation: Error: OutputParser failed

Method: Error: OutputParser failed

Result: Error: OutputParser failed

Conclusion: Error: OutputParser failed

Abstract: Leveraging offline data is an attractive way to accelerate online sequential decision-making. However, it is crucial to account for latent states in users or environments in the offline data, and latent bandits form a compelling model for doing so. In this light, we design end-to-end latent bandit algorithms capable of handing uncountably many latent states. We focus on a linear latent contextual bandit $-$ a linear bandit where each user has its own high-dimensional reward parameter in $\mathbb{R}^{d_A}$, but reward parameters across users lie in a low-rank latent subspace of dimension $d_K \ll d_A$. First, we provide an offline algorithm to learn this subspace with provable guarantees. We then present two online algorithms that utilize the output of this offline algorithm to accelerate online learning. The first enjoys $\tilde{O}(\min(d_A\sqrt{T}, d_K\sqrt{T}(1+\sqrt{d_AT/d_KN})))$ regret guarantees, so that the effective dimension is lower when the size $N$ of the offline dataset is larger. We prove a matching lower bound on regret, showing that our algorithm is minimax optimal. The second is a practical algorithm that enjoys only a slightly weaker guarantee, but is computationally efficient. We also establish the efficacy of our methods using experiments on both synthetic data and real-life movie recommendation data from MovieLens. Finally, we theoretically establish the generality of the latent bandit model by proving a de Finetti theorem for stateless decision processes.

[951] A Law of Next-Token Prediction in Large Language Models

Hangfeng He, Weijie J. Su

Main category: cs.LG

TL;DR: LLMs exhibit a universal law where each layer contributes equally to prediction accuracy improvement, observed across diverse architectures and pre-training data.

DetailsMotivation: Understanding how LLMs process input data internally is challenging due to their black-box nature, requiring quantitative analysis of their internal mechanisms.

Method: Introduced a precise quantitative law that governs contextualized token embeddings learning through intermediate layers in pre-trained LLMs for next-token prediction.

Result: Found that each layer contributes equally to enhancing prediction accuracy from lowest to highest layer, a universal phenomenon across diverse open-source LLMs regardless of architecture or pre-training data.

Conclusion: This law provides new perspectives and actionable insights to guide LLM development and applications, including model scaling, pre-training tasks, and interpretation.

Abstract: Large language models (LLMs) have been widely employed across various application domains, yet their black-box nature poses significant challenges to understanding how these models process input data internally to make predictions. In this paper, we introduce a precise and quantitative law that governs the learning of contextualized token embeddings through intermediate layers in pre-trained LLMs for next-token prediction. Our findings reveal that each layer contributes equally to enhancing prediction accuracy, from the lowest to the highest layer – a universal phenomenon observed across a diverse array of open-source LLMs, irrespective of their architectures or pre-training data. We demonstrate that this law offers new perspectives and actionable insights to inform and guide practices in LLM development and applications, including model scaling, pre-training tasks, and interpretation.

[952] Decomposing heterogeneous dynamical systems with graph neural networks

Cédric Allier, Magdalena C. Schneider, Michael Innerberger, Larissa Heinrich, John A. Bogovic, Stephan Saalfeld

Main category: cs.LG

TL;DR: Graph neural networks can learn interaction rules and latent heterogeneity from observable dynamics to virtually decompose complex systems and infer governing equations.

DetailsMotivation: Natural dynamical systems are complex with heterogeneous components and diverse interactions, making it difficult to understand their underlying governing rules.

Method: Simple graph neural networks are designed to jointly learn interaction rules and latent heterogeneity from observable dynamics, enabling virtual decomposition of complex systems.

Result: The approach was tested with simulation experiments of interacting moving particles, vector fields, and signaling networks, successfully learning latent heterogeneity and dynamics.

Conclusion: This method shows promise as a generally applicable tool to uncover governing rules underlying complex natural dynamics, though currently validated with simulated data.

Abstract: Natural physical, chemical, and biological dynamical systems are often complex, with heterogeneous components interacting in diverse ways. We show how simple graph neural networks can be designed to jointly learn the interaction rules and the latent heterogeneity from observable dynamics. The learned latent heterogeneity and dynamics can be used to virtually decompose the complex system which is necessary to infer and parameterize the underlying governing equations. We tested the approach with simulation experiments of interacting moving particles, vector fields, and signaling networks. While our current aim is to better understand and validate the approach with simulated data, we anticipate it to become a generally applicable tool to uncover the governing rules underlying complex dynamics observed in nature.

[953] Exploring Response Uncertainty in MLLMs: An Empirical Evaluation under Misleading Scenarios

Yunkai Dang, Mengxi Gao, Yibo Yan, Xin Zou, Yanggan Gu, Jungang Li, Jingyu Wang, Peijie Jiang, Aiwei Liu, Jia Liu, Xuming Hu

Main category: cs.LG

TL;DR: MLLMs are highly vulnerable to misleading information, with 65% of correct answers being overturned by deceptive cues. The study proposes a benchmark and fine-tuning method that significantly reduces this vulnerability.

DetailsMotivation: Existing studies focus on visual-textual misalignment but neglect MLLMs' ability to maintain correct answers when faced with misleading information, revealing a critical vulnerability.

Method: Two-stage evaluation pipeline: (1) get original responses on clean inputs, (2) inject explicit and implicit misleading instructions. Created MUB benchmark and fine-tuned models on 2000-sample dataset.

Result: Average misleading rates exceed 86% across models. Fine-tuning reduced rates to 6.97% (explicit) and 32.77% (implicit), improving consistency by 29.37% on deceptive inputs.

Conclusion: MLLMs show significant response uncertainty to misleading cues, but targeted fine-tuning can substantially improve their robustness while maintaining performance on standard benchmarks.

Abstract: Multimodal large language models (MLLMs) have recently achieved state-of-the-art performance on tasks ranging from visual question answering to video understanding. However, existing studies have concentrated mainly on visual-textual misalignment, leaving largely unexplored the MLLMs’ ability to preserve an originally correct answer when confronted with misleading information. We reveal a response uncertainty phenomenon: across nine standard datasets, twelve state-of-the-art open-source MLLMs overturn a previously correct answer in 65% of cases after receiving a single deceptive cue. To systematically quantify this vulnerability, we propose a two-stage evaluation pipeline: (1) elicit each model’s original response on unperturbed inputs; (2) inject explicit (false-answer hints) and implicit (contextual contradictions) misleading instructions, and compute the misleading rate - the fraction of correct-to-incorrect flips. Leveraging the most susceptible examples, we curate the Multimodal Uncertainty Benchmark (MUB), a collection of image-question pairs stratified into low, medium, and high difficulty based on how many of twelve state-of-the-art MLLMs they mislead. Extensive evaluation on twelve open-source and five closed-source models reveals a high uncertainty: average misleading rates exceed 86%, with explicit cues over 67.19% and implicit cues over 80.67%. To reduce the misleading rate, we then fine-tune all open-source MLLMs on a compact 2000-sample mixed-instruction dataset, reducing misleading rates to 6.97% (explicit) and 32.77% (implicit), boosting consistency by nearly 29.37% on highly deceptive inputs, and slightly improving accuracy on standard benchmarks. Our code is available at https://github.com/Yunkaidang/uncertainty

[954] Re-examining learning linear functions in context

Omar Naim, Guilhem Fouilhé, Nicholas Asher

Main category: cs.LG

TL;DR: Transformers fail to generalize beyond training distribution in linear function ICL, challenging algorithmic learning narratives

DetailsMotivation: To understand how in-context learning actually works in transformers, particularly for linear functions, as current understanding is limited

Method: Used controlled setup with synthetic training data and trained GPT-2-like transformers from scratch to study ICL of univariate linear functions

Result: Models failed to generalize beyond training distribution, contradicting the prevailing view that transformers use algorithmic approaches like linear regression

Conclusion: Proposed a mathematically precise hypothesis about what models actually learn, highlighting fundamental limitations in inferring abstract task structures

Abstract: In-context learning (ICL) has emerged as a powerful paradigm for easily adapting Large Language Models (LLMs) to various tasks. However, our understanding of how ICL works remains limited. We explore a simple model of ICL in a controlled setup with synthetic training data to investigate ICL of univariate linear functions. We experiment with a range of GPT-2-like transformer models trained from scratch. Our findings challenge the prevailing narrative that transformers adopt algorithmic approaches like linear regression to learn a linear function in-context. These models fail to generalize beyond their training distribution, highlighting fundamental limitations in their capacity to infer abstract task structures. Our experiments lead us to propose a mathematically precise hypothesis of what the model might be learning.

[955] Adversarial Attacks and Defenses in Multivariate Time-Series Forecasting for Smart and Connected Infrastructures

Pooja Krishan, Rohan Mohapatra, Sanchari Das, Saptarshi Sengupta

Main category: cs.LG

TL;DR: Adversarial attacks can fool time-series forecasting models with subtle input modifications. The paper demonstrates attack feasibility using FGSM/BIM methods and develops robust defenses through adversarial training, achieving significant RMSE improvements.

DetailsMotivation: Deep learning models are vulnerable to adversarial attacks that cause incorrect predictions with high confidence, posing security risks for time-series forecasting applications.

Method: Used untargeted white-box attacks (FGSM and BIM) to poison training inputs, then developed robust models through adversarial training and model hardening techniques.

Result: Achieved 72.41% and 94.81% decrease in RMSE for electricity and hard disk datasets respectively after implementing adversarial defenses, demonstrating successful attack transferability.

Conclusion: Adversarial attacks pose serious threats to time-series forecasting, but robust defenses through adversarial training can effectively mitigate these security concerns across different domains.

Abstract: The emergence of deep learning models has revolutionized various industries over the last decade, leading to a surge in connected devices and infrastructures. However, these models can be tricked into making incorrect predictions with high confidence, leading to disastrous failures and security concerns. To this end, we explore the impact of adversarial attacks on multivariate time-series forecasting and investigate methods to counter them. Specifically, we employ untargeted white-box attacks, namely the Fast Gradient Sign Method (FGSM) and the Basic Iterative Method (BIM), to poison the inputs to the training process, effectively misleading the model. We also illustrate the subtle modifications to the inputs after the attack, which makes detecting the attack using the naked eye quite difficult. Having demonstrated the feasibility of these attacks, we develop robust models through adversarial training and model hardening. We are among the first to showcase the transferability of these attacks and defenses by extrapolating our work from the benchmark electricity data to a larger, 10-year real-world data used for predicting the time-to-failure of hard disks. Our experimental results confirm that the attacks and defenses achieve the desired security thresholds, leading to a 72.41% and 94.81% decrease in RMSE for the electricity and hard disk datasets respectively after implementing the adversarial defenses.

[956] Learning in complex action spaces without policy gradients

Arash Tavakoli, Sina Ghiassian, Nemanja Rakićević

Main category: cs.LG

TL;DR: Policy gradient and action-value methods are equivalent in small action spaces but diverge in complex ones. The paper shows this superiority comes from universal principles that can be applied to action-value methods via QMLE framework.

DetailsMotivation: To understand why policy gradient methods outperform action-value methods in complex action spaces and demonstrate this superiority stems from universal principles rather than intrinsic qualities of policy gradients.

Method: Identified three universal principles from policy gradients and created QMLE (Q-learning with maximum likelihood estimation) framework to incorporate these principles into action-value methods without using policy gradients.

Result: QMLE achieves comparable computational cost to policy gradient methods in complex action spaces and shows strong performance on DeepMind Control Suite, competing with state-of-the-art methods like DMPO and D4PG.

Conclusion: The apparent superiority of policy gradients in complex action spaces is not intrinsic but comes from universal principles that can be successfully applied to action-value methods through frameworks like QMLE.

Abstract: While conventional wisdom holds that policy gradient methods are better suited to complex action spaces than action-value methods, foundational work has shown that the two paradigms are equivalent in small, finite action spaces (O’Donoghue et al., 2017; Schulman et al., 2017a). This raises the question of why their computational applicability and performance diverge as the complexity of the action space increases. We hypothesize that the apparent superiority of policy gradients in such settings stems not from intrinsic qualities of the paradigm but from universal principles that can also be applied to action-value methods, enabling similar functions. We identify three such principles and provide a framework for incorporating them into action-value methods. To support our hypothesis, we instantiate this framework in what we term QMLE, for Q-learning with maximum likelihood estimation. Our results show that QMLE can be applied to complex action spaces at a computational cost comparable to that of policy gradient methods, all without using policy gradients. Furthermore, QMLE exhibits strong performance on the DeepMind Control Suite, even when compared to state-of-the-art methods such as DMPO and D4PG.

[957] Beyond the Kolmogorov Barrier: A Learnable Weighted Hybrid Autoencoder for Model Order Reduction

Nithin Somasekharan, Shaowu Pan

Main category: cs.LG

TL;DR: A hybrid autoencoder combining SVD with deep learning using learnable weights to overcome convergence issues in high-dimensional physical system representation learning.

DetailsMotivation: To address poor convergence behavior of deep autoencoders as latent space rank increases, overcoming the Kolmogorov barrier for reduced-order modeling of complex physical systems.

Method: Proposed learnable weighted hybrid autoencoder that combines singular value decomposition (SVD) with deep autoencoders through learnable weighting parameters.

Result: Significantly improved generalization performance on chaotic PDE systems (1D Kuramoto-Sivashinsky and forced isotropic turbulence), with trained models showing sharpness thousands of times smaller than other models.

Conclusion: The hybrid approach with learnable weights is essential for effective representation learning and offers significant improvements for surrogate modeling when combined with time series techniques like Koopman operators and LSTMs.

Abstract: Representation learning for high-dimensional, complex physical systems aims to identify a low-dimensional intrinsic latent space, which is crucial for reduced-order modeling and modal analysis. To overcome the well-known Kolmogorov barrier, deep autoencoders (AEs) have been introduced in recent years, but they often suffer from poor convergence behavior as the rank of the latent space increases. To address this issue, we propose the learnable weighted hybrid autoencoder, a hybrid approach that combines the strengths of singular value decomposition (SVD) with deep autoencoders through a learnable weighted framework. We find that the introduction of learnable weighting parameters is essential – without them, the resulting model would either collapse into a standard POD or fail to exhibit the desired convergence behavior. Interestingly, we empirically find that our trained model has a sharpness thousands of times smaller compared to other models. Our experiments on classical chaotic PDE systems, including the 1D Kuramoto-Sivashinsky and forced isotropic turbulence datasets, demonstrate that our approach significantly improves generalization performance compared to several competing methods. Additionally, when combining with time series modeling techniques (e.g., Koopman operator, LSTM), the proposed technique offers significant improvements for surrogate modeling of high-dimensional multi-scale PDE systems.

[958] Training and Evaluating with Human Label Variation: An Empirical Study

Kemal Kurniawan, Meladel Mistica, Timothy Baldwin, Jey Han Lau

Main category: cs.LG

TL;DR: New fuzzy set-based evaluation metrics for human label variation (HLV) are proposed and tested as differentiable training objectives, but traditional methods using disaggregated annotations or soft labels outperform them across multiple datasets.

DetailsMotivation: Human label variation challenges the single ground truth assumption in machine learning, and there's a need to understand which methods and metrics perform best in HLV settings.

Method: Proposed new differentiable evaluation metrics based on fuzzy set theory, tested them as training objectives, and conducted extensive experiments across 6 HLV datasets comparing 14 training methods and 6 evaluation metrics.

Result: Training on disaggregated annotations or soft labels performed best across metrics, outperforming training with the proposed differentiable metrics. The proposed soft micro F1 score was identified as one of the best metrics for HLV data.

Conclusion: While the proposed fuzzy set-based metrics show promise, traditional approaches using disaggregated data or soft labels remain more effective for handling human label variation in model training and evaluation.

Abstract: Human label variation (HLV) challenges the standard assumption that a labelled instance has a single ground truth, instead embracing the natural variation in human annotation to train and evaluate models. While various training methods and metrics for HLV have been proposed, it is still unclear which methods and metrics perform best in what settings. We propose new evaluation metrics for HLV leveraging fuzzy set theory. Since these new proposed metrics are differentiable, we then in turn experiment with employing these metrics as training objectives. We conduct an extensive study over 6 HLV datasets testing 14 training methods and 6 evaluation metrics. We find that training on either disaggregated annotations or soft labels performs best across metrics, outperforming training using the proposed training objectives with differentiable metrics. We also show that our proposed soft micro F1 score is one of the best metrics for HLV data.

[959] Rotate, Clip, and Partition: Towards W2A4KV4 Quantization by Integrating Rotation and Learnable Non-uniform Quantizer

Euntae Choi, Sumin Song, Woosang Lim, Sungjoo Yoo

Main category: cs.LG

TL;DR: RCP is a quantization-aware training method that achieves extreme compression of LLMs to W2A4KV4 (2-bit weights, 4-bit activations, 4-bit KV cache) with minimal performance loss.

DetailsMotivation: To enable efficient deployment of large language models on memory-constrained devices by achieving extreme compression while maintaining model performance.

Method: Integrates rotation techniques with novel non-uniform weight quantizer design using Learnable Direct Partitioning (LDP), and develops specialized GPU kernel for GEMV operations on non-uniform W2A4 quantization.

Result: Compresses LLaMA-2-7B to W2A4KV4 with only 2.84 WikiText2 perplexity loss and 5.29x memory reduction. Successfully quantizes mobile-targeted LLaMA-3.2 and domain-specific models without convergence issues.

Conclusion: RCP enables extreme LLM compression to W2A4KV4 configuration with minimal performance degradation, making large models deployable on memory-constrained devices.

Abstract: We propose Rotate, Clip, and Partition (RCP), a quantization-aware training (QAT) approach that first realizes extreme compression of LLMs with W2A4KV4(2-bit weight, 4-bit activation, and 4-bit KV cache) configuration. RCP integrates recent rotation techniques with a novel non-uniform weight quantizer design, by quantitatively analyzing the impact of random rotation on 2-bit weight quantization. Our weight quantizer features Learnable Direct Partitioning (LDP), which introduces learnable parameters to directly learn non-uniform intervals jointly with LLM weights. We also present a specialized GPU kernel that supports GEMV on non-uniform W2A4. Experiments show that RCP can compress LLaMA-2-7B to W2A4KV4 with a loss of only 2.84 WikiText2 ppl and 5.29 times reduced memory footprint. Furthermore, RCP can quantize challenging mobile-targeted LLaMA-3.2 models and domain-specific WizardCoder-7B and MetaMath-7B with no critical problems such as convergence failure and repetition. Code is available at https://github.com/ songsm921/RCP.

[960] Skill-Enhanced Reinforcement Learning Acceleration from Heterogeneous Demonstrations

Hanping Zhang, Yuhong Guo

Main category: cs.LG

TL;DR: SeRLA is a two-stage method that uses skill-level adversarial PU learning to extract knowledge from expert and low-cost demonstrations, then employs skill-based soft actor-critic for efficient RL training with data enhancement to handle sparse data.

DetailsMotivation: Limited availability of expert demonstration data hinders effective Learning from Demonstration (LfD) for rapid reinforcement learning acceleration.

Method: Two-stage approach: 1) Skill-level adversarial Positive-Unlabeled learning to extract skill priors from expert and low-cost demonstrations, 2) Skill-based soft actor-critic algorithm for downstream RL training with skill-level data enhancement.

Result: Achieves state-of-the-art performance in accelerating reinforcement learning on downstream tasks, particularly in early training phase across multiple standard RL benchmarks.

Conclusion: SeRLA effectively addresses data scarcity in LfD by leveraging both expert and low-cost demonstrations through skill-level learning and data enhancement techniques.

Abstract: Learning from Demonstration (LfD) is a well-established problem in Reinforcement Learning (RL), which aims to facilitate rapid RL by leveraging expert demonstrations to pre-train the RL agent. However, the limited availability of expert demonstration data often hinders its ability to effectively aid downstream RL learning. To address this problem, we propose a novel two-stage method dubbed as Skill-enhanced Reinforcement Learning Acceleration (SeRLA). SeRLA introduces a skill-level adversarial Positive-Unlabeled (PU) learning model that extracts useful skill prior knowledge by learning from both expert demonstrations and general low-cost demonstrations in the offline prior learning stage. Building on this, it employs a skill-based soft actor-critic algorithm to leverage the acquired priors for efficient training of a skill policy network in the downstream online RL stage. In addition, we propose a simple skill-level data enhancement technique to mitigate data sparsity and further improve both skill prior learning and skill policy training. Experiments across multiple standard RL benchmarks demonstrate that SeRLA achieves state-of-the-art performance in accelerating reinforcement learning on downstream tasks, particularly in the early training phase.

[961] More Bang for the Buck: Process Reward Modeling with Entropy-Driven Uncertainty

Lang Cao, Renhong Chen, Yingtian Zou, Chao Peng, Wu Ning, Huacong Xu, Qian Chen, Yuxian Wang, Peishuo Su, Mofan Peng, Zijie Chen, Yitong Li

Main category: cs.LG

TL;DR: EDU-PRM is a novel entropy-driven framework that automatically segments reasoning steps using predictive entropy, eliminating manual annotations and achieving state-of-the-art performance on math reasoning tasks with superior efficiency.

DetailsMotivation: Existing Process Reward Models require costly manual step annotations and static partitioning, which limits scalability and efficiency in complex reasoning tasks.

Method: Uses entropy-driven training to automatically anchor step boundaries at tokens with high predictive entropy, enabling dynamic segmentation without manual annotations.

Result: Achieves 65.5% accuracy on MATH test set (surpassing baselines), 67.3% accuracy with EDU sampling (47% token reduction), and 88.4% SOTA on ProcessBench using <1.5% training data.

Conclusion: EDU-PRM provides a scalable, annotation-efficient paradigm for process supervision in mathematical reasoning, enabling efficient complex reasoning without manual step labeling.

Abstract: We introduce the Entropy Driven Uncertainty Process Reward Model (EDU-PRM), a novel entropy-driven training framework for process reward modeling that enables dynamic, uncertainty-aligned segmentation of complex reasoning steps, eliminating the need for costly manual step annotations. Unlike previous Process Reward Models (PRMs) that rely on static partitioning and human labeling, EDU-PRM automatically anchors step boundaries at tokens with high predictive entropy. On the MATH test set, EDU-PRM achieves 65.5% accuracy, surpassing strong public PRM baselines such as Math-Shepherd PRM (61.7%) and Omega PRM (62.4%) under the High Temperature (HT) Sample + BON setting. Furthermore, when replacing HT sampling with EDU sampling, EDU-PRM further improves both accuracy and efficiency: at N=64, accuracy increases from 64.7% (HT Sample + BON) to 67.3% (EDU Sample + BON), while the number of generated tokens is reduced by 47%, demonstrating a superior accuracy-cost balance. On the ProcessBench test set, EDU-PRM achieves a new state-of-the-art accuracy of 88.4% using less than 1.5% of the Qwen2.5-Math-PRM-72B training data, surpassing the previous best of 87.8%. In summary, EDU-PRM provides a scalable and annotation-efficient paradigm for process supervision in mathematical reasoning, opening new avenues for efficient complex reasoning on math.

[962] ViSymRe: Vision-guided Multimodal Symbolic Regression

Da Li, Junping Yin, Jin Xu, Xinxin Li, Juan Zhang

Main category: cs.LG

TL;DR: ViSymRe is a vision-guided multimodal symbolic regression model that bridges the modality gap between datasets and mathematical expressions by incorporating expression graphs as a third resource, achieving better performance than dataset-only baselines.

DetailsMotivation: Traditional symbolic regression models face challenges like low efficiency and overfitting, while recent LLM-based approaches struggle with modality gaps between input datasets and output expressions. The paper aims to address these limitations by incorporating visual guidance.

Method: ViSymRe introduces a multimodal approach that extracts “virtual vision” from datasets and incorporates expression graphs as a third modality to bridge the gap between datasets and mathematical expressions, without requiring expression graphs during inference.

Result: Evaluation on multiple benchmarks shows ViSymRe achieves more competitive performance than state-of-the-art dataset-only baselines, producing expressions that fit datasets well while being simple and structurally accurate.

Conclusion: The vision-guided multimodal approach successfully addresses modality gap challenges in symbolic regression, producing high-quality mathematical expressions that balance accuracy, simplicity, and structural correctness.

Abstract: Extracting simple mathematical expression from an observational dataset to describe complex natural phenomena is one of the core objectives of artificial intelligence (AI). This field is known as symbolic regression (SR). Traditional SR models are based on genetic programming (GP) or reinforcement learning (RL), facing well-known challenges, such as low efficiency and overfitting. Recent studies have integrated SR with large language models (LLMs), enabling fast zero-shot inference by learning mappings from millions of dataset-expression pairs. However, since the input and output are inherently different modalities, such models often struggle to converge effectively. In this paper, we introduce ViSymRe, a vision-guided multimodal SR model that incorporates the third resource, expression graph, to bridge the modality gap. Different from traditional multimodal models, ViSymRe is trained to extract vision, termed virtual vision, from datasets, without relying on the global availability of expression graphs, which addresses the essential challenge of visual SR, i.e., expression graphs are not available during inference. Evaluation results on multiple mainstream benchmarks show that ViSymRe achieves more competitive performance than the state-of-the-art dataset-only baselines. The expressions predicted by ViSymRe not only fit the dataset well but are also simple and structurally accurate, goals that SR models strive to achieve.

[963] APEX$^2$: Adaptive and Extreme Summarization for Personalized Knowledge Graphs

Zihao Li, Dongqi Fu, Mengting Ai, Jingrui He

Main category: cs.LG

TL;DR: APEX² is a scalable framework for adaptive personalized knowledge graph summarization that handles evolving user interests with extremely small size constraints (≤0.1% compression), outperforming existing methods in accuracy and efficiency.

DetailsMotivation: Existing PKG summarization methods assume static user interests and fail when size constraints are extremely small, unable to distinguish immediate interests or guarantee utility.

Method: APEX² constructs initial PKG then continuously tracks interest shifts to adjust previous summaries, designed with robust theoretical guarantees for adaptive summarization.

Result: Evaluated on benchmark KGs with up to 12M triples and ≤0.1% compression ratios, APEX² outperforms state-of-the-art baselines in query-answering accuracy and efficiency.

Conclusion: APEX² successfully addresses limitations of existing methods by providing adaptive summarization with theoretical guarantees for evolving user interests under extreme size constraints.

Abstract: Knowledge graphs (KGs), which store an extensive number of relational facts, serve various applications. Recently, personalized knowledge graphs (PKGs) have emerged as a solution to optimize storage costs by customizing their content to align with users’ specific interests within particular domains. In the real world, on one hand, user queries and their underlying interests are inherently evolving, requiring PKGs to adapt continuously; on the other hand, the summarization is constantly expected to be as small as possible in terms of storage cost. However, the existing PKG summarization methods implicitly assume that the user’s interests are constant and do not shift. Furthermore, when the size constraint of PKG is extremely small, the existing methods cannot distinguish which facts are more of immediate interest and guarantee the utility of the summarized PKG. To address these limitations, we propose APEX$^2$, a highly scalable PKG summarization framework designed with robust theoretical guarantees to excel in adaptive summarization tasks with extremely small size constraints. To be specific, after constructing an initial PKG, APEX$^2$ continuously tracks the interest shift and adjusts the previous summary. We evaluate APEX$^2$ under an evolving query setting on benchmark KGs containing up to 12 million triples, summarizing with compression ratios $\leq 0.1%$. The experiments show that APEX outperforms state-of-the-art baselines in terms of both query-answering accuracy and efficiency. Code is available at https://github.com/iDEA-iSAIL-Lab-UIUC/APEX.

[964] Goal-Conditioned Data Augmentation for Offline Reinforcement Learning

Xingshuai Huang, Di Wu, Benoit Boulet

Main category: cs.LG

TL;DR: GODA is a goal-conditioned diffusion-based data augmentation method that enhances offline RL by generating higher-quality samples from suboptimal datasets using return-oriented guidance and adaptive conditioning.

DetailsMotivation: Offline RL struggles with suboptimal datasets that lack sufficient high-quality demonstrations, limiting policy learning performance.

Method: Goal-conditioned diffusion with return-oriented goal conditions, controllable scaling for return guidance, and adaptive gated conditioning for noisy inputs.

Result: Demonstrated effectiveness on D4RL benchmark and real-world traffic signal control tasks, outperforming state-of-the-art data augmentation methods.

Conclusion: GODA successfully maximizes utility of limited optimal demonstrations and enhances offline RL performance through quality data augmentation.

Abstract: Offline reinforcement learning (RL) enables policy learning from pre-collected offline datasets, relaxing the need to interact directly with the environment. However, limited by the quality of offline datasets, it generally fails to learn well-qualified policies in suboptimal datasets. To address datasets with insufficient optimal demonstrations, we introduce Goal-cOnditioned Data Augmentation (GODA), a novel goal-conditioned diffusion-based method for augmenting samples with higher quality. Leveraging recent advancements in generative modelling, GODA incorporates a novel return-oriented goal condition with various selection mechanisms. Specifically, we introduce a controllable scaling technique to provide enhanced return-based guidance during data sampling. GODA learns a comprehensive distribution representation of the original offline datasets while generating new data with selectively higher-return goals, thereby maximizing the utility of limited optimal demonstrations. Furthermore, we propose a novel adaptive gated conditioning method for processing noisy inputs and conditions, enhancing the capture of goal-oriented guidance. We conduct experiments on the D4RL benchmark and real-world challenges, specifically traffic signal control (TSC) tasks, to demonstrate GODA’s effectiveness in enhancing data quality and superior performance compared to state-of-the-art data augmentation methods across various offline RL algorithms.

[965] Fine-tuning Quantized Neural Networks with Zeroth-order Optimization

Sifeng Shang, Jiayi Zhou, Chenyu Lin, Minxian Li, Kaiyang Zhou

Main category: cs.LG

TL;DR: QZO enables memory-efficient 4-bit LLM fine-tuning using zeroth-order optimization and quantization, reducing memory by 18x compared to 16-bit full-parameter fine-tuning.

DetailsMotivation: GPU memory is a bottleneck for adapting large language models to downstream tasks due to exponential model size growth, requiring minimization of memory usage on weights, gradients, and optimizer states.

Method: Proposes Quantized Zeroth-order Optimization (QZO) that uses zeroth-order optimization to eliminate gradients and optimizer states, combined with model quantization (bfloat16 to int4). QZO perturbs continuous quantization scale for gradient estimation and uses directional derivative clipping to stabilize training.

Result: QZO reduces total memory cost by more than 18x for 4-bit LLMs compared to 16-bit full-parameter fine-tuning, enabling fine-tuning of Llama-2-13B within a single 24GB GPU.

Conclusion: QZO provides a simple yet effective unified framework for memory-efficient training that works with both scalar-based and codebook-based quantization methods, making large model fine-tuning more accessible with limited GPU resources.

Abstract: As the size of large language models grows exponentially, GPU memory has become a bottleneck for adapting these models to downstream tasks. In this paper, we aim to push the limits of memory-efficient training by minimizing memory usage on model weights, gradients, and optimizer states, within a unified framework. Our idea is to eliminate both gradients and optimizer states using zeroth-order optimization, which approximates gradients by perturbing weights during forward passes to identify gradient directions. To minimize memory usage on weights, we employ model quantization, e.g., converting from bfloat16 to int4. However, directly applying zeroth-order optimization to quantized weights is infeasible due to the precision gap between discrete weights and continuous gradients, which would otherwise require de-quantization and re-quantization. To overcome this challenge, we propose Quantized Zeroth-order Optimization (QZO), a simple yet effective approach that perturbs the continuous quantization scale for gradient estimation and uses a directional derivative clipping method to stabilize training. QZO is orthogonal to both scalar-based and codebook-based post-training quantization methods. Compared to full-parameter fine-tuning in 16 bits, QZO can reduce the total memory cost by more than 18$\times$ for 4-bit LLMs, and enables fine-tuning Llama-2-13B within a single 24GB GPU. Code will be released publicly.

[966] How Can I Publish My LLM Benchmark Without Giving the True Answers Away?

Takashi Ishida, Thanawat Lodkaew, Ikko Yamane

Main category: cs.LG

TL;DR: Proposes a method to publish LLM benchmarks without disclosing ground-truth answers by injecting randomness through multiple logically correct answers, enabling open evaluation while detecting data contamination.

DetailsMotivation: Current benchmark publishing risks contaminating future LLMs through unintentional training data inclusion, and private benchmarks require trust in a single organization while still allowing test-set overfitting.

Method: Inject randomness by preparing several logically correct answers for each question and including only one as the solution, reducing the Bayes accuracy ceiling to detect contamination.

Result: Experimental evidence shows the method can accurately detect data contamination across various benchmarks, models, and training methodologies.

Conclusion: This approach enables open benchmark publishing while maintaining evaluation integrity and providing a reliable mechanism for detecting data contamination in LLMs.

Abstract: Publishing a large language model (LLM) benchmark on the Internet risks contaminating future LLMs: the benchmark may be unintentionally (or intentionally) used to train or select a model. A common mitigation is to keep the benchmark private and let participants submit their models or predictions to the organizers. However, this strategy will require trust in a single organization and still permits test-set overfitting through repeated queries. To overcome this issue, we propose a way to publish benchmarks without completely disclosing the ground-truth answers to the questions, while still maintaining the ability to openly evaluate LLMs. Our main idea is to inject randomness to the answers by preparing several logically correct answers, and only include one of them as the solution in the benchmark. This reduces the best possible accuracy, i.e., Bayes accuracy, of the benchmark. Not only is this helpful to keep us from disclosing the ground truth, but this approach also offers a test for detecting data contamination. In principle, even fully capable models should not surpass the Bayes accuracy. If a model surpasses this ceiling despite this expectation, this is a strong signal of data contamination. We present experimental evidence that our method can detect data contamination accurately on a wide range of benchmarks, models, and training methodologies.

[967] The Complexity of Learning Sparse Superposed Features with Feedback

Akash Kumar

Main category: cs.LG

TL;DR: This paper investigates whether learned features from deep networks can be efficiently retrieved using relative triplet comparisons from an agent like an LLM, establishing tight bounds on feedback complexity and validating through experiments.

DetailsMotivation: To understand if underlying learned features of deep networks can be efficiently extracted through feedback mechanisms, particularly using relative triplet comparisons from agents like large language models.

Method: Analyze feedback complexity for learning feature matrices in sparse settings using relative triplet comparisons, establish theoretical bounds for constructed activations and distributional information, and validate through experiments on Recursive Feature Machines and dictionary extraction from sparse autoencoders.

Result: Established tight bounds for feature retrieval when agents can construct activations, demonstrated strong upper bounds in sparse scenarios with limited distributional feedback, and validated theoretical findings through successful experiments on two distinct applications.

Conclusion: The research shows that learned features from deep networks can be efficiently retrieved through relative triplet feedback from agents, with proven theoretical bounds and practical validation across different applications.

Abstract: The success of deep networks is crucially attributed to their ability to capture latent features within a representation space. In this work, we investigate whether the underlying learned features of a model can be efficiently retrieved through feedback from an agent, such as a large language model (LLM), in the form of relative \tt{triplet comparisons}. These features may represent various constructs, including dictionaries in LLMs or a covariance matrix of Mahalanobis distances. We analyze the feedback complexity associated with learning a feature matrix in sparse settings. Our results establish tight bounds when the agent is permitted to construct activations and demonstrate strong upper bounds in sparse scenarios when the agent’s feedback is limited to distributional information. We validate our theoretical findings through experiments on two distinct applications: feature recovery from Recursive Feature Machines and dictionary extraction from sparse autoencoders trained on Large Language Models.

[968] Harnessing Vision Models for Time Series Analysis: A Survey

Jingchao Ni, Ziming Zhao, ChengAo Shen, Hanghang Tong, Dongjin Song, Wei Cheng, Dongsheng Luo, Haifeng Chen

Main category: cs.LG

TL;DR: Survey paper exploring the use of vision models (LVMs/VLMs) instead of LLMs for time series analysis, discussing encoding methods, modeling approaches, and future directions.

DetailsMotivation: The discrepancy between continuous time series and discrete token spaces of LLMs, plus challenges in modeling multivariate correlations, make vision models a promising alternative for time series analysis.

Method: Comprehensive survey with dual-view taxonomy: how to encode time series as images, and how to model imaged time series for various tasks. Addresses pre- and post-processing challenges.

Result: Provides systematic overview of vision model applications in time series analysis, highlighting advantages over LLMs and identifying key research questions and methodologies.

Conclusion: Vision models offer significant potential for time series analysis, with the survey outlining future directions to advance this emerging research area.

Abstract: Time series analysis has witnessed the inspiring development from traditional autoregressive models, deep learning models, to recent Transformers and Large Language Models (LLMs). Efforts in leveraging vision models for time series analysis have also been made along the way but are less visible to the community due to the predominant research on sequence modeling in this domain. However, the discrepancy between continuous time series and the discrete token space of LLMs, and the challenges in explicitly modeling the correlations of variates in multivariate time series have shifted some research attentions to the equally successful Large Vision Models (LVMs) and Vision Language Models (VLMs). To fill the blank in the existing literature, this survey discusses the advantages of vision models over LLMs in time series analysis. It provides a comprehensive and in-depth overview of the existing methods, with dual views of detailed taxonomy that answer the key research questions including how to encode time series as images and how to model the imaged time series for various tasks. Additionally, we address the challenges in the pre- and post-processing steps involved in this framework and outline future directions to further advance time series analysis with vision models.

[969] A Markov Categorical Framework for Language Modeling

Yifan Zhang

Main category: cs.LG

TL;DR: A unified theoretical framework using Markov categories to explain how autoregressive language models work, connecting training objectives, representation geometry, and practical capabilities.

DetailsMotivation: To develop a unified theory explaining the internal mechanisms of autoregressive language models and how training shapes their representations and capabilities.

Method: Introduces a compositional analytical framework using Markov categories to model single-step generation, connecting training objectives, representation space geometry, and model capabilities through information theory.

Result: Provides information-theoretic rationale for multi-token prediction methods, clarifies how NLL training forces models to learn conditional uncertainty, and reveals that NLL functions as implicit spectral contrastive learning that structures representation spaces.

Conclusion: Offers a powerful new framework to understand information flow in language models and how training objectives shape internal geometry, bridging learning theory with practical model success.

Abstract: Autoregressive language models achieve remarkable performance, yet a unified theory explaining their internal mechanisms–how training shapes their representations and enables complex behaviors–remains elusive. We introduce a new analytical framework that models the single-step generation process as a composition of information-processing stages using the language of Markov categories. This compositional perspective provides a unified mathematical language to connect three critical aspects of language modeling that are typically studied in isolation: the training objective, the geometry of the learned representation space, and practical model capabilities. First, our framework provides a precise information-theoretic rationale for the success of multi-token prediction methods like speculative decoding, quantifying the “information surplus” a model’s hidden state contains about tokens beyond the immediate next one. Second, we clarify how the standard negative log-likelihood (NLL) objective compels the model to learn not just the next word, but also the data’s intrinsic conditional uncertainty, a process we formalize using categorical entropy. Our central result reveals that NLL training functions as an implicit form of spectral contrastive learning. We prove that, for common model architectures, this simple predictive objective forces the model to sculpt a geometrically structured representation space, implicitly aligning representations with the eigenspectrum of a “predictive similarity” operator. This work offers a powerful new lens to understand how information flows through a model and how the training objective shapes its internal geometry, thereby bridging the gap between learning theory and the practical success of large language models.

[970] A Gap Between the Gaussian RKHS and Neural Networks: An Infinite-Center Asymptotic Analysis

Akash Kumar, Rahul Parhi, Mikhail Belkin

Main category: cs.LG

TL;DR: The paper shows that on unbounded domains, Gaussian RKHS functions can have infinite norm in neural network Banach spaces, revealing a fundamental gap where kernel methods can represent functions that neural networks cannot.

DetailsMotivation: To investigate the functional-space relationship between kernel methods and neural networks on unbounded domains, contrasting with known results on bounded domains where Gaussian RKHS strictly embeds into neural network Banach spaces.

Method: Theoretical analysis establishing that certain functions in the Gaussian reproducing kernel Hilbert space (RKHS) have infinite norm in the neural network Banach space when considered on unbounded domains like ℝᵈ.

Result: Demonstrates a nontrivial gap between kernel methods and neural networks - functions that are easily representable by kernel methods (Gaussian RKHS) cannot be represented by neural networks due to infinite norm in the neural network Banach space.

Conclusion: On unbounded domains, the relationship between kernel methods and neural networks is fundamentally different from bounded domains, with kernel methods having representational capabilities that exceed those of neural networks for certain functions.

Abstract: Recent works have characterized the function-space inductive bias of infinite-width bounded-norm single-hidden-layer neural networks as a kind of bounded-variation-type space. This novel neural network Banach space encompasses many classical multivariate function spaces, including certain Sobolev spaces and the spectral Barron spaces. Notably, this Banach space also includes functions that exhibit less classical regularity, such as those that only vary in a few directions. On bounded domains, it is well-established that the Gaussian reproducing kernel Hilbert space (RKHS) strictly embeds into this Banach space, demonstrating a clear gap between the Gaussian RKHS and the neural network Banach space. It turns out that when investigating these spaces on unbounded domains, e.g., all of $\mathbb{R}^d$, the story is fundamentally different. We establish the following fundamental result: Certain functions that lie in the Gaussian RKHS have infinite norm in the neural network Banach space. This provides a nontrivial gap between kernel methods and neural networks by exhibiting functions that kernel methods easily represent, whereas neural networks cannot.

[971] LLM Assertiveness can be Mechanistically Decomposed into Emotional and Logical Components

Hikaru Tsujimura, Arush Tagade

Main category: cs.LG

TL;DR: Mechanistic analysis of LLM assertiveness reveals two orthogonal components (emotional and logical) that parallel psychological dual-route models, with steering vectors showing distinct causal effects on prediction accuracy.

DetailsMotivation: LLMs often display overconfidence with unwarranted certainty in high-stakes contexts, requiring investigation into the internal mechanisms behind this assertive behavior.

Method: Used open-sourced Llama 3.2 models fine-tuned on human-annotated assertiveness datasets, extracted residual activations across all layers, computed similarity metrics to localize assertive representations, and derived steering vectors from identified components.

Result: Identified layers most sensitive to assertiveness contrasts and revealed that high-assertive representations decompose into emotional and logical clusters. Emotional steering vectors broadly influence prediction accuracy, while logical vectors have more localized effects.

Conclusion: Provides mechanistic evidence for multi-component structure of LLM assertiveness and highlights potential avenues for mitigating overconfident behavior through targeted interventions.

Abstract: Large Language Models (LLMs) often display overconfidence, presenting information with unwarranted certainty in high-stakes contexts. We investigate the internal basis of this behavior via mechanistic interpretability. Using open-sourced Llama 3.2 models fine-tuned on human annotated assertiveness datasets, we extract residual activations across all layers, and compute similarity metrics to localize assertive representations. Our analysis identifies layers most sensitive to assertiveness contrasts and reveals that high-assertive representations decompose into two orthogonal sub-components of emotional and logical clusters-paralleling the dual-route Elaboration Likelihood Model in Psychology. Steering vectors derived from these sub-components show distinct causal effects: emotional vectors broadly influence prediction accuracy, while logical vectors exert more localized effects. These findings provide mechanistic evidence for the multi-component structure of LLM assertiveness and highlight avenues for mitigating overconfident behavior.

[972] Mirage or Method? How Model-Task Alignment Induces Divergent RL Conclusions

Haoze Wu, Cheng Wang, Wenshuo Zhao, Junxian He

Main category: cs.LG

TL;DR: Counterintuitive RL phenomena in LLMs only work when models already have strong task alignment; standard RL remains robust across all settings.

DetailsMotivation: Recent RL advances in LLMs show surprising phenomena like single-example learning and negative-only training, but it's unclear when these actually work versus when they fail.

Method: Systematic examination of counterintuitive claims through rigorous experiments across different model architectures and task domains, measuring Model-Task Alignment via pass@k accuracy.

Result: Counterintuitive results only occur when models already exhibit strong model-task alignment. These techniques fail in challenging regimes where standard RL methods remain effective.

Conclusion: Model-Task Alignment is the key factor determining when counterintuitive RL phenomena work. Standard RL training is consistently robust, while novel techniques are only effective in already-aligned scenarios.

Abstract: Recent advances in applying reinforcement learning (RL) to large language models (LLMs) have led to substantial progress. In particular, a series of remarkable yet often counterintuitive phenomena have been reported in LLMs, exhibiting patterns not typically observed in traditional RL settings. For example, notable claims include that a single training example can match the performance achieved with an entire dataset, that the reward signal does not need to be very accurate, and that training solely with negative samples can match or even surpass sophisticated reward-based methods. However, the precise conditions under which these observations hold - and, critically, when they fail - remain unclear. In this work, we identify a key factor that differentiates RL observations: whether the pretrained model already exhibits strong Model-Task Alignment, as measured by pass@k accuracy on the evaluated task. Through a systematic and comprehensive examination of a series of counterintuitive claims, supported by rigorous experimental validation across different model architectures and task domains, our findings show that while standard RL training remains consistently robust across settings, many of these counterintuitive results arise only when the model and task already exhibit strong model-task alignment. In contrast, these techniques fail to drive substantial learning in more challenging regimes, where standard RL methods remain effective.

[973] To See a World in a Spark of Neuron: Disentangling Multi-task Interference for Training-free Model Merging

Zitao Fang, Guodong DU, Shuyang Yu, Yifei Guo, Yiwei Zhang, Yiyao Cao, Jing Li, Ho-Kin Tang, Sim Kuan Goh

Main category: cs.LG

TL;DR: NeuroMerging is a novel model merging framework that addresses task interference by decomposing representations into neuronal subspaces for input sensitivity and task adaptability, enabling training-free fusion across diverse tasks with superior performance.

DetailsMotivation: Fine-tuning pre-trained models improves task-specific performance but harms generalization. Model merging techniques suffer from task interference due to overlooking neuronal mechanisms, connectivity, and activation patterns.

Method: Decomposed task-specific representations into two complementary neuronal subspaces (input sensitivity and task adaptability). Developed NeuroMerging framework to mitigate task interference within these neuronal subspaces for training-free model fusion.

Result: Achieved superior performance compared to existing methods on multi-task benchmarks across both natural language and vision domains through extensive experiments.

Conclusion: Aligning neuronal mechanisms is crucial for effective model merging. NeuroMerging offers new insights into mitigating task interference and improving knowledge fusion without requiring additional training.

Abstract: Fine-tuning pre-trained models on targeted datasets enhances task-specific performance but often comes at the expense of generalization. Model merging techniques, which integrate multiple fine-tuned models into a single multi-task model through task arithmetic, offer a promising solution. However, task interference remains a fundamental challenge, leading to performance degradation and suboptimal merged models. Existing approaches largely overlooked the fundamental roles of neurons, their connectivity, and activation, resulting in a merging process and a merged model that does not consider how neurons relay and process information. In this work, we present the first study that relies on neuronal mechanisms for model merging. Specifically, we decomposed task-specific representations into two complementary neuronal subspaces that regulate input sensitivity and task adaptability. Leveraging this decomposition, we introduced NeuroMerging, a novel merging framework developed to mitigate task interference within neuronal subspaces, enabling training-free model fusion across diverse tasks. Through extensive experiments, we demonstrated that NeuroMerging achieved superior performance compared to existing methods on multi-task benchmarks across both natural language and vision domains. Our findings highlighted the importance of aligning neuronal mechanisms in model merging, offering new insights into mitigating task interference and improving knowledge fusion. Our project is available at https://ZzzitaoFang.github.io/projects/NeuroMerging/.

[974] A Rollout-Based Algorithm and Reward Function for Resource Allocation in Business Processes

Jeroen Middelhuis, Zaharah Bukhsh, Ivo Adan, Remco Dijkman

Main category: cs.LG

TL;DR: Proposed rollout-based DRL algorithm with direct reward decomposition for business process resource allocation, eliminating need for reward engineering and achieving optimal policies.

DetailsMotivation: Existing DRL methods are unsuitable for dynamic business process environments and rely on engineered reward functions that may misalign with objectives, leading to suboptimal policies.

Method: Rollout-based DRL algorithm that iteratively improves policy by evaluating execution trajectories, with reward function directly decomposing cycle time minimization objective.

Result: Achieved optimal policy in six test scenarios and outperformed/matched best heuristics on realistically sized business process models.

Conclusion: The proposed approach successfully addresses reward misalignment issues and demonstrates effectiveness in optimizing resource allocation for business processes.

Abstract: Resource allocation plays a critical role in minimizing cycle time and improving the efficiency of business processes. Recently, Deep Reinforcement Learning (DRL) has emerged as a powerful technique to optimize resource allocation policies in business processes. In the DRL framework, an agent learns a policy through interaction with the environment, guided solely by reward signals that indicate the quality of its decisions. However, existing algorithms are not suitable for dynamic environments such as business processes. Furthermore, existing DRL-based methods rely on engineered reward functions that approximate the desired objective, but a misalignment between reward and objective can lead to undesired decisions or suboptimal policies. To address these issues, we propose a rollout-based DRL algorithm and a reward function to optimize the objective directly. Our algorithm iteratively improves the policy by evaluating execution trajectories following different actions. Our reward function directly decomposes the objective function of minimizing the cycle time, such that trial-and-error reward engineering becomes unnecessary. We evaluated our method in six scenarios, for which the optimal policy can be computed, and on a set of increasingly complex, realistically sized process models. The results show that our algorithm can learn the optimal policy for the scenarios and outperform or match the best heuristics on the realistically sized business processes.

[975] Tilus: A Tile-Level GPGPU Programming Language for Low-Precision Computation

Yaoyao Ding, Bohan Hou, Xiao Zhang, Allan Lin, Tianqi Chen, Cody Yu Hao, Yida Wang, Gennady Pekhimenko

Main category: cs.LG

TL;DR: Tilus is a new domain-specific language for GPU computing that supports arbitrary low-precision data types (1-8 bits) and outperforms existing solutions by up to 2.61x.

DetailsMotivation: Existing low-precision computation approaches for LLM serving are limited to power-of-two bit widths and suffer from suboptimal performance due to high-level GPU programming abstractions that restrict critical optimizations.

Method: Tilus introduces a domain-specific language with thread-block-level programming model, hierarchical memory space, novel algebraic layout system, and support for diverse low-precision data types. It compiles to efficient GPU programs through automatic vectorization and instruction selection.

Result: Tilus achieves performance improvements of 1.75x over Triton, 2.61x over Ladder, 1.29x over QuantLLM, and 1.03x over Marlin, demonstrating efficient support for full spectrum of low-precision data types.

Conclusion: Tilus successfully addresses limitations of existing approaches by providing a specialized language that enables fine-grained optimizations for arbitrary low-precision computations, significantly improving LLM serving efficiency.

Abstract: Serving Large Language Models (LLMs) is critical for AI-powered applications, yet it demands substantial computational resources, particularly in memory bandwidth and computational throughput. Low-precision computation has emerged as a key technique to improve efficiency while reducing resource consumption. Existing approaches for generating low-precision kernels are limited to weight bit widths that are powers of two and suffer from suboptimal performance because of high-level GPU programming abstractions. These abstractions restrict critical optimizations, such as fine-grained register management and optimized memory access patterns, that are essential for efficient low-precision computations. In this paper, we introduce Tilus, a domain-specific language designed for General-Purpose GPU (GPGPU) computing that supports low-precision data types with arbitrary bit widths from 1 to 8 while maintaining GPU programmability. Tilus features a thread-block-level programming model, a hierarchical memory space, a novel algebraic layout system, and extensive support for diverse low-precision data types. Tilus programs are compiled into highly efficient GPU programs through automatic vectorization and instruction selection. Extensive experiments demonstrate that Tilus efficiently supports a full spectrum of low-precision data types, and outperforms state-of-the-art low-precision kernels. Compared to existing compilers such as Triton and Ladder, as well as hand-optimized kernels such as QuantLLM and Marlin, Tilus achieves performance improvements of: $1.75\times$, $2.61\times$, $1.29\times$ and $1.03\times$, respectively. We open-source Tilus at https://github.com/NVIDIA/tilus.

[976] InDiD: Instant Disorder Detection via Representation Learning

Evgenia Romanenkova, Alexander Stepikin, Matvey Morozov, Alexey Zaytsev

Main category: cs.LG

TL;DR: Proposes a differentiable loss function for change point detection that balances detection delay and false alarms, enabling representation learning in deep models for semi-structured sequential data like videos and sensor streams.

DetailsMotivation: Classic change point detection approaches underperform on semi-structured sequential data because they cannot process its structure without proper representation, and there's a need to detect disorders as fast as possible.

Method: A principled loss function that approximates classic rigorous solutions but is differentiable, allowing representation learning for deep models. Applied to synthetic sequences, real-world sensor data, and carefully labeled video data with change points.

Result: Outperforms baselines significantly - for explosion detection in video, achieves F1 score of 0.53 compared to baseline scores of 0.31 and 0.35. Complex data requires meaningful representations tailored for CPD task.

Conclusion: The proposed approach provides effective representations for change point detection in semi-structured sequential data, demonstrating superior performance over existing baselines across different data types including video surveillance.

Abstract: For sequential data, a change point is a moment of abrupt regime switch in data streams. Such changes appear in different scenarios, including simpler data from sensors and more challenging video surveillance data. We need to detect disorders as fast as possible. Classic approaches for change point detection (CPD) might underperform for semi-structured sequential data because they cannot process its structure without a proper representation. We propose a principled loss function that balances change detection delay and time to a false alarm. It approximates classic rigorous solutions but is differentiable and allows representation learning for deep models. We consider synthetic sequences, real-world data sensors and videos with change points. We carefully labelled available data with change point moments for video data and released it for the first time. Experiments suggest that complex data require meaningful representations tailored for the specificity of the CPD task – and our approach provides them outperforming considered baselines. For example, for explosion detection in video, the F1 score for our method is 0.53 compared to baseline scores of 0.31 and 0.35.

[977] Memory-adaptive Depth-wise Heterogeneous Federated Learning

Kai Zhang, Yutong Dai, Hongyi Wang, Eric Xing, Xun Chen, Lichao Sun

Main category: cs.LG

TL;DR: FeDepth introduces memory-adaptive depth-wise learning for federated learning, decomposing models into blocks based on client memory budgets and training sequentially, achieving significant accuracy improvements over width-slimming approaches.

DetailsMotivation: Address performance degradation in federated learning caused by heterogeneous device memory limitations, where current width-slimming techniques negatively impact global model performance during aggregation.

Method: Adaptively decomposes full model into blocks according to each client’s memory budget, then trains blocks sequentially to obtain a full inference model (depth-wise learning approach).

Result: Outperforms state-of-the-art approaches with 5%+ improvement on CIFAR-10 and over 10% improvement on CIFAR-100 top-1 accuracy, and demonstrates effectiveness on ViT through depth-wise fine-tuning.

Conclusion: Depth-wise training strategy is highly effective for improving global model performance in federated learning with heterogeneous devices, highlighting the importance of memory-aware techniques.

Abstract: Federated learning is a promising paradigm that allows multiple clients to collaboratively train a model without sharing the local data. However, the presence of heterogeneous devices in federated learning, such as mobile phones and IoT devices with varying memory capabilities, would limit the scale and hence the performance of the model could be trained. The mainstream approaches to address memory limitations focus on width-slimming techniques, where different clients train subnetworks with reduced widths locally and then the server aggregates the subnetworks. The global model produced from these methods suffers from performance degradation due to the negative impact of the actions taken to handle the varying subnetwork widths in the aggregation phase. In this paper, we introduce a memory-adaptive depth-wise learning solution in FL called FeDepth, which adaptively decomposes the full model into blocks according to the memory budgets of each client and trains blocks sequentially to obtain a full inference model. Our method outperforms state-of-the-art approaches, achieving 5% and more than 10% improvements in top-1 accuracy on CIFAR-10 and CIFAR-100, respectively. We also demonstrate the effectiveness of depth-wise fine-tuning on ViT. Our findings highlight the importance of memory-aware techniques for federated learning with heterogeneous devices and the success of depth-wise training strategy in improving the global model’s performance.

[978] FairPO: Robust Preference Optimization for Fair Multi-Label Learning

Soumen Kumar Mondal, Akshit Varmora, Prateek Chanda, Ganesh Ramakrishnan

Main category: cs.LG

TL;DR: FairPO is a fairness framework for multi-label classification that uses preference optimization to reduce bias by prioritizing underperforming label groups while maintaining overall performance.

DetailsMotivation: To address fairness issues in multi-label classification where certain label groups may be systematically underperforming or biased, ensuring equitable treatment across diverse label categories.

Method: Partitions labels into privileged/non-privileged groups, employs DPO-inspired preference-based loss to differentiate true positives from negatives in privileged groups, uses robust optimization to dynamically adjust training emphasis toward poorer-performing groups.

Result: The framework aims to mitigate bias while preserving baseline classification performance for non-privileged labels, promoting fairer treatment across all label categories.

Conclusion: FairPO provides an effective approach to fairness in multi-label classification, with plans to extend to alternative loss formulations (SimPO, CPO) and multilabel generation capabilities for ambiguous inputs.

Abstract: We propose FairPO, a novel framework designed to promote fairness in multi-label classification by directly optimizing preference signals with a group robustness perspective. In our framework, the set of labels is partitioned into privileged and non-privileged groups, and a preference-based loss inspired by Direct Preference Optimization (DPO) is employed to more effectively differentiate true positive labels from confusing negatives within the privileged group, while preserving baseline classification performance for non-privileged labels. By framing the learning problem as a robust optimization over groups, our approach dynamically adjusts the training emphasis toward groups with poorer performance, thereby mitigating bias and ensuring a fairer treatment across diverse label categories. In addition, we outline plans to extend this approach by investigating alternative loss formulations such as Simple Preference Optimisation (SimPO) and Contrastive Preference Optimization (CPO) to exploit reference-free reward formulations and contrastive training signals. Furthermore, we plan to extend FairPO with multilabel generation capabilities, enabling the model to dynamically generate diverse and coherent label sets for ambiguous inputs.

[979] Knowledge-integrated AutoEncoder Model

Teddy Lazebnik, Liron Simon-Keren

Main category: cs.LG

TL;DR: KiAE integrates external knowledge into AutoEncoders to create controllable embedding spaces with desired properties for downstream tasks, outperforming 9 existing models in reconstruction accuracy.

DetailsMotivation: Traditional AutoEncoders produce black-box embedding spaces that lack controllability and may not possess properties needed for downstream tasks, limiting their effectiveness in data encoding.

Method: Proposes Knowledge-integrated AutoEncoder (KiAE) that leverages domain-specific information to preserve desired distance and neighborhood properties in the embedding space during learning.

Result: Evaluated on three large-scale datasets from different scientific fields, KiAE outperformed nine existing encoding models in reconstruction accuracy and effectively captured underlying structures and relationships.

Conclusion: KiAE successfully integrates external knowledge to generate more useful representations, demonstrating superior performance over traditional AutoEncoder approaches while providing controllable embedding spaces.

Abstract: Data encoding is a common and central operation in most data analysis tasks. The performance of other models downstream in the computational process highly depends on the quality of data encoding. One of the most powerful ways to encode data is using the neural network AutoEncoder (AE) architecture. However, the developers of AE cannot easily influence the produced embedding space, as it is usually treated as a black box technique. This means the embedding space is uncontrollable and does not necessarily possess the properties desired for downstream tasks. This paper introduces a novel approach for developing AE models that can integrate external knowledge sources into the learning process, possibly leading to more accurate results. The proposed Knowledge-integrated AutoEncoder (KiAE) model can leverage domain-specific information to make sure the desired distance and neighborhood properties between samples are preservative in the embedding space. The proposed model is evaluated on three large-scale datasets from three scientific fields and is compared to nine existing encoding models. The results demonstrate that the KiAE model effectively captures the underlying structures and relationships between the input data and external knowledge, meaning it generates a more useful representation. This leads to outperforming the rest of the models in terms of reconstruction accuracy.

[980] ORBIT-2: Scaling Exascale Vision Foundation Models for Weather and Climate Downscaling

Xiao Wang, Jong-Youl Choi, Takuya Kurihaya, Isaac Lyngaas, Hong-Jun Yoon, Xi Xiao, David Pugmire, Ming Fan, Nasik M. Nafi, Aristeidis Tsaris, Ashwin M. Aji, Maliha Hossain, Mohamed Wahib, Dali Wang, Peter Thornton, Prasanna Balaprakash, Moetasim Ashfaq, Dan Lu

Main category: cs.LG

TL;DR: ORBIT-2 is a scalable foundation model for global hyper-resolution climate downscaling that addresses computational limitations of Vision Transformers through two innovations: Reslim architecture and TILES algorithm, achieving linear complexity and massive parallel processing.

DetailsMotivation: Sparse observations and coarse-resolution climate models limit regional decision-making, while existing AI methods struggle with generalization across variables/geographies and face quadratic complexity constraints of Vision Transformer self-attention.

Method: ORBIT-2 incorporates two key innovations: (1) Residual Slim ViT (Reslim) - lightweight architecture with residual learning and Bayesian regularization; (2) TILES - tile-wise sequence scaling algorithm that reduces self-attention complexity from quadratic to linear.

Result: ORBIT-2 scales to 10 billion parameters across 65,536 GPUs, achieving 4.1 exaFLOPS throughput and 74-98% strong scaling efficiency. It supports 0.9 km global resolution and processes 4.2 billion tokens. Achieves R² scores of 0.98-0.99 on 7 km benchmarks.

Conclusion: ORBIT-2 demonstrates breakthrough scalability and efficiency for climate downscaling, enabling hyper-resolution global modeling with linear computational complexity and high accuracy against observational data.

Abstract: Sparse observations and coarse-resolution climate models limit effective regional decision-making, underscoring the need for robust downscaling. However, existing AI methods struggle with generalization across variables and geographies and are constrained by the quadratic complexity of Vision Transformer (ViT) self-attention. We introduce ORBIT-2, a scalable foundation model for global, hyper-resolution climate downscaling. ORBIT-2 incorporates two key innovations: (1) Residual Slim ViT (Reslim), a lightweight architecture with residual learning and Bayesian regularization for efficient, robust prediction; and (2) TILES, a tile-wise sequence scaling algorithm that reduces self-attention complexity from quadratic to linear, enabling long-sequence processing and massive parallelism. ORBIT-2 scales to 10 billion parameters across 65,536 GPUs, achieving up to 4.1 exaFLOPS sustained throughput and 74–98% strong scaling efficiency. It supports downscaling to 0.9 km global resolution and processes sequences up to 4.2 billion tokens. On 7 km resolution benchmarks, ORBIT-2 achieves high accuracy with $R^2$ scores in the range of 0.98–0.99 against observational data.

[981] Sampling, Diffusions, and Stochastic Localization

Andrea Montanari

Main category: cs.LG

TL;DR: This paper connects diffusion sampling methods with stochastic localization, providing a unified framework for understanding and generalizing sampling algorithms in high-dimensional distributions.

DetailsMotivation: To bridge the gap between diffusion-based sampling techniques and stochastic localization methods, which have been developed separately but share fundamental connections in high-dimensional sampling problems.

Method: The authors generalize algorithmic stochastic localization processes and establish clear connections between diffusion methods and stochastic localization, creating a unified theoretical framework.

Result: A comprehensive framework that unifies various sampling schemes, provides new insights into their relationships, and offers a simple yet broadly applicable construction method.

Conclusion: The unified viewpoint connecting diffusions and stochastic localization offers valuable insights for sampling algorithms and enables generalization of known sampling schemes in a coherent framework.

Abstract: Diffusions are a successful technique to sample from high-dimensional distributions. The target distribution can be either explicitly given or learnt from a collection of samples. They implement a diffusion process whose endpoint is a sample from the target distribution. The drift of the diffusion process is typically represented as a neural network. Stochastic localization is a successful technique to prove mixing of Markov Chains and other functional inequalities in high dimension. An algorithmic version of stochastic localization was recently proposed in order to sample from certain statistical mechanics models. This expository article has three objectives: $(i)$~Generalize the algorithmic construction to other stochastic localization processes. This construction is both simple and broadly applicable; $(ii)$~Clarify the connection between diffusions and stochastic localization. This allows to derive several known sampling schemes in a unified fashion; $(iii)$~Describe the insights that follow from this unified viewpoint.

[982] Mask-PINNs: Mitigating Internal Covariate Shift in Physics-Informed Neural Networks

Feilong Jiang, Xiaonan Hou, Jianqiao Ye, Min Xia

Main category: cs.LG

TL;DR: Mask-PINNs address internal covariate shift in Physics-Informed Neural Networks using learnable masks that regulate feature distributions while preserving physical constraints, improving accuracy, stability, and enabling wider networks.

DetailsMotivation: Internal covariate shift disrupts feature distributions and limits model expressiveness in PINNs, and conventional normalization methods like Batch/Layer Normalization distort physical consistency required for reliable PDE solutions.

Method: Proposed Mask-PINNs architecture with a learnable mask function that regulates feature distributions through a modulation mechanism while preserving physical constraints. Theoretical analysis shows the mask suppresses feature representation expansion.

Result: Validated on multiple PDE benchmarks (convection, wave propagation, Helmholtz equations) across diverse activation functions. Shows consistent improvements in prediction accuracy, convergence stability, and robustness. Enables effective use of wider networks.

Conclusion: Mask-PINNs overcome a key limitation in existing PINN frameworks by addressing internal covariate shift without compromising physical consistency, leading to more stable and effective training with improved performance across various PDE problems.

Abstract: Physics-Informed Neural Networks (PINNs) have emerged as a powerful framework for solving partial differential equations (PDEs) by embedding physical laws directly into the loss function. However, as a fundamental optimization issue, internal covariate shift (ICS) hinders the stable and effective training of PINNs by disrupting feature distributions and limiting model expressiveness. Unlike standard deep learning tasks, conventional remedies for ICS – such as Batch Normalization and Layer Normalization – are not directly applicable to PINNs, as they distort the physical consistency required for reliable PDE solutions. To address this issue, we propose Mask-PINNs, a novel architecture that introduces a learnable mask function to regulate feature distributions while preserving the underlying physical constraints of PINNs. We provide a theoretical analysis showing that the mask suppresses the expansion of feature representations through a carefully designed modulation mechanism. Empirically, we validate the method on multiple PDE benchmarks – including convection, wave propagation, and Helmholtz equations – across diverse activation functions. Our results show consistent improvements in prediction accuracy, convergence stability, and robustness. Furthermore, we demonstrate that Mask-PINNs enable the effective use of wider networks, overcoming a key limitation in existing PINN frameworks.

[983] K-Tensors: Clustering Positive Semi-Definite Matrices

Hanchao Zhang, Xiaomeng Ju, Baoyi Shi, Lingsong Meng, Thaddeus Tarpey

Main category: cs.LG

TL;DR: K-Tensors is a novel clustering algorithm for symmetric positive semi-definite matrices that preserves geometric and spectral information using a specialized divergence, with applications in brain connectivity analysis.

DetailsMotivation: Conventional clustering methods for SPSD matrices lose critical geometric and spectral information through vectorization or transformations, necessitating an approach that respects the intrinsic geometry of these matrices.

Method: The K-Tensors algorithm introduces a divergence that preserves shape and eigenstructure information, identifies structured subsets with common principal component representations, and yields principal SPSD tensors as representative matrices.

Result: The algorithm is shown to be self-consistent under mild distribution assumptions and converges to a local optimum. Applied to rs-fMRI data, it successfully clusters brain connectivity matrices to discover subject groups with shared connectivity structures.

Conclusion: K-Tensors provides an effective clustering approach for SPSD matrices that maintains their geometric properties, with demonstrated practical utility in neuroimaging applications for identifying meaningful patterns in brain connectivity data.

Abstract: This paper presents a new clustering algorithm for symmetric positive semi-definite (SPSD) matrices, called K-Tensors. The method identifies structured subsets of the SPSD cone characterized by common principal component (CPC) representations, where each subset corresponds to matrices sharing a common eigenstructure. Unlike conventional clustering approaches that rely on vectorization or transformations of SPSD matrices, thereby losing critical geometric and spectral information, K-Tensors introduces a divergence that respects the intrinsic geometry of SPSD matrices. This divergence preserves the shape and eigenstructure information and yields principal SPSD tensors, defined as a set of representative matrices that summarize the distribution of SPSD matrices. By exploring its theoretical properties, we show that the proposed clustering algorithm is self-consistent under mild distribution assumptions and converges to a local optimum. We demonstrate the use of the algorithm through an application to resting-state functional magnetic resonance imaging (rs-fMRI) data from the Human Connectome Project, where we cluster brain connectivity matrices to discover groups of subjects with shared connectivity structures.

[984] GEN: A Practical Alternative to Graph Transformers for Long-Range Graph Modeling

Shuo Wang, Ge Cheng, Yun Zhang

Main category: cs.LG

TL;DR: Graph Elimination Networks (GENs) combine edge-wise and hop-wise self-attention to achieve Graph Transformer-like long-range modeling while maintaining MPNN efficiency, outperforming baselines on long-range graph tasks.

DetailsMotivation: MPNNs struggle with long-distance information propagation, while Graph Transformers have quadratic computational costs that limit scalability. There's a need for efficient long-range modeling in graph neural networks.

Method: GENs use parallel edge-wise and hop-wise self-attention with multiplicative composition. The Graph Elimination Algorithm prevents double counting across hops and extracts disentangled multi-hop features for hop-wise attention within a bounded K-hop receptive field.

Result: On LRGB benchmark, GENs outperform MPNN baselines by 7.7pp on PascalVOC-SP and 6.0pp on COCO-SP, achieving performance comparable to state-of-the-art Graph Transformers. On OGBN-Products, GENs support full-batch training while sparse-attention baselines struggle with memory limits.

Conclusion: GENs provide an efficient and practical alternative to Graph Transformers for large, sparse graphs, achieving strong long-range modeling capabilities while maintaining computational efficiency.

Abstract: Message Passing Neural Networks (MPNNs) model local relations effectively but struggle to propagate information over long distances. Graph Transformers (GTs) mitigate this via global self-attention, yet their quadratic cost in the number of nodes limits scalability. We propose Graph Elimination Networks (GENs), an MPNN variant that approximates GT-like long-range modeling while maintaining high efficiency. GENs combine edge-wise and hop-wise self-attention in parallel; their multiplicative composition yields an attention kernel separable across edge and hop factors within a bounded K-hop receptive field. To enable hop-wise attention, we introduce the Graph Elimination Algorithm (GEA), which prevents double counting across hops, ensuring that each round injects the k-hop incremental contribution exactly once. Taking differences between successive rounds recovers the k-hop increment and yields disentangled multi-hop features as inputs for hop-wise attention. This preserves clearer structural distinctions across hop distances and enables more faithful modeling of pairwise dependencies between distant nodes within the K-hop neighborhood. On the Long-Range Graph Benchmark (LRGB), GENs outperform strong MPNN baselines by 7.7 and 6.0 percentage points (pp) on PascalVOC-SP and COCO-SP, and achieve performance on par with or better than state-of-the-art Graph Transformers. On OGBN-Products, GENs support full-batch training/inference, while sparse-attention baselines like Exphormer struggle with memory limits under comparable budgets, highlighting GENs as a practical alternative for large, sparse graphs.

[985] Deep Transductive Outlier Detection

Simon Klüttermann, Emmanuel Müller

Main category: cs.LG

TL;DR: Doust is the first end-to-end transductive deep learning algorithm for outlier detection that leverages unlabeled test data to achieve significant performance improvements, achieving 89% ROC-AUC on ADBench benchmark and outperforming 21 competitors by ~10%.

DetailsMotivation: Outlier detection is a core ML challenge, and while transductive learning has shown promise in related tasks, it remains largely unexplored for modern outlier detection. The paper aims to explore how leveraging test data during training can boost outlier detection accuracy.

Method: Developed Doust - an end-to-end transductive deep learning algorithm that explicitly leverages unlabeled test data during training to improve outlier detection performance.

Result: Achieved 89% average ROC-AUC on the comprehensive ADBench benchmark, outperforming all 21 competitors by approximately 10%. Analysis revealed substantial performance gains in favorable conditions but identified limitations with very low contamination rates unless datasets are sufficiently large.

Conclusion: Transductive learning shows significant potential for outlier detection with substantial performance improvements, though very low contamination rates can limit gains unless datasets are large enough. Doust establishes transductive learning as a promising approach for modern outlier detection.

Abstract: Outlier detection (OD) is one of the core challenges in machine learning. Transductive learning, which leverages test data during training, has shown promise in related machine learning tasks, yet remains largely unexplored for modern OD. We present Doust, the first end-to-end transductive deep learning algorithm for outlier detection, which explicitly leverages unlabeled test data to boost accuracy. On the comprehensive ADBench benchmark, Doust achieves an average ROC-AUC of $89%$, outperforming all 21 competitors by roughly $10%$. Our analysis identifies both the potential and a limitation of transductive OD: while performance gains can be substantial in favorable conditions, very low contamination rates can hinder improvements unless the dataset is sufficiently large.

[986] Introducing ‘Inside’ Out of Distribution

Teddy Lazebnik

Main category: cs.LG

TL;DR: Paper introduces inside vs outside OOD distinction, shows outside OOD causes greater performance degradation than inside OOD on synthetic datasets using RMSE and F1 metrics.

DetailsMotivation: Current OOD studies focus only on extrapolatory (outside) OOD, neglecting interpolatory (inside) OOD cases, creating a gap in understanding comprehensive OOD detection.

Method: Used synthetically-generated datasets with both inside and outside OOD, analyzed using normalized Root Mean Squared Error (RMSE) and F1 score as performance metrics.

Result: Different inside-outside OOD profiles have unique effects on ML performance, with outside OOD generally causing greater performance degradation on average.

Conclusion: Distinguishing between inside and outside OOD is crucial for developing effective counter-OOD methods in machine learning.

Abstract: Detecting and understanding out-of-distribution (OOD) samples is crucial in machine learning (ML) to ensure reliable model performance. Current OOD studies primarily focus on extrapolatory (outside) OOD, neglecting potential cases of interpolatory (inside) OOD. In this study, we introduce a novel perspective on OOD by suggesting it can be divided into inside and outside cases. We examine the inside-outside OOD profiles of datasets and their impact on ML model performance, using normalized Root Mean Squared Error (RMSE) and F1 score as the performance metrics on syntetically-generated datasets with both inside and outside OOD. Our analysis demonstrates that different inside-outside OOD profiles lead to unique effects on ML model performance, with outside OOD generally causing greater performance degradation, on average. These findings highlight the importance of distinguishing between inside and outside OOD for developing effective counter-OOD methods.

[987] Optimal Parallelization of Boosting

Arthur da Cunha, Mikael Møller Høgsgaard, Kasper Green Larsen

Main category: cs.LG

TL;DR: This paper closes the gap between theoretical lower bounds and practical performance in parallel Boosting algorithms, providing improved lower bounds and a matching algorithm that achieves near-optimal parallel complexity across the entire tradeoff spectrum.

DetailsMotivation: There exists a significant gap between theoretical lower bounds and actual performance of parallel Boosting algorithms in the tradeoff between training rounds and parallel work per round, which needs to be addressed.

Method: The authors provide improved lower bounds on parallel complexity of weak-to-strong learners and develop a parallel Boosting algorithm that matches these bounds across the entire p vs. t tradeoff spectrum.

Result: The work essentially closes the persistent gap between theoretical lower bounds and algorithm performance, achieving matching performance up to logarithmic factors throughout the tradeoff space.

Conclusion: This research settles the true parallel complexity of nearly sample-optimal Boosting algorithms, providing both optimal theoretical bounds and practical algorithms that achieve them.

Abstract: Recent works on the parallel complexity of Boosting have established strong lower bounds on the tradeoff between the number of training rounds $p$ and the total parallel work per round $t$. These works have also presented highly non-trivial parallel algorithms that shed light on different regions of this tradeoff. Despite these advancements, a significant gap persists between the theoretical lower bounds and the performance of these algorithms across much of the tradeoff space. In this work, we essentially close this gap by providing both improved lower bounds on the parallel complexity of weak-to-strong learners, and a parallel Boosting algorithm whose performance matches these bounds across the entire $p$ vs.~$t$ compromise spectrum, up to logarithmic factors. Ultimately, this work settles the true parallel complexity of Boosting algorithms that are nearly sample-optimal.

[988] Are Hourly PM2.5 Forecasts Sufficiently Accurate to Plan Your Day? Individual Decision Making in the Face of Increasing Wildfire Smoke

Renato Berlinghieri, David R. Burt, Paolo Giani, Arlene M. Fiore, Tamara Broderick

Main category: cs.LG

TL;DR: Evaluation of six PM2.5 forecasting methods during 2023 US wildfire season shows room for improvement in helping individuals reduce air pollution exposure through better hourly forecasts.

DetailsMotivation: Increasing wildfire frequency due to climate change creates health risks from air pollution, and reliable hourly air quality forecasts could help people plan activities to reduce exposure, similar to how weather forecasts are used.

Method: Evaluated six existing PM2.5 forecasting methods (physical simulation, ensembling, and AI) across continental US during 2023 fire season, focusing on decision-making scenarios like whether/when to go outside. Used visualizations and developed new evaluation metrics.

Result: Found meaningful room for improvement in PM2.5 forecasting accuracy and usefulness for individual decision-making regarding pollution exposure avoidance.

Conclusion: PM2.5 forecasting could be improved through better physical models, incorporating more data sources, and leveraging artificial intelligence tools to provide more reliable hourly forecasts for public health protection.

Abstract: Wildfire frequency is increasing as the climate changes, and the resulting air pollution poses health risks. Just as people routinely use hourly weather forecasts to plan their day’s activities around precipitation, reliable hourly air quality forecasts could help individuals reduce their exposure to air pollution. In the present work, we evaluate six existing forecasts of ground-level fine particulate matter (PM2.5) within the continental United States during the 2023 fire season. We include forecasts using physical simulation, ensembling, and artificial intelligence. We focus our evaluation on individual decisions, such as (1) whether to go outside on a day with potentially high PM2.5 or (2) when to go outside for the lowest PM2.5 exposure. Our evaluation consists of both visualizations of hourly PM2.5 forecasts in particular locations as well as metrics summarizing forecast skill for the two tasks above. As part of our analysis, we introduce a new evaluation metric for the task of deciding when to go outside. We find meaningful room for improvement in PM2.5 forecasting, which might be realized by improving physical models, incorporating more data sources, and using artificial intelligence tools.

[989] Adaptive and oblivious statistical adversaries are equivalent

Guy Blanc, Gregory Valiant

Main category: cs.LG

TL;DR: Sample-adaptive and sample-oblivious adversaries are equivalent for statistical tasks under corruption, differing only by polynomial factors in sample size.

DetailsMotivation: To resolve the fundamental question about whether sample-adaptive adversaries (who know the sample contents) are more powerful than sample-oblivious adversaries in corrupting statistical learning tasks.

Method: Prove equivalence by constructing algorithm A’ that uses a polynomially larger sample and runs original algorithm A on a uniformly random subsample, maintaining computational efficiency.

Result: For all types of corruptions, sample-adaptive and sample-oblivious adversaries are equivalent up to polynomial factors in sample size.

Conclusion: The main open question from previous works is resolved - adaptive and oblivious adversaries have equivalent corruption power in statistical tasks.

Abstract: We resolve a fundamental question about the ability to perform a statistical task, such as learning, when an adversary corrupts the sample. Such adversaries are specified by the types of corruption they can make and their level of knowledge about the sample. The latter distinguishes between sample-adaptive adversaries which know the contents of the sample when choosing the corruption, and sample-oblivious adversaries, which do not. We prove that for all types of corruptions, sample-adaptive and sample-oblivious adversaries are \emph{equivalent} up to polynomial factors in the sample size. This resolves the main open question introduced by [BLMT22] and further explored in [CHL+23]. Specifically, consider any algorithm $A$ that solves a statistical task even when a sample-oblivious adversary corrupts its input. We show that there is an algorithm $A’$ that solves the same task when the corresponding sample-adaptive adversary corrupts its input. The construction of $A’$ is simple and maintains the computational efficiency of $A$: It requests a polynomially larger sample than $A$ uses and then runs $A$ on a uniformly random subsample.

[990] FIT-GNN: Faster Inference Time for GNNs that ‘FIT’ in Memory Using Coarsening

Shubhajit Roy, Hrriday Ruparel, Kishan Ved, Anirban Dasgupta

Main category: cs.LG

TL;DR: Novel graph coarsening approach for GNN scalability that reduces inference computational costs and memory usage while maintaining competitive performance.

DetailsMotivation: Existing GNN scalability methods focus on training phase but neglect inference computational costs, creating bottlenecks for real-world deployment especially on low-resource devices.

Method: Proposes two graph coarsening methods (Extra Nodes and Cluster Nodes) to create smaller graphs for efficient inference, extending application to graph-level tasks including classification and regression.

Result: Achieves orders of magnitude improvement in single-node inference time, significantly reduces memory consumption, enables efficient training/inference on low-resource devices where conventional methods fail.

Conclusion: Graph coarsening effectively addresses GNN scalability challenges during inference phase, providing substantial computational benefits while maintaining competitive model performance across various graph tasks.

Abstract: Scalability of Graph Neural Networks (GNNs) remains a significant challenge. To tackle this, methods like coarsening, condensation, and computation trees are used to train on a smaller graph, resulting in faster computation. Nonetheless, prior research has not adequately addressed the computational costs during the inference phase. This paper presents a novel approach to improve the scalability of GNNs by reducing computational burden during the inference phase using graph coarsening. We demonstrate two different methods – Extra Nodes and Cluster Nodes. Our study extends the application of graph coarsening for graph-level tasks, including graph classification and graph regression. We conduct extensive experiments on multiple benchmark datasets to evaluate the performance of our approach. Our results show that the proposed method achieves orders of magnitude improvements in single-node inference time compared to traditional approaches. Furthermore, it significantly reduces memory consumption for node and graph classification and regression tasks, enabling efficient training and inference on low-resource devices where conventional methods are impractical. Notably, these computational advantages are achieved while maintaining competitive performance relative to baseline models.

[991] A theoretical framework for self-supervised contrastive learning for continuous dependent data

Alexander Marusov, Aleksandr Yugay, Alexey Zaytsev

Main category: cs.LG

TL;DR: Novel theoretical framework for contrastive self-supervised learning that handles continuous dependent data with semantic closeness between samples, outperforming existing methods on temporal and spatio-temporal benchmarks.

DetailsMotivation: Traditional contrastive SSL methods assume semantic independence between samples, which doesn't hold for dependent data like temporal and spatio-temporal domains that exhibit complex correlations.

Method: Proposed a framework with hard and soft ground truth similarity measures, derived analytical form for estimated similarity matrix, and introduced dependency-aware loss functions for dependent data.

Result: Outperformed TS2Vec on UEA and UCR benchmarks with 4.17% and 2.08% accuracy improvements respectively, and achieved 7% higher ROC-AUC on drought classification with complex spatio-temporal patterns.

Conclusion: The theoretically grounded dependency-aware loss functions effectively capture spatio-temporal dependencies, making the approach superior for dependent data SSL applications.

Abstract: Self-supervised learning (SSL) has emerged as a powerful approach to learning representations, particularly in the field of computer vision. However, its application to dependent data, such as temporal and spatio-temporal domains, remains underexplored. Besides, traditional contrastive SSL methods often assume \emph{semantic independence between samples}, which does not hold for dependent data exhibiting complex correlations. We propose a novel theoretical framework for contrastive SSL tailored to \emph{continuous dependent data}, which allows the nearest samples to be semantically close to each other. In particular, we propose two possible \textit{ground truth similarity measures} between objects – \emph{hard} and \emph{soft} closeness. Under it, we derive an analytical form for the \textit{estimated similarity matrix} that accommodates both types of closeness between samples, thereby introducing dependency-aware loss functions. We validate our approach, \emph{Dependent TS2Vec}, on temporal and spatio-temporal downstream problems. Given the dependency patterns presented in the data, our approach surpasses modern ones for dependent data, highlighting the effectiveness of our theoretically grounded loss functions for SSL in capturing spatio-temporal dependencies. Specifically, we outperform TS2Vec on the standard UEA and UCR benchmarks, with accuracy improvements of $4.17$% and $2.08$%, respectively. Furthermore, on the drought classification task, which involves complex spatio-temporal patterns, our method achieves a $7$% higher ROC-AUC score.

[992] FedSPD: A Soft-clustering Approach for Personalized Decentralized Federated Learning

I-Cheng Lin, Osman Yagan, Carlee Joe-Wong

Main category: cs.LG

TL;DR: FedSPD is a personalized federated learning algorithm for decentralized settings that uses clustering to achieve consensus on distinct data clusters while personalizing models for individual clients, reducing communication costs and improving performance in low-connectivity networks.

DetailsMotivation: Traditional federated learning relies on a central server, while decentralized frameworks assume shared models. Personalizing models for individual clients can enhance performance with heterogeneous data distributions, but existing approaches have high communication costs.

Method: FedSPD uses a clustering-based framework that enables consensus on models for distinct data clusters while personalizing to unique mixtures of these clusters at different clients. It allows selective model updates based on data distribution.

Result: Experimental results show FedSPD outperforms multiple decentralized variants of personalized federated learning algorithms, especially in low-connectivity network scenarios.

Conclusion: FedSPD provides an efficient personalized federated learning solution for decentralized settings with theoretical convergence guarantees, reduced communication costs, and improved performance in challenging network conditions.

Abstract: Federated learning has recently gained popularity as a framework for distributed clients to collaboratively train a machine learning model using local data. While traditional federated learning relies on a central server for model aggregation, recent advancements adopt a decentralized framework, enabling direct model exchange between clients and eliminating the single point of failure. However, existing decentralized frameworks often assume all clients train a shared model. Personalizing each client’s model can enhance performance, especially with heterogeneous client data distributions. We propose FedSPD, an efficient personalized federated learning algorithm for the decentralized setting, and show that it learns accurate models even in low-connectivity networks. To provide theoretical guarantees on convergence, we introduce a clustering-based framework that enables consensus on models for distinct data clusters while personalizing to unique mixtures of these clusters at different clients. This flexibility, allowing selective model updates based on data distribution, substantially reduces communication costs compared to prior work on personalized federated learning in decentralized settings. Experimental results on real-world datasets show that FedSPD outperforms multiple decentralized variants of personalized federated learning algorithms, especially in scenarios with low-connectivity networks.

[993] Uncertainty in Supply Chain Digital Twins: A Quantum-Classical Hybrid Approach

Abdullah Abdullah, Fannya Ratana Sandjaja, Ayesha Abdul Majeed, Gyan Wickremasinghe, Karen Rafferty, Vishal Sharma

Main category: cs.LG

TL;DR: Quantum-classical hybrid ML models for uncertainty quantification, showing how quantum feature transformation affects uncertainty propagation with varying qubit counts.

DetailsMotivation: There's a gap in understanding how quantum feature transformations impact uncertainty quantification within quantum-classical hybrid machine learning architectures, particularly for complex applications like supply chain digital twins and financial risk assessment.

Method: Applied existing uncertainty quantification techniques to different models within a hybrid quantum-classical framework, examining how quantum feature transformation affects uncertainty propagation. Tested with increasing qubit counts from 4 to 16.

Result: Increasing qubits from 4 to 16 showed varied model responsiveness to outlier detection samples, which is critical for resilient decision-making in dynamic environments.

Conclusion: Quantum computing techniques can effectively transform data features for uncertainty quantification, particularly when combined with classical methods, demonstrating their potential for complex dynamic applications.

Abstract: This study investigates uncertainty quantification (UQ) using quantum-classical hybrid machine learning (ML) models for applications in complex and dynamic fields, such as attaining resiliency in supply chain digital twins and financial risk assessment. Although quantum feature transformations have been integrated into ML models for complex data tasks, a gap exists in determining their impact on UQ within their hybrid architectures (quantum-classical approach). This work applies existing UQ techniques for different models within a hybrid framework, examining how quantum feature transformation affects uncertainty propagation. Increasing qubits from 4 to 16 shows varied model responsiveness to outlier detection (OD) samples, which is a critical factor for resilient decision-making in dynamic environments. This work shows how quantum computing techniques can transform data features for UQ, particularly when combined with classical methods.

[994] What Is the Point of Equality in Machine Learning Fairness? Beyond Equality of Opportunity

Youjin Kong

Main category: cs.LG

TL;DR: This paper critiques the exclusive focus on distributive equality in ML fairness research and proposes a multifaceted egalitarian framework that integrates both distributive and relational equality to address allocative and representational harms.

DetailsMotivation: Current ML fairness research primarily focuses on distributive equality, which fails to address representational harms and explain why reinforcing social hierarchies through ML systems is wrong. The paper aims to provide a more comprehensive ethical foundation.

Method: The paper draws on critical social and political philosophy to develop a multifaceted egalitarian framework that integrates distributive equality (addressing allocative harms) and relational equality (addressing representational harms and social hierarchies).

Result: The proposed framework offers a more comprehensive ethical foundation for ML fairness that can address both allocative harms (economic loss) and representational harms (stereotypes, erasure, social stratification).

Conclusion: A complete approach to ML fairness requires moving beyond distributive equality alone and incorporating relational equality to challenge structural inequality and foster a society where people relate as equals, with practical implementation pathways across the ML pipeline.

Abstract: Fairness in machine learning (ML) has become a rapidly growing area of research. But why, in the first place, is unfairness in ML wrong? And why should we care about improving fairness? Most fair-ML research implicitly appeals to distributive equality: the idea that desirable benefits and goods, such as opportunities (e.g., Barocas et al., 2023), should be equally distributed across society. Unfair ML models, then, are seen as wrong because they unequally distribute such benefits. This paper argues that this exclusive focus on distributive equality offers an incomplete and potentially misleading ethical foundation. Grounding ML fairness in egalitarianism–the view that equality is a fundamental moral and social ideal–requires challenging structural inequality: systematic, institutional, and durable arrangements that privilege some groups while disadvantaging others. Structural inequality manifests through ML systems in two primary forms: allocative harms (e.g., economic loss) and representational harms (e.g., stereotypes, erasure). While distributive equality helps address allocative harms, it fails to explain why representational harms are wrong–why it is wrong for ML systems to reinforce social hierarchies that stratify people into superior and inferior groups–and why ML systems should aim to foster a society where people relate as equals (i.e., relational equality). To address these limitations, the paper proposes a multifaceted egalitarian framework for ML fairness that integrates both distributive and relational equality. Drawing on critical social and political philosophy, this framework offers a more comprehensive ethical foundation for tackling the full spectrum of harms perpetuated by ML systems. The paper also outlines practical pathways for implementing the framework across the entire ML pipeline.

[995] Contrastive MIM: A Contrastive Mutual Information Framework for Unified Generative and Discriminative Representation Learning

Micha Livne

Main category: cs.LG

TL;DR: cMIM is a contrastive extension of Mutual Information Machine that combines generative and discriminative learning, eliminating the need for data augmentation and being robust to batch size, while outperforming MIM and InfoNCE on classification/regression tasks.

DetailsMotivation: Existing representation learning methods like contrastive learning and MIM have trade-offs - MIM underperforms on discriminative tasks compared to state-of-the-art alternatives despite its generative strengths.

Method: Augments MIM with a novel contrastive objective to enforce global discriminative structure while retaining generative capabilities. Also introduces informative embeddings technique to extract enriched representations from encoder-decoder models.

Result: cMIM consistently outperforms MIM and InfoNCE in classification and regression tasks while preserving comparable reconstruction quality.

Conclusion: cMIM provides a unified framework for learning representations that are simultaneously effective for both discriminative and generative applications.

Abstract: Learning representations that generalize well to unknown downstream tasks is a central challenge in representation learning. Existing approaches such as contrastive learning, self-supervised masking, and denoising auto-encoders address this challenge with varying trade-offs. In this paper, we introduce the {contrastive Mutual Information Machine} (cMIM), a probabilistic framework that augments the Mutual Information Machine (MIM) with a novel contrastive objective. While MIM maximizes mutual information between inputs and latent variables and encourages clustering of latent codes, its representations underperform on discriminative tasks compared to state-of-the-art alternatives. cMIM addresses this limitation by enforcing global discriminative structure while retaining MIM’s generative strengths. We present two main contributions: (1) we propose cMIM, a contrastive extension of MIM that eliminates the need for positive data augmentation and is robust to batch size, unlike InfoNCE-based methods; (2) we introduce {informative embeddings}, a general technique for extracting enriched representations from encoder–decoder models that substantially improve discriminative performance without additional training, and which apply broadly beyond MIM. Empirical results demonstrate that cMIM consistently outperforms MIM and InfoNCE in classification and regression tasks, while preserving comparable reconstruction quality. These findings suggest that cMIM provides a unified framework for learning representations that are simultaneously effective for discriminative and generative applications.

[996] DiffKV: Differentiated Memory Management for Large Language Models with Parallel KV Compaction

Yanqi Zhang, Yuwei Hu, Runyuan Zhao, John C. S. Lui, Haibo Chen

Main category: cs.LG

TL;DR: DiffKV is a novel KV cache compression framework that uses three levels of differentiation to achieve 2.7-5.7x compression with near-lossless accuracy and 1.9-5.4x throughput improvement.

DetailsMotivation: LLMs face high serving costs due to memory demands, with KV cache being the main bottleneck. Existing compression techniques treat keys and values uniformly and discard unimportant tokens entirely, missing fine-grained distinctions.

Method: DiffKV exploits three differentiation levels: different impact of keys vs values on attention, varying token importance, and diverse sparsity patterns across attention heads. It includes an on-GPU memory manager that compacts fragmented memory in parallel.

Result: Achieves 2.7-5.7x KV cache compression with near-lossless accuracy on complex reasoning workloads, and 1.9-5.4x throughput improvement on mainstream LLMs including thinking models.

Conclusion: DiffKV effectively addresses KV cache memory bottlenecks by leveraging fine-grained differentiation and efficient memory management, translating sparsity into performance gains for LLM serving.

Abstract: Large language models (LLMs) demonstrate remarkable capabilities but face substantial serving costs due to their high memory demands, with the key-value (KV) cache being a primary bottleneck. State-of-the-art KV cache compression techniques, such as quantization and pruning, apply uniform treatment to both keys and values, and discard unimportant tokens entirely, overlooking the fine-grained distinctions in the significance of individual KV cache components. To address such limitations, we introduce \textit{DiffKV}, a novel framework for efficient KV cache compression that exploits three levels of differentiation in the KV cache: (1) the differing impact of keys and values on attention computation, (2) the varying importance of tokens, and (3) the diverse dynamic sparsity patterns across attention heads. These levels of differentiation introduce irregular memory usage patterns across different requests and attention heads, posing significant scalability challenges for memory management. To address these challenges, DiffKV proposes an on-GPU memory manager that compacts fragmented free memory list into contiguous regions in parallel, effectively translating sparsity in the KV cache into performance gains. We evaluate DiffKV on several mainstream LLMs, including the emerging thinking models that generate extended chains of thought. DiffKV is able to compress the KV cache by $2.7\times$ to $5.7\times$ with near-lossless accuracy on complex workloads requiring sophisticated reasoning and long-generation capabilities, and enhances throughput by $1.9\times$ to $5.4\times$. Source codes of DiffKV are available at https://github.com/zyqCSL/DiffKV.

[997] Iterative Distillation for Reward-Guided Fine-Tuning of Diffusion Models in Biomolecular Design

Xingyu Su, Xiner Li, Masatoshi Uehara, Sunwoo Kim, Yulai Zhao, Gabriele Scalia, Ehsan Hajiramezanali, Tommaso Biancalani, Degui Zhi, Shuiwang Ji

Main category: cs.LG

TL;DR: A distillation-based fine-tuning framework for diffusion models that enables optimization for arbitrary reward functions in biomolecular design, addressing instability and sample efficiency issues of RL methods.

DetailsMotivation: Real-world biomolecular design requires optimization beyond high-fidelity generation, often involving non-differentiable reward functions like physics-based simulations or scientific knowledge rewards. Existing RL methods suffer from instability, low sample efficiency, and mode collapse.

Method: Iterative distillation-based framework that treats the problem as policy distillation: collects off-policy data during roll-in, simulates reward-based soft-optimal policies during roll-out, and updates the model by minimizing KL divergence between simulated soft-optimal policy and current model policy.

Result: Enhanced training stability and sample efficiency compared to existing RL-based methods. Demonstrates effectiveness and superior reward optimization across diverse tasks in protein, small molecule, and regulatory DNA design.

Conclusion: The proposed off-policy formulation with KL divergence minimization provides a stable and efficient approach for fine-tuning diffusion models to optimize for arbitrary reward functions in biomolecular design applications.

Abstract: We address the problem of fine-tuning diffusion models for reward-guided generation in biomolecular design. While diffusion models have proven highly effective in modeling complex, high-dimensional data distributions, real-world applications often demand more than high-fidelity generation, requiring optimization with respect to potentially non-differentiable reward functions such as physics-based simulation or rewards based on scientific knowledge. Although RL methods have been explored to fine-tune diffusion models for such objectives, they often suffer from instability, low sample efficiency, and mode collapse due to their on-policy nature. In this work, we propose an iterative distillation-based fine-tuning framework that enables diffusion models to optimize for arbitrary reward functions. Our method casts the problem as policy distillation: it collects off-policy data during the roll-in phase, simulates reward-based soft-optimal policies during roll-out, and updates the model by minimizing the KL divergence between the simulated soft-optimal policy and the current model policy. Our off-policy formulation, combined with KL divergence minimization, enhances training stability and sample efficiency compared to existing RL-based methods. Empirical results demonstrate the effectiveness and superior reward optimization of our approach across diverse tasks in protein, small molecule, and regulatory DNA design.

[998] Addressing Key Challenges of Adversarial Attacks and Defenses in the Tabular Domain: A Methodological Framework for Coherence and Consistency

Yael Itzhakev, Amit Giloni, Yuval Elovici, Asaf Shabtai

Main category: cs.LG

TL;DR: A novel approach for evaluating tabular adversarial attacks using Class-Specific Anomaly Detection (CSAD) and SHAP explainability to ensure sample coherence and detect subtle perturbations.

DetailsMotivation: Tabular data has complex feature interdependencies that make adversarial attacks challenging, and existing evaluation metrics fail to account for sample coherence and class-specific distribution preservation.

Method: Proposed CSAD technique for perturbing dependent features while maintaining coherence, integrated with SHAP explainability for anomaly detection. Evaluated various black-box query-based and transferability-based gradient attacks across four target models on benchmark datasets.

Result: The approach revealed key differences in attacker risk, effort, and attack quality, providing insights into strengths, limitations, and trade-offs for both attackers and defenders in tabular domain.

Conclusion: The findings establish a foundation for future research on adversarial attacks and defense development in tabular data, addressing the unique challenges of feature interdependencies and sample coherence.

Abstract: Machine learning models trained on tabular data are vulnerable to adversarial attacks, even in realistic scenarios where attackers only have access to the model’s outputs. Since tabular data contains complex interdependencies among features, it presents a unique challenge for adversarial samples which must maintain coherence and respect these interdependencies to remain indistinguishable from benign data. Moreover, existing attack evaluation metrics-such as the success rate, perturbation magnitude, and query count-fail to account for this challenge. To address those gaps, we propose a technique for perturbing dependent features while preserving sample coherence. In addition, we introduce Class-Specific Anomaly Detection (CSAD), an effective novel anomaly detection approach, along with concrete metrics for assessing the quality of tabular adversarial attacks. CSAD evaluates adversarial samples relative to their predicted class distribution, rather than a broad benign distribution. It ensures that subtle adversarial perturbations, which may appear coherent in other classes, are correctly identified as anomalies. We integrate SHAP explainability techniques to detect inconsistencies in model decision-making, extending CSAD for SHAP-based anomaly detection. Our evaluation incorporates both anomaly detection rates with SHAP-based assessments to provide a more comprehensive measure of adversarial sample quality. We evaluate various attack strategies, examining black-box query-based and transferability-based gradient attacks across four target models. Experiments on benchmark tabular datasets reveal key differences in the attacker’s risk and effort and attack quality, offering insights into the strengths, limitations, and trade-offs faced by attackers and defenders. Our findings lay the groundwork for future research on adversarial attacks and defense development in the tabular domain.

[999] Investigating Parameter-Efficiency of Hybrid QuGANs Based on Geometric Properties of Generated Sea Route Graphs

Tobias Rohe, Florian Burger, Michael Kölle, Sebastian Wölckert, Maximilian Zorn, Claudia Linnhoff-Popien

Main category: cs.LG

TL;DR: Quantum-classical hybrid GANs (QuGANs) can generate shipping route graphs with similar quality to larger classical GANs, learning geometric properties efficiently but struggling with data variance.

DetailsMotivation: The demand for artificially generated data is high, and quantum computing's probabilistic nature offers potential for generative AI applications like shipping route generation.

Method: Used quantum-classical hybrid generative adversarial networks (QuGANs) to generate shipping route graphs, comparing them with classical GANs while focusing on parameter efficiency.

Result: QuGANs quickly learned underlying geometric properties and distributions of shipping data, achieving similar result quality to larger classical GANs but had difficulties introducing variance into sampled data.

Conclusion: QuGANs demonstrate potential for quantum computing applications in generative AI, showing parameter efficiency advantages while highlighting current limitations in data variance generation.

Abstract: The demand for artificially generated data for the development, training and testing of new algorithms is omnipresent. Quantum computing (QC), does offer the hope that its inherent probabilistic functionality can be utilised in this field of generative artificial intelligence. In this study, we use quantum-classical hybrid generative adversarial networks (QuGANs) to artificially generate graphs of shipping routes. We create a training dataset based on real shipping data and investigate to what extent QuGANs are able to learn and reproduce inherent distributions and geometric features of this data. We compare hybrid QuGANs with classical Generative Adversarial Networks (GANs), with a special focus on their parameter efficiency. Our results indicate that QuGANs are indeed able to quickly learn and represent underlying geometric properties and distributions, although they seem to have difficulties in introducing variance into the sampled data. Compared to classical GANs of greater size, measured in the number of parameters used, some QuGANs show similar result quality. Our reference to concrete use cases, such as the generation of shipping data, provides an illustrative example and demonstrate the potential and diversity in which QC can be used.

[1000] chebgreen: Learning and Interpolating Continuous Empirical Green’s Functions from Data

Harshwardhan Praveen, Jacob Brown, Christopher Earls

Main category: cs.LG

TL;DR: chebgreen is a mesh-independent data-driven library that learns Empirical Green’s Functions for unknown 1D PDEs using Rational Neural Networks and Chebyshev basis interpolation.

DetailsMotivation: To model one-dimensional systems with unknown governing partial differential equations and associated control parameters, where traditional analytical methods are not applicable.

Method: Learns Empirical Green’s Function as a Rational Neural Network, constructs bivariate representation in Chebyshev basis, and interpolates singular functions on Quasimatrices manifold with Lagrange polynomial interpolation for singular values.

Result: The method successfully uncovers Green’s function at unseen control parameter values through interpolation of singular functions and values.

Conclusion: chebgreen provides an effective data-driven approach for modeling 1D systems with unknown PDEs by learning and interpolating Green’s functions in a mathematical framework.

Abstract: In this work, we present a mesh-independent, data-driven library, chebgreen, to mathematically model one-dimensional systems, possessing an associated control parameter, and whose governing partial differential equation is unknown. The proposed method learns an Empirical Green’s Function for the associated, but hidden, boundary value problem, in the form of a Rational Neural Network from which we subsequently construct a bivariate representation in a Chebyshev basis. We uncover the Green’s function, at an unseen control parameter value, by interpolating the left and right singular functions within a suitable library, expressed as points on a manifold of Quasimatrices, while the associated singular values are interpolated with Lagrange polynomials.

[1001] Memory Capacity of Nonlinear Recurrent Networks: Is it Informative?

Giovanni Ballarin, Lyudmila Grigoryeva, Juan-Pablo Ortega

Main category: cs.LG

TL;DR: The paper shows that memory capacity (MC) of random nonlinear RNNs can vary arbitrarily within bounds based solely on input scale, making existing MC definitions practically useless for distinguishing RNN performance.

DetailsMotivation: Previous work showed linear RNNs have maximal memory capacity almost surely, questioning MC's usefulness as a performance metric for distinguishing RNNs processing stochastic signals.

Method: Analysis of memory capacity in random nonlinear recurrent neural networks, examining how MC values vary within established upper and lower bounds depending on input process scale.

Result: Memory capacity of random nonlinear RNNs yields arbitrary values within known bounds that depend exclusively on the scale of the input process.

Conclusion: The existing definition of memory capacity for both linear and nonlinear RNNs has no practical value for distinguishing network performance.

Abstract: The total memory capacity (MC) of linear recurrent neural networks (RNNs) has been proven to be equal to the rank of the corresponding Kalman controllability matrix, and it is almost surely maximal for connectivity and input weight matrices drawn from regular distributions. This fact questions the usefulness of this metric in distinguishing the performance of linear RNNs in the processing of stochastic signals. This work shows that the MC of random nonlinear RNNs yields arbitrary values within established upper and lower bounds depending exclusively on the scale of the input process. This confirms that the existing definition of MC in linear and nonlinear cases has no practical value.

[1002] Debiasing Guidance for Discrete Diffusion with Sequential Monte Carlo

Cheuk Kit Lee, Paul Jeha, Jes Frellsen, Pietro Lio, Michael Samuel Albergo, Francisco Vargas

Main category: cs.LG

TL;DR: A Sequential Monte Carlo algorithm for unbiased sampling from target distributions in discrete diffusion models, providing better control while maintaining quality compared to guidance methods.

DetailsMotivation: Current guidance methods in discrete diffusion models fail to properly sample from target distributions proportional to p₀(x₀)p(ζ|x₀)^α, creating a need for more effective control mechanisms.

Method: Introduces a Sequential Monte Carlo algorithm that leverages both unconditional and guided diffusion processes to generate unbiased samples from the target distribution.

Result: Validated on low-dimensional distributions, controlled images, and text generation. For text, achieves strong control while maintaining low perplexity compared to guidance-based approaches.

Conclusion: The proposed Sequential Monte Carlo method successfully addresses limitations of current guidance techniques, enabling unbiased sampling from target distributions with improved control capabilities.

Abstract: Discrete diffusion models are a class of generative models that produce samples from an approximated data distribution within a discrete state space. Often, there is a need to target specific regions of the data distribution. Current guidance methods aim to sample from a distribution with mass proportional to $p_0(x_0) p(\zeta|x_0)^\alpha$ but fail to achieve this in practice. We introduce a Sequential Monte Carlo algorithm that generates unbiasedly from this target distribution, utilising the learnt unconditional and guided process. We validate our approach on low-dimensional distributions, controlled images and text generations. For text generation, our method provides strong control while maintaining low perplexity compared to guidance-based approaches.

[1003] A Comprehensive Survey on Imbalanced Data Learning

Xinyi Gao, Dongting Xie, Yihang Zhang, Zhengren Wang, Conghui He, Hongzhi Yin, Wentao Zhang

Main category: cs.LG

TL;DR: Survey paper analyzing imbalanced data distributions in machine learning, categorizing solutions into four approaches: data re-balancing, feature representation, training strategy, and ensemble learning.

DetailsMotivation: Imbalanced data distributions are prevalent in real-world data and severely hinder ML performance by biasing decision-making processes. The paper aims to deepen understanding and facilitate research on handling imbalanced data.

Method: Systematic analysis of various real-world data formats and categorization of existing research into four distinct solution categories. Provides overview of open-source libraries and identifies current challenges.

Result: Structured framework that helps researchers comprehensively understand imbalance across diverse data formats and provides clearer path toward achieving specific research goals.

Conclusion: The survey offers novel insights to foster future advancements in handling imbalanced data, highlighting this as a critical area of study in machine learning.

Abstract: With the expansion of data availability, machine learning (ML) has achieved remarkable breakthroughs in both academia and industry. However, imbalanced data distributions are prevalent in various types of raw data and severely hinder the performance of ML by biasing the decision-making processes. To deepen the understanding of imbalanced data and facilitate the related research and applications, this survey systematically analyzes various real-world data formats and concludes existing researches for different data formats into four distinct categories: data re-balancing, feature representation, training strategy, and ensemble learning. This structured analysis helps researchers comprehensively understand the pervasive nature of imbalance across diverse data formats, thereby paving a clearer path toward achieving specific research goals. We provide an overview of relevant open-source libraries, spotlight current challenges, and offer novel insights aimed at fostering future advancements in this critical area of study.

[1004] Shortcut Learning Susceptibility in Vision Classifiers

Pirzada Suhail, Vrinda Goel, Amit Sethi

Main category: cs.LG

TL;DR: Systematic evaluation of CNN, MLP, and ViT architectures’ susceptibility to shortcut learning using controlled datasets with artificial shortcuts, revealing CNN’s relative robustness and ViT’s tendency to ignore actual features when shortcuts are present.

DetailsMotivation: Shortcut learning poses significant challenges to building robust ML models across vision, NLP, and speech applications, where models exploit spurious correlations instead of learning meaningful features.

Method: Introduce deliberate positional and intensity-based shortcuts correlated with class labels, train models on modified dataset, test on shortcut-containing and clean test sets, use network inversion for qualitative analysis, and evaluate across different learning rates.

Result: CNNs at lower learning rates show more resistance to shortcut features, while ViTs (especially without positional encodings) almost completely ignore distinctive image features when shortcuts are available.

Conclusion: Different architectures exhibit varying susceptibility to shortcut learning, with CNNs demonstrating better robustness against spurious correlations compared to ViTs in the presence of artificial shortcuts.

Abstract: Shortcut learning, where machine learning models exploit spurious correlations in data instead of capturing meaningful features, poses a significant challenge to building robust and generalizable models. This phenomenon is prevalent across various machine learning applications, including vision, natural language processing, and speech recognition, where models may find unintended cues that minimize training loss but fail to capture the underlying structure of the data. Vision classifiers based on Convolutional Neural Networks (CNNs), Multi-Layer Perceptrons (MLPs), and Vision Transformers (ViTs) leverage distinct architectural principles to process spatial and structural information, making them differently susceptible to shortcut learning. In this study, we systematically evaluate these architectures by introducing deliberate shortcuts into the dataset that are correlated with class labels both positionally and via intensity, creating a controlled setup to assess whether models rely on these artificial cues or learn actual distinguishing features. We perform both quantitative evaluation by training on the shortcut-modified dataset and testing on two different test sets-one containing the same shortcuts and another without them-to determine the extent of reliance on shortcuts. Additionally, qualitative evaluation is performed using network inversion-based reconstruction techniques to analyze what the models internalize in their weights, aiming to reconstruct the training data as perceived by the classifiers. Further, we evaluate susceptibility to shortcut learning across different learning rates. Our analysis reveals that CNNs at lower learning rates tend to be more reserved against entirely picking up shortcut features, while ViTs, particularly those without positional encodings, almost entirely ignore the distinctive image features in the presence of shortcuts.

[1005] A Survey on Human-Centered Evaluation of Explainable AI Methods in Clinical Decision Support Systems

Alessandro Gambetti, Qiwei Han, Hong Shen, Claudia Soares

Main category: cs.LG

TL;DR: Survey paper on human-centered evaluations of Explainable AI methods in Clinical Decision Support Systems, analyzing XAI methodologies, evaluation frameworks, and clinical adoption challenges.

DetailsMotivation: XAI is crucial for CDSS to enhance transparency and trust, but the effectiveness of existing XAI methods in real-world medical settings remains underexplored.

Method: Conducted a comprehensive survey categorizing existing works based on XAI methodologies, evaluation frameworks, and clinical adoption challenges.

Result: Revealed key challenges in integrating XAI into healthcare workflows and identified gaps in current evaluation approaches.

Conclusion: Proposed a structured framework to align XAI evaluation methods with clinical stakeholder needs to improve real-world adoption and effectiveness.

Abstract: Explainable AI (XAI) has become a crucial component of Clinical Decision Support Systems (CDSS) to enhance transparency, trust, and clinical adoption. However, while many XAI methods have been proposed, their effectiveness in real-world medical settings remains underexplored. This paper provides a survey of human-centered evaluations of Explainable AI methods in Clinical Decision Support Systems. By categorizing existing works based on XAI methodologies, evaluation frameworks, and clinical adoption challenges, we offer a structured understanding of the landscape. Our findings reveal key challenges in the integration of XAI into healthcare workflows and propose a structured framework to align the evaluation methods of XAI with the clinical needs of stakeholders.

[1006] Minimal Ranks, Maximum Confidence: Parameter-efficient Uncertainty Quantification for LoRA

Patryk Marszałek, Klaudia Bałazy, Jacek Tabor, Tomasz Kuśmierczyk

Main category: cs.LG

TL;DR: Proposes a parameter-efficient Bayesian LoRA method via subspace inference that enables effective uncertainty quantification in low-dimensional spaces while maintaining computational efficiency.

DetailsMotivation: Standard LoRA lacks uncertainty quantification, leading to overconfident models. Bayesian LoRA variants address this but increase parameter count and training difficulty, offsetting efficiency gains.

Method: Novel parameter-efficient Bayesian LoRA via subspace inference, projecting weight space to enable uncertainty modeling in very low-dimensional parameter spaces.

Result: Achieves strong performance with improved calibration and generalization while maintaining computational efficiency. Weight covariances exhibit low ranks.

Conclusion: Effective uncertainty quantification can be achieved in low-dimensional spaces with appropriate weight space projection, making Bayesian LoRA both efficient and well-calibrated.

Abstract: Low-Rank Adaptation (LoRA) enables parameter-efficient fine-tuning of large language models by decomposing weight updates into low-rank matrices, significantly reducing storage and computational overhead. While effective, standard LoRA lacks mechanisms for uncertainty quantification, leading to overconfident and poorly calibrated models. Bayesian variants of LoRA address this limitation, but at the cost of a significantly increased number of trainable parameters, partially offsetting the original efficiency gains. Additionally, these models are harder to train and may suffer from unstable convergence. In this work, we propose a novel parameter-efficient Bayesian LoRA via subspace inference, demonstrating that effective uncertainty quantification can be achieved in very low-dimensional parameter spaces. The proposed method achieves strong performance with improved calibration and generalization while maintaining computational efficiency. Our empirical findings show that, with the appropriate projection of the weight space: (1) uncertainty can be effectively modeled in a low-dimensional space, and (2) weight covariances exhibit low ranks.

[1007] Convergence Analysis of Aggregation-Broadcast in LoRA-enabled Distributed Fine-Tuning

Xin Chen, Shuaijun Chen, Omid Tavallaie, Nguyen Tran, Shuhuang Xiang, Albert Zomaya

Main category: cs.LG

TL;DR: The paper provides a unified convergence analysis for LoRA-based Federated Learning, categorizing aggregation methods into Sum-Product and Product-Sum types, and establishes both weak and strong convergence conditions with theoretical guarantees.

DetailsMotivation: While Low-Rank Adaptation (LoRA) reduces communication overhead in Federated Learning by updating only a small number of parameters, how to effectively aggregate these LoRA-updated local models remains an understudied critical problem.

Method: The authors categorize existing aggregation methods into Sum-Product (SP) and Product-Sum (PS) types, define the Aggregation-Broadcast Operator (ABO), and derive both weak and strong convergence conditions under mild assumptions through theoretical analysis.

Result: Theoretical analysis shows that SP and PS aggregation methods satisfy weak and strong convergence conditions respectively, but differ in achieving optimal convergence rates. Extensive experiments on standard benchmarks validate these theoretical findings.

Conclusion: The paper provides a principled theoretical framework for understanding LoRA-based FL aggregation strategies, proving convergence guarantees for different aggregation methods and offering insights into their performance differences.

Abstract: Federated Learning (FL) enables collaborative model training across decentralized data sources while preserving data privacy. However, the growing size of Machine Learning (ML) models poses communication and computation challenges in FL. Low-Rank Adaptation (LoRA) has recently been introduced into FL as an efficient fine-tuning method, reducing communication overhead by updating only a small number of trainable parameters. Despite its effectiveness, how to aggregate LoRA-updated local models on the server remains a critical and understudied problem. In this paper, we provide a unified convergence analysis for LoRA-based FL. We first categories the current aggregation method into two major type: Sum-Product (SP) and Product-Sum (PS). Then we formally define the Aggregation-Broadcast Operator (ABO) and derive both weak and strong convergence condition under mild assumptions. Furthermore, we present both weak and strong convergence condition that guarantee convergence of the local model and the global model respectively. These theoretical analyze offer a principled understanding of various aggregation strategies. Notably, we prove that the SP and PS aggregation methods satisfy the weak and strong convergence condition respectively, but differ in their ability to achieve the optimal convergence rate. Extensive experiments on standard benchmarks validate our theoretical findings.

[1008] Tighten The Lasso: A Convex Hull Volume-based Anomaly Detection Method

Uri Itai, Asael Bar Ilan, Teddy Lazebnik

Main category: cs.LG

TL;DR: Novel OOD detection method using convex hull volume changes to distinguish between in-distribution and out-of-distribution data, achieving performance comparable to SOTA methods with computational efficiency analysis.

DetailsMotivation: Detecting out-of-distribution data is crucial for maintaining model reliability and robustness in machine learning systems.

Method: Leverages convex hull property by observing that OOD samples marginally increase the convex hull’s volume. Establishes decision boundary by iteratively computing CH volume as samples are removed, stopping when removal doesn’t significantly alter volume.

Result: Evaluated against seven widely used anomaly detection methods across ten datasets, demonstrating performance comparable to state-of-the-art techniques.

Conclusion: Proposed algorithm provides effective OOD detection with comparable performance to SOTA methods, plus introduces computationally efficient criterion to identify datasets where it outperforms existing approaches.

Abstract: Detecting out-of-distribution (OOD) data is a critical task for maintaining model reliability and robustness. In this study, we propose a novel anomaly detection algorithm that leverages the convex hull (CH) property of a dataset by exploiting the observation that OOD samples marginally increase the CH’s volume compared to in-distribution samples. Thus, we establish a decision boundary between OOD and in-distribution data by iteratively computing the CH’s volume as samples are removed, stopping when such removal does not significantly alter the CH’s volume. The proposed algorithm is evaluated against seven widely used anomaly detection methods across ten datasets, demonstrating performance comparable to state-of-the-art (SOTA) techniques. Furthermore, we introduce a computationally efficient criterion for identifying datasets where the proposed method outperforms existing SOTA approaches.

[1009] A DbC Inspired Neurosymbolic Layer for Trustworthy Agent Design

Claudiu Leoveanu-Condrei

Main category: cs.LG

TL;DR: Introduces a contract layer for LLMs that applies Design by Contract principles to provide semantic and type guarantees on inputs/outputs, with probabilistic remediation to ensure compliance.

DetailsMotivation: LLMs produce fluent outputs but lack verifiable guarantees, creating reliability issues for applications requiring predictable behavior.

Method: Adapts Design by Contract and type-theoretic principles to create a mediating contract layer that specifies semantic and type requirements, using probabilistic remediation to steer generation toward compliance.

Result: Enables probabilistic contract satisfaction and semantic validation through programmer-specified conditions on well-typed data structures, treating LLMs as both semantic parsers and probabilistic black-box components.

Conclusion: Establishes that any two agents satisfying the same contracts are functionally equivalent with respect to those contracts, providing a foundation for reliable LLM-based systems.

Abstract: Generative models, particularly Large Language Models (LLMs), produce fluent outputs yet lack verifiable guarantees. We adapt Design by Contract (DbC) and type-theoretic principles to introduce a contract layer that mediates every LLM call. Contracts stipulate semantic and type requirements on inputs and outputs, coupled with probabilistic remediation to steer generation toward compliance. The layer exposes the dual view of LLMs as semantic parsers and probabilistic black-box components. Contract satisfaction is probabilistic and semantic validation is operationally defined through programmer-specified conditions on well-typed data structures. More broadly, this work postulates that any two agents satisfying the same contracts are \emph{functionally equivalent} with respect to those contracts.

[1010] Scalable Graph Attention-based Instance Selection via Mini-Batch Sampling and Hierarchical Hashing

Zahiriddin Rustamov, Ayham Zaitouny, Nazar Zaki

Main category: cs.LG

TL;DR: GAIS uses graph attention mechanisms for scalable instance selection, achieving 96%+ reduction while maintaining model performance through efficient mini-batch and hierarchical hashing approaches.

DetailsMotivation: Current instance selection methods struggle with complex relationships in high-dimensional spaces and scalability with large datasets containing millions of instances.

Method: Graph attention-based instance selection (GAIS) with two scalable graph construction approaches: distance-based mini-batch sampling with strategic batch processing, and hierarchical hashing with random projections for efficient similarity computation.

Result: Experiments on 39 datasets show reduction rates above 96% while maintaining or improving model performance compared to state-of-the-art methods. Mini-batch approach optimal for large datasets, multi-view variants excel on complex high-dimensional data.

Conclusion: Attention-based importance scoring effectively identifies instances important for maintaining decision boundaries without computationally expensive pairwise comparisons, providing scalable instance selection for large datasets.

Abstract: Instance selection (IS) addresses the critical challenge of reducing dataset size while keeping informative characteristics, becoming increasingly important as datasets grow to millions of instances. Current IS methods often struggle with capturing complex relationships in high-dimensional spaces and scale with large datasets. This paper introduces a graph attention-based instance selection (GAIS) method that uses attention mechanisms to identify informative instances through their structural relationships in graph representations. We present two approaches for scalable graph construction: a distance-based mini-batch sampling technique that achieves dataset-size-independent complexity through strategic batch processing, and a hierarchical hashing approach that enables efficient similarity computation through random projections. The mini-batch approach keeps class distributions through stratified sampling, while the hierarchical hashing method captures relationships at multiple granularities through single-level, multi-level, and multi-view variants. Experiments across 39 datasets show that GAIS achieves reduction rates above 96% while maintaining or improving model performance relative to state-of-the-art IS methods. The findings show that the distance-based mini-batch approach offers an optimal efficiency for large-scale datasets, while multi-view variants excel on complex, high-dimensional data, demonstrating that attention-based importance scoring can effectively identify instances important for maintaining decision boundaries while avoiding computationally prohibitive pairwise comparisons.

[1011] Armijo Line-search Can Make (Stochastic) Gradient Descent Provably Faster

Sharan Vaswani, Reza Babanezhad

Main category: cs.LG

TL;DR: Armijo line-search for gradient descent provides faster convergence than fixed step-size methods for smooth functions, achieving linear rates for convex logistic regression and matching specialized algorithms for non-convex gradient domination problems.

DetailsMotivation: Standard gradient descent requires knowing the global smoothness constant L, while Armijo line-search adapts to local smoothness and can achieve better convergence rates without this knowledge.

Method: Analyzes gradient descent with Armijo line-search (GD-LS) compared to GD with fixed 1/L step-size, focusing on functions satisfying non-uniform smoothness conditions, convex objectives (logistic regression, classification), and non-convex objectives with gradient domination.

Result: GD-LS achieves linear convergence for convex logistic regression/classification (improving over sublinear GD(1/L)), matches specialized algorithm convergence for non-convex gradient domination problems, and stochastic GD with line-search converges under interpolation.

Conclusion: Armijo line-search provides significant convergence improvements over fixed step-size gradient descent across various problem classes, making it a superior practical choice that adapts to local smoothness properties.

Abstract: Armijo line-search (Armijo-LS) is a standard method to set the step-size for gradient descent (GD). For smooth functions, Armijo-LS alleviates the need to know the global smoothness constant L and adapts to the ``local’’ smoothness, enabling GD to converge faster. Existing theoretical analyses show that GD with Armijo-LS (GD-LS) can result in constant factor improvements over GD with a 1/L step-size (denoted as GD(1/L)). We strengthen these results and show that if the objective function satisfies a certain non-uniform smoothness condition, GD-LS can result in a faster convergence rate than GD(1/L). In particular, we prove that for convex objectives corresponding to logistic regression and multi-class classification, GD-LS can converge to the optimum at a linear rate, and hence improves over the sublinear convergence of GD(1/L). Furthermore, for non-convex objectives satisfying gradient domination (e.g., those corresponding to the softmax policy gradient in RL or generalized linear models with a logistic link function), GD-LS can match the fast convergence of algorithms tailored for these specific settings. Finally, we analyze the convergence of stochastic GD with a stochastic line-search on convex losses under the interpolation assumption.

[1012] Breaking Free: Decoupling Forced Systems with Laplace Neural Networks

Bernd Zimmering, Cecília Coelho, Vaibhav Gupta, Maria Maleshkova, Oliver Niggemann

Main category: cs.LG

TL;DR: Laplace-Net is a neural framework that uses Laplace transforms to model forced dynamical systems, enabling decoupled learning of internal dynamics, external inputs, and initial conditions for improved interpretability and transferability.

DetailsMotivation: Modelling forced dynamical systems with external inputs is critical across engineering, finance, and natural sciences, but existing approaches lack interpretability and flexibility for new forcing signals.

Method: A decoupled, solver-free neural framework leveraging Laplace transform-based approach to decompose internal dynamics, external inputs, and initial values into established theoretical concepts.

Result: Experimental results on eight benchmark datasets (linear, non-linear, delayed systems) show improved accuracy and robustness compared to state-of-the-art approaches, especially with complex and unseen inputs.

Conclusion: Laplace-Net provides enhanced interpretability, transferability for new forcing signals, and flexibility for applications ranging from controller adaptation to long-horizon forecasting.

Abstract: Modelling forced dynamical systems - where an external input drives the system state - is critical across diverse domains such as engineering, finance, and the natural sciences. In this work, we propose Laplace-Net, a decoupled, solver-free neural framework for learning forced and delay-aware systems. It leverages a Laplace transform-based approach to decompose internal dynamics, external inputs, and initial values into established theoretical concepts, enhancing interpretability. Laplace-Net promotes transferability since the system can be rapidly re-trained or fine-tuned for new forcing signals, providing flexibility in applications ranging from controller adaptation to long-horizon forecasting. Experimental results on eight benchmark datasets - including linear, non-linear, and delayed systems - demonstrate the method’s improved accuracy and robustness compared to state-of-the-art approaches, particularly in handling complex and previously unseen inputs.

[1013] Class Unbiasing for Generalization in Medical Diagnosis

Lishi Zuo, Man-Wai Mak, Lu Yi, Youzhi Tu

Main category: cs.LG

TL;DR: Proposes a method to mitigate class-feature bias and class imbalance simultaneously using class-wise inequality loss and group distributionally robust optimization.

DetailsMotivation: Medical diagnosis models can fail due to bias, specifically class-feature bias where models rely on features correlated with only some classes, leading to poor generalization.

Method: Class-wise inequality loss to promote equal contributions from positive and negative class samples, combined with class-weighted training that upweights underperforming classes.

Result: Empirical demonstration on synthetic and real-world datasets shows class-feature bias negatively impacts performance, and the proposed method effectively mitigates both biases.

Conclusion: The approach improves model generalization ability by addressing both class-feature bias and class imbalance simultaneously.

Abstract: Medical diagnosis might fail due to bias. In this work, we identified class-feature bias, which refers to models’ potential reliance on features that are strongly correlated with only a subset of classes, leading to biased performance and poor generalization on other classes. We aim to train a class-unbiased model (Cls-unbias) that mitigates both class imbalance and class-feature bias simultaneously. Specifically, we propose a class-wise inequality loss which promotes equal contributions of classification loss from positive-class and negative-class samples. We propose to optimize a class-wise group distributionally robust optimization objective-a class-weighted training objective that upweights underperforming classes-to enhance the effectiveness of the inequality loss under class imbalance. Through synthetic and real-world datasets, we empirically demonstrate that class-feature bias can negatively impact model performance. Our proposed method effectively mitigates both class-feature bias and class imbalance, thereby improving the model’s generalization ability.

[1014] Membership Inference Attacks on Large-Scale Models: A Survey

Hengyu Wu, Yang Cao

Main category: cs.LG

TL;DR: First comprehensive survey of Membership Inference Attacks (MIAs) on Large Language Models and Large Multimodal Models, analyzing attacks across different model types, adversarial knowledge levels, strategies, and pipeline stages.

DetailsMotivation: Privacy risks of large-scale models remain underexplored despite increasing deployment. MIAs are important for exposing privacy risks but lack systematic surveys for large-scale models unlike classic ML models.

Method: Comprehensive review analyzing MIAs by model type (LLMs/LMMs), adversarial knowledge, attack strategies, and across multiple pipeline stages including pre-training, fine-tuning, alignment, and RAG.

Result: Provides systematic categorization and analysis of MIA effectiveness and limitations in large-scale models, identifying current attack capabilities and vulnerabilities.

Conclusion: Identifies open challenges and proposes future research directions for strengthening privacy resilience in large-scale models through improved MIA defense mechanisms.

Abstract: As large-scale models such as Large Language Models (LLMs) and Large Multimodal Models (LMMs) see increasing deployment, their privacy risks remain underexplored. Membership Inference Attacks (MIAs), which reveal whether a data point was used in training the target model, are an important technique for exposing or assessing privacy risks and have been shown to be effective across diverse machine learning algorithms. However, despite extensive studies on MIAs in classic models, there remains a lack of systematic surveys addressing their effectiveness and limitations in large-scale models. To address this gap, we provide the first comprehensive review of MIAs targeting LLMs and LMMs, analyzing attacks by model type, adversarial knowledge, and strategy. Unlike prior surveys, we further examine MIAs across multiple stages of the model pipeline, including pre-training, fine-tuning, alignment, and Retrieval-Augmented Generation (RAG). Finally, we identify open challenges and propose future research directions for strengthening privacy resilience in large-scale models.

[1015] Network Inversion for Generating Confidently Classified Counterfeits

Pirzada Suhail, Pravesh Khaparde, Amit Sethi

Main category: cs.LG

TL;DR: Paper introduces Confidently Classified Counterfeits (CCCs) - synthetic samples that models confidently classify despite being significantly different from training data, challenging assumptions in OOD detection.

DetailsMotivation: Traditional adversarial methods are input-dependent and fail to ensure both high confidence and meaningful deviation from training data. Need to understand model confidence behavior under synthetic OOD conditions.

Method: Extend network inversion techniques by replacing soft vector conditioning with one-hot class conditioning and adding KL divergence loss between one-hot label and classifier’s output distribution.

Result: Models can assign high confidence to entirely synthetic, out-of-distribution inputs, challenging core assumptions of confidence-based OOD detection techniques.

Conclusion: High-confidence outputs do not necessarily imply in-distribution data, highlighting need for more robust uncertainty estimation in safety-critical applications.

Abstract: In vision classification, generating inputs that elicit confident predictions is key to understanding model behavior and reliability, especially under adversarial or out-of-distribution (OOD) conditions. While traditional adversarial methods rely on perturbing existing inputs to fool a model, they are inherently input-dependent and often fail to ensure both high confidence and meaningful deviation from the training data. In this work, we extend network inversion techniques to generate Confidently Classified Counterfeits (CCCs), synthetic samples that are confidently classified by the model despite being significantly different from the training distribution and independent of any specific input. We alter inversion technique by replacing soft vector conditioning with one-hot class conditioning and introducing a Kullback-Leibler divergence loss between the one-hot label and the classifier’s output distribution. CCCs offer a model-centric perspective on confidence, revealing that models can assign high confidence to entirely synthetic, out-of-distribution inputs. This challenges the core assumption behind many OOD detection techniques based on thresholding prediction confidence, which assume that high-confidence outputs imply in-distribution data, and highlights the need for more robust uncertainty estimation in safety-critical applications.

[1016] Grid2Guide: A* Enabled Small Language Model for Indoor Navigation

Md. Wasiul Haque, Sagar Dasgupta, Mizanur Rahman

Main category: cs.LG

TL;DR: Grid2Guide: A hybrid indoor navigation framework combining A* algorithm with Small Language Model to generate human-readable route instructions without external infrastructure.

DetailsMotivation: Reliable indoor navigation remains challenging in complex environments without external positioning signals or dedicated infrastructure.

Method: Creates binary occupancy matrix from indoor map, uses A* algorithm to compute optimal path, then transforms steps into natural language instructions using Small Language Model.

Result: Experimental evaluations demonstrate effectiveness in producing accurate and timely navigation guidance across various indoor scenarios.

Conclusion: Validated as a lightweight, infrastructure-free solution for real-time indoor navigation support.

Abstract: Reliable indoor navigation remains a significant challenge in complex environments, particularly where external positioning signals and dedicated infrastructures are unavailable. This research presents Grid2Guide, a hybrid navigation framework that combines the A* search algorithm with a Small Language Model (SLM) to generate clear, human-readable route instructions. The framework first conducts a binary occupancy matrix from a given indoor map. Using this matrix, the A* algorithm computes the optimal path between origin and destination, producing concise textual navigation steps. These steps are then transformed into natural language instructions by the SLM, enhancing interpretability for end users. Experimental evaluations across various indoor scenarios demonstrate the method’s effectiveness in producing accurate and timely navigation guidance. The results validate the proposed approach as a lightweight, infrastructure-free solution for real-time indoor navigation support.

[1017] Learnable cut flow for high energy physics

Jing Li, Hao Sun

Main category: cs.LG

TL;DR: LCF is a neural network that makes traditional cut flow methods differentiable and data-driven, combining interpretability with neural network power while providing feature importance insights.

DetailsMotivation: Neural networks are powerful but opaque black boxes, while traditional cut flow methods are interpretable but require manual tuning. There's a need to merge both approaches' strengths.

Method: Learnable Cut Flow (LCF) transforms cut selection into differentiable process using parallel and sequential strategies, with modified loss function using mask operations instead of hard cuts. Includes Learnable Importance metric for feature importance.

Result: LCF accurately learns cut boundaries, assigns higher importance to discriminative features, handles redundant features robustly, and performs effectively in real-world scenarios, though initially underperforms other methods on diboson dataset.

Conclusion: LCF successfully bridges traditional cut flow methods and neural networks, providing interpretable insights into training process and feature importance while maintaining performance.

Abstract: Neural networks have emerged as a powerful paradigm for tasks in high energy physics, yet their opaque training process renders them as a black box. In contrast, the traditional cut flow method offers simplicity and interpretability but requires extensive manual tuning to identify optimal cut boundaries. To merge the strengths of both approaches, we propose the Learnable Cut Flow (LCF), a neural network that transforms the traditional cut selection into a fully differentiable, data-driven process. LCF implements two cut strategies-parallel, where observable distributions are treated independently, and sequential, where prior cuts shape subsequent ones-to flexibly determine optimal boundaries. Building on this strategy, we introduce the Learnable Importance, a metric that quantifies feature importance and adjusts their contributions to the loss accordingly, offering model-driven insights unlike ad-hoc metrics. To ensure differentiability, a modified loss function replaces hard cuts with mask operations, preserving data shape throughout the training process. LCF is tested on six varied mock datasets and a realistic diboson vs. QCD dataset. Results demonstrate that LCF 1. accurately learns cut boundaries across typical feature distributions in both parallel and sequential strategies, 2. assigns higher importance to discriminative features with minimal overlap, 3. handles redundant or correlated features robustly, and 4. performs effectively in real-world scenarios. In the diboson dataset, LCF initially underperforms boosted decision trees and multiplayer perceptrons when using all observables. LCF bridges the gap between traditional cut flow method and modern black-box neural networks, delivering actionable insights into the training process and feature importance. Source code and experimental data are available at https://github.com/Star9daisy/learnable-cut-flow.

[1018] UQGNN: Uncertainty Quantification of Graph Neural Networks for Multivariate Spatiotemporal Prediction

Dahai Yu, Dingyi Zhuang, Lin Jiang, Rongchao Xu, Xinyue Ye, Yuheng Bu, Shenhao Wang, Guang Wang

Main category: cs.LG

TL;DR: UQGNN is a novel Graph Neural Network that addresses uncertainty quantification in multivariate spatiotemporal prediction, capturing correlations between different urban phenomena and achieving state-of-the-art performance.

DetailsMotivation: Most existing spatiotemporal prediction models are deterministic and lack uncertainty quantification, while probabilistic models typically focus on single phenomena, ignoring correlations between heterogeneous urban events.

Method: Proposes UQGNN with two key components: (1) Interaction-aware Spatiotemporal Embedding Module using multivariate diffusion graph convolutional network and interaction-aware temporal convolutional network, and (2) multivariate probabilistic prediction module for estimating both mean values and uncertainties.

Result: Extensive experiments on four real-world datasets from Shenzhen, NYC, and Chicago show UQGNN outperforms state-of-the-art baselines in both prediction accuracy and uncertainty quantification, achieving 5% improvement on Shenzhen dataset.

Conclusion: UQGNN effectively addresses the research gap by providing uncertainty-aware multivariate spatiotemporal prediction that captures complex interactions between different urban phenomena, demonstrating superior performance across multiple real-world datasets.

Abstract: Spatiotemporal prediction plays a critical role in numerous real-world applications such as urban planning, transportation optimization, disaster response, and pandemic control. In recent years, researchers have made significant progress by developing advanced deep learning models for spatiotemporal prediction. However, most existing models are deterministic, i.e., predicting only the expected mean values without quantifying uncertainty, leading to potentially unreliable and inaccurate outcomes. While recent studies have introduced probabilistic models to quantify uncertainty, they typically focus on a single phenomenon (e.g., taxi, bike, crime, or traffic crashes), thereby neglecting the inherent correlations among heterogeneous urban phenomena. To address the research gap, we propose a novel Graph Neural Network with Uncertainty Quantification, termed UQGNN for multivariate spatiotemporal prediction. UQGNN introduces two key innovations: (i) an Interaction-aware Spatiotemporal Embedding Module that integrates a multivariate diffusion graph convolutional network and an interaction-aware temporal convolutional network to effectively capture complex spatial and temporal interaction patterns, and (ii) a multivariate probabilistic prediction module designed to estimate both expected mean values and associated uncertainties. Extensive experiments on four real-world multivariate spatiotemporal datasets from Shenzhen, New York City, and Chicago demonstrate that UQGNN consistently outperforms state-of-the-art baselines in both prediction accuracy and uncertainty quantification. For example, on the Shenzhen dataset, UQGNN achieves a 5% improvement in both prediction accuracy and uncertainty quantification.

[1019] Explainable post-training bias mitigation with distribution-based fairness metrics

Ryan Franks, Alexey Miroshnikov, Konstandinos Kotsiopoulos

Main category: cs.LG

TL;DR: A novel optimization framework for post-processing models to achieve distribution-based fairness constraints without retraining, particularly effective for gradient-boosted decision trees.

DetailsMotivation: To efficiently produce demographically blind and explainable models across various fairness levels without the computational cost of retraining existing models.

Method: Uses stochastic gradient descent-based optimization framework for post-processing, applicable to various model types with emphasis on gradient-boosted decision trees. Includes interpretable global bias metrics.

Result: Empirically tested on various datasets and compared favorably against other methods, demonstrating effectiveness in achieving fairness constraints through post-processing.

Conclusion: The framework successfully enables efficient production of fair and explainable models across different fairness levels through post-processing, avoiding the need for model retraining while maintaining interpretability.

Abstract: We develop a novel optimization framework with distribution-based fairness constraints for efficiently producing demographically blind, explainable models across a wide range of fairness levels. This is accomplished through post-processing, avoiding the need for retraining. Our framework, which is based on stochastic gradient descent, can be applied to a wide range of model types, with a particular emphasis on the post-processing of gradient-boosted decision trees. Additionally, we design a broad class of interpretable global bias metrics compatible with our method by building on previous work. We empirically test our methodology on a variety of datasets and compare it to other methods.

[1020] Optimal Control of Probabilistic Dynamics Models via Mean Hamiltonian Minimization

David Leeftink, Çağatay Yıldız, Steffen Ridderbusch, Max Hinne, Marcel van Gerven

Main category: cs.LG

TL;DR: Probabilistic Hamiltonian approach for optimal control with learned dynamics models that minimizes mean Hamiltonian under epistemic uncertainty

DetailsMotivation: Optimal control of non-linear continuous-time systems requires careful treatment under epistemic uncertainty when exact system dynamics are unknown

Method: Translate probabilistic Pontryagin maximum principle to minimize mean Hamiltonian with respect to posterior distribution over system dynamics, using multiple shooting numerical method scalable to large probabilistic models including ensemble neural ODEs

Result: Reduced trial costs in offline model-based RL tasks and competitive performance in online scenarios compared to other baselines

Conclusion: Bridges optimal control and reinforcement learning, providing principled and practical framework for controlling uncertain systems with learned dynamics

Abstract: Without exact knowledge of the true system dynamics, optimal control of non-linear continuous-time systems requires careful treatment under epistemic uncertainty. In this work, we translate a probabilistic interpretation of the Pontryagin maximum principle to the challenge of optimal control with learned probabilistic dynamics models. Our framework provides a principled treatment of epistemic uncertainty by minimizing the mean Hamiltonian with respect to a posterior distribution over the system dynamics. We propose a multiple shooting numerical method that leverages mean Hamiltonian minimization and is scalable to large-scale probabilistic dynamics models, including ensemble neural ordinary differential equations. Comparisons against other baselines in online and offline model-based reinforcement learning tasks show that our probabilistic Hamiltonian approach leads to reduced trial costs in offline settings and achieves competitive performance in online scenarios. By bridging optimal control and reinforcement learning, our approach offers a principled and practical framework for controlling uncertain systems with learned dynamics.

[1021] Gaussian Mixture Flow Matching Models

Hansheng Chen, Kai Zhang, Hao Tan, Zexiang Xu, Fujun Luan, Leonidas Guibas, Gordon Wetzstein, Sai Bi

Main category: cs.LG

TL;DR: GMFlow proposes Gaussian mixture flow matching to improve few-step sampling and address over-saturation issues in diffusion models by predicting multi-modal flow velocity distributions instead of single Gaussian means.

DetailsMotivation: Current diffusion and flow matching models underperform in few-step sampling due to discretization error and produce over-saturated colors under classifier-free guidance, limiting their practical efficiency and quality.

Method: GMFlow predicts dynamic Gaussian mixture parameters to capture multi-modal flow velocity distribution using KL divergence loss, generalizes previous models, and introduces GM-SDE/ODE solvers with analytic denoising distributions for precise sampling.

Result: GMFlow achieves state-of-the-art performance with Precision of 0.942 using only 6 sampling steps on ImageNet 256x256, consistently outperforming flow matching baselines in generation quality.

Conclusion: The proposed Gaussian mixture flow matching framework effectively addresses limitations of existing models, providing superior few-step sampling performance and mitigating over-saturation issues through probabilistic guidance.

Abstract: Diffusion models approximate the denoising distribution as a Gaussian and predict its mean, whereas flow matching models reparameterize the Gaussian mean as flow velocity. However, they underperform in few-step sampling due to discretization error and tend to produce over-saturated colors under classifier-free guidance (CFG). To address these limitations, we propose a novel Gaussian mixture flow matching (GMFlow) model: instead of predicting the mean, GMFlow predicts dynamic Gaussian mixture (GM) parameters to capture a multi-modal flow velocity distribution, which can be learned with a KL divergence loss. We demonstrate that GMFlow generalizes previous diffusion and flow matching models where a single Gaussian is learned with an $L_2$ denoising loss. For inference, we derive GM-SDE/ODE solvers that leverage analytic denoising distributions and velocity fields for precise few-step sampling. Furthermore, we introduce a novel probabilistic guidance scheme that mitigates the over-saturation issues of CFG and improves image generation quality. Extensive experiments demonstrate that GMFlow consistently outperforms flow matching baselines in generation quality, achieving a Precision of 0.942 with only 6 sampling steps on ImageNet 256$\times$256.

[1022] Hierarchical Evaluation Function: A Multi-Metric Approach for Optimizing Demand Forecasting Models

Adolfo González, Víctor Parada

Main category: cs.LG

TL;DR: Proposes Hierarchical Evaluation Function (HEF) that combines R2, MAE, and RMSE with adaptive penalties to improve demand forecasting model evaluation, outperforming traditional MAE in complex environments.

DetailsMotivation: Traditional metrics like MAE and RMSE used in isolation can lead to biased evaluations and limited model robustness in inventory management with demand uncertainty and constraints.

Method: Developed HEF composite function integrating R2, MAE, and RMSE with hierarchical framework and adaptive penalties. Tested with Grid Search, PSO, and Optuna optimization on Walmart, M3, M4, and M5 datasets.

Result: HEF consistently outperformed MAE in global metrics (R2, Global Relative Precision, RMSE, RMSSE), improving explanatory power and stability against extreme errors, though MAE remained simpler and more computationally efficient.

Conclusion: HEF provides a robust, adaptive alternative for model selection and hyperparameter optimization in highly variable environments, offering a solid framework for demand forecasting.

Abstract: Inventory management in dynamic and competitive business environments presents multidimensional challenges, particularly in the face of demand uncertainty and logistical and financial constraints. In this context, accurate demand forecasting is critical for optimizing resources and anticipating market fluctuations. However, the isolated use of traditional metrics such as Mean Absolute Error (MAE) or Root Mean Squared Error (RMSE) can lead to biased evaluations and limit model robustness. To address this limitation, we propose the Hierarchical Evaluation Function (HEF), a composite function that integrates R2, MAE, and RMSE under a hierarchical and dynamic framework, complemented by adaptive penalties. The study implements HEF in the optimization of multiple prediction models, applying Grid Search, Particle Swarm Optimization (PSO), and Optuna, and evaluating their performance on reference databases (Walmart, M3, M4, and M5). The results, validated using statistical tests, confirm that HEF consistently outperforms the MAE used as the evaluation function in global metrics such as R2, Global Relative Precision, RMSE, and RMSSE, improving explanatory power and stability against extreme errors. In contrast, the MAE retains advantages in simplicity and computational efficiency. In summary, HEF constitutes a robust and adaptive alternative for highly variable environments, providing a solid framework for model selection and hyperparameter optimization.

[1023] Responsible Machine Learning via Mixed-Integer Optimization

Nathan Justin, Qingshi Sun, Andrés Gómez, Phebe Vayanos

Main category: cs.LG

TL;DR: This tutorial paper introduces mixed-integer optimization (MIO) as a framework for building responsible machine learning models that address fairness, transparency, and robustness concerns while maintaining performance.

DetailsMotivation: As ML systems are deployed in increasingly critical and sensitive domains affecting individuals and society, there is a growing need for responsible ML methods that address fairness, transparency, and robustness concerns while providing guaranteed performance.

Method: The paper proposes using mixed-integer optimization (MIO) to embed responsible ML considerations directly into the learning process, enabling the creation of inherently transparent models that can incorporate fairness constraints and other domain-specific requirements.

Result: MIO provides a powerful framework for building responsible ML models that align with core principles of responsible ML while maintaining performance, with practical strategies and tools available for efficiently solving MIO problems.

Conclusion: The paper concludes by discussing current limitations and open research questions, providing suggestions for future work in using MIO for responsible machine learning development.

Abstract: In the last few decades, Machine Learning (ML) has achieved significant success across domains ranging from healthcare, sustainability, and the social sciences, to criminal justice and finance. But its deployment in increasingly sophisticated, critical, and sensitive areas affecting individuals, the groups they belong to, and society as a whole raises critical concerns around fairness, transparency and robustness, among others. As the complexity and scale of ML systems and of the settings in which they are deployed grow, so does the need for responsible ML methods that address these challenges while providing guaranteed performance in deployment. Mixed-integer optimization (MIO) offers a powerful framework for embedding responsible ML considerations directly into the learning process while maintaining performance. For example, it enables learning of inherently transparent models that can conveniently incorporate fairness or other domain specific constraints. This tutorial paper provides an accessible and comprehensive introduction to this topic discussing both theoretical and practical aspects. It outlines some of the core principles of responsible ML, their importance in applications, and the practical utility of MIO for building ML models that align with these principles. Through examples and mathematical formulations, it illustrates practical strategies and available tools for efficiently solving MIO problems for responsible ML. It concludes with a discussion on current limitations and open research questions, providing suggestions for future work.

[1024] Adaptively Robust LLM Inference Optimization under Prediction Uncertainty

Zixi Chen, Yinyu Ye, Zijie Zhou

Main category: cs.LG

TL;DR: Optimizing LLM inference scheduling with ML-based output length prediction to minimize latency and prevent memory overflow, using adaptive algorithms that achieve log-scale competitive ratio.

DetailsMotivation: LLM inference is an online multi-task service that consumes significant energy, with unknown output lengths making scheduling challenging. Efficient scheduling is crucial to reduce latency and power consumption while handling high request volumes.

Method: Proposed two algorithms: A_max (conservative approach using upper bound predictions) and A_min (adaptive algorithm using lower bound predictions with dynamic refinement during inference). Uses ML to predict output length intervals.

Result: A_min achieves log-scale competitive ratio and performs nearly as well as hindsight scheduler in simulations. It’s more robust than A_max, especially when prediction accuracy decreases, and relies only on lower bounds which are easier to predict accurately.

Conclusion: The adaptive algorithm A_min provides efficient and robust LLM inference scheduling by leveraging ML predictions of output length intervals, achieving near-optimal performance while preventing memory overflow with practical design advantages.

Abstract: We study the problem of optimizing Large Language Model (LLM) inference scheduling to minimize total latency. LLM inference is an online and multi-task service process and also heavily energy consuming by which a pre-trained LLM processes input requests and generates output tokens sequentially. Therefore, it is vital to improve its scheduling efficiency and reduce the power consumption while a great amount of prompt requests are arriving. A key challenge in LLM inference scheduling is that while the prompt length is known upon arrival, the output length, which critically impacts memory usage and processing time, is unknown. To address this uncertainty, we propose algorithms that leverage machine learning to predict output lengths, assuming the prediction provides an interval classification (min-max range) for each request. We first design a conservative algorithm, $\mathcal{A}{\max}$, which schedules requests based on the upper bound of predicted output lengths to prevent memory overflow. However, this approach is overly conservative: as prediction accuracy decreases, performance degrades significantly due to potential overestimation. To overcome this limitation, we propose $\mathcal{A}{\min}$, an adaptive algorithm that initially treats the predicted lower bound as the output length and dynamically refines this estimate during inferencing. We prove that $\mathcal{A}{\min}$ achieves a log-scale competitive ratio. Through numerical simulations, we demonstrate that $\mathcal{A}{\min}$ often performs nearly as well as the hindsight scheduler, highlighting both its efficiency and robustness in practical scenarios. Moreover, $\mathcal{A}_{\min}$ relies solely on the lower bound of the prediction interval–an advantageous design choice since upper bounds on output length are typically more challenging to predict accurately.

[1025] A Causality- and Frequency-Aware Deep Learning Framework for Wave Elevation Prediction Behind Floating Breakwaters

Jianxin Zhang, Lianzi Jiang, Xinyu Han, Xiangrong Wang

Main category: cs.LG

TL;DR: E2E-FANet is a novel neural network that improves wave elevation prediction behind floating breakwaters by incorporating frequency-aware modeling and causal relationships between wave-structure interactions.

DetailsMotivation: Existing deep learning approaches have limited generalization capability under unseen operating conditions for predicting nonlinear wave fields behind floating breakwaters, which is crucial for coastal engineering optimization and safety.

Method: Proposes E2E-FANet with three key components: 1) Dual-Basis Frequency Mapping module using orthogonal cosine/sine bases for adaptive time-frequency representation, 2) Exogenous-to-Endogenous Cross-Attention module to model causal influence of breakwater motion on waves, and 3) Temporal-wise Attention mechanism for capturing complex dependencies.

Result: Extensive experiments show E2E-FANet achieves superior predictive accuracy and robust generalization across diverse wave conditions and varying relative water density conditions compared to mainstream models.

Conclusion: The work emphasizes the importance of integrating causality and frequency-aware modeling in deep learning architectures for modeling nonlinear dynamic systems in coastal engineering applications.

Abstract: Predicting the elevations of nonlinear wave fields behind floating breakwaters (FBs) is crucial for optimizing coastal engineering structures, enhancing safety, and improving design efficiency. Existing deep learning approaches exhibit limited generalization capability under unseen operating conditions. To address this challenge, this study proposes the Exogenous-to-Endogenous Frequency-Aware Network (E2E-FANet), a novel end-to-end neural network designed to model relationships between waves and structures. First, the Dual-Basis Frequency Mapping (DBFM) module leverages orthogonal cosine and sine bases to generate an adaptive time-frequency representation, enabling the model to effectively disentangle the evolving spectral components of wave signals. Second, the Exogenous-to-Endogenous Cross-Attention (E2ECA) module employs cross attention to explicitly model the unidirectional causal influence of floating breakwater motion on wave elevations. Additionally, a Temporal-wise Attention (TA) mechanism is incorporated that adaptively captures complex dependencies in endogenous variables. Extensive experiments, including generalization tests across diverse wave conditions and adaptability tests under varying relative water density (RW) conditions, demonstrate that E2E-FANet achieves superior predictive accuracy and robust generalization compared to mainstream models. This work emphasizes the importance of integrating causality and frequency-aware modeling in deep learning architectures for modeling nonlinear dynamics systems.

[1026] Bridging Generalization and Personalization in Wearable Human Activity Recognition via On-Device Few-Shot Learning

Pixi Kang, Julian Moosmann, Mengxi Liu, Bo Zhou, Michele Magno, Paul Lukowicz, Sizhen Bian

Main category: cs.LG

TL;DR: A novel on-device few-shot learning framework for wearable human activity recognition that bridges generalization across users and personalization for individuals, achieving significant accuracy improvements with minimal computation on resource-constrained devices.

DetailsMotivation: Conventional HAR models fail to generalize across diverse users and struggle with user-specific variations, leading to degraded performance. There's a need for solutions that can both generalize well and personalize efficiently for individual users on wearable devices.

Method: Proposes a two-stage approach: first trains a generalizable representation across users, then rapidly adapts to new users with few labeled samples by updating lightweight classifier layers directly on resource-constrained devices (RISC-V GAP9 microcontroller).

Result: Achieved accuracy improvements of 3.73% on RecGym, 17.38% on QVAR-Gesture, and 3.70% on Ultrasound-Gesture datasets through post-deployment adaptation. Demonstrates robust on-device learning with minimal computation and memory costs.

Conclusion: Few-shot on-device learning enables scalable, user-aware, and energy-efficient wearable human activity recognition by seamlessly uniting generalization and personalization, making it practical for real-world deployment on resource-constrained devices.

Abstract: Human Activity Recognition (HAR) with wearable devices requires both strong generalization across diverse users and efficient personalization for individuals. However, conventional HAR models often fail to generalize when faced with user-specific variations, leading to degraded performance. To address this challenge, we propose a novel on-device few-shot learning framework that bridges generalization and personalization in wearable HAR. Our method first trains a generalizable representation across users and then rapidly adapts to new users with only a few labeled samples, updating lightweight classifier layers directly on resource-constrained devices. This approach achieves robust on-device learning with minimal computation and memory cost, making it practical for real-world deployment. We implement our framework on the energy-efficient RISC-V GAP9 microcontroller and evaluate it on three benchmark datasets (RecGym, QVAR-Gesture, Ultrasound-Gesture). Across these scenarios, post-deployment adaptation improves accuracy by 3.73%, 17.38%, and 3.70%, respectively. These results demonstrate that few-shot on-device learning enables scalable, user-aware, and energy-efficient wearable human activity recognition by seamlessly uniting generalization and personalization \footnote{https://github.com/kangpx/onlineTiny2023}.

[1027] Symmetry-Breaking Descent for Invariant Cost Functionals

Mikhail Osipov

Main category: cs.LG

TL;DR: A variational method for optimizing discontinuous G-invariant cost functionals using symmetry-breaking deformations without gradients or training labels.

DetailsMotivation: Many machine learning tasks involve optimizing cost functionals that are invariant under symmetry groups but may be discontinuous or non-differentiable, making gradient-based methods inapplicable.

Method: Construct explicit deformations via a control field that minimizes an auxiliary energy functional, generating flows normal to the G-orbit to cross decision boundaries of invariant costs.

Result: Shows that symmetry-breaking deformations can reduce cost under mild conditions, with both geometric and weakly-coupled variants analyzed.

Conclusion: Provides a principled, gradient-free approach for optimizing discontinuous invariant functionals using Lie-algebraic variational flows at test time.

Abstract: We study the problem of reducing a task cost functional $W : H^s(M) \to \mathbb{R}$, not assumed continuous or differentiable, defined over Sobolev-class signals $S \in H^s(M) $, in the presence of a global symmetry group $G \subset \mathrm{Diff}(M)$. The group acts on signals by pullback, and the cost $W$ is invariant under this action. Such scenarios arise in machine learning and related optimization tasks, where performance metrics may be discontinuous or model-internal. We propose a variational method that exploits the symmetry structure to construct explicit deformations of the input signal. A deformation control field $ \phi: M \to \mathbb R^d$, obtained by minimizing an auxiliary energy functional, induces a flow that generically lies in the normal space (with respect to the $L^2$ inner product) to the $G$-orbit of $S$, and hence is a natural candidate to cross the decision boundary of the $G $-invariant cost. We analyze two variants of the coupling term: (1) purely geometric, independent of $W$, and (2) weakly coupled to $W$. Under mild conditions, we show that symmetry-breaking deformations of the signal can reduce the cost. Our approach requires no gradient backpropagation or training labels and operates entirely at test time. It provides a principled tool for optimizing discontinuous invariant cost functionals via Lie-algebraic variational flows.

[1028] Bigger Isn’t Always Memorizing: Early Stopping Overparameterized Diffusion Models

Alessandro Favero, Antonio Sclocchi, Matthieu Wyart

Main category: cs.LG

TL;DR: Diffusion models achieve generalization before memorization in overparameterized regimes, with memorization time proportional to dataset size, enabling effective early stopping for optimal generalization.

DetailsMotivation: To understand how diffusion models generalize rather than just memorize training data, especially in overparameterized settings where memorization would be expected.

Method: Analyzed diffusion models across image and language domains, plus a simple probabilistic context-free grammar model, tracking generalization vs memorization dynamics during training.

Result: Found that generalization occurs before memorization, with memorization time scaling linearly with dataset size. Created a phase diagram showing this competition between time scales.

Conclusion: Early stopping based on dataset size can optimize generalization while preventing memorization, with implications for hyperparameter transfer and privacy-sensitive applications.

Abstract: Diffusion probabilistic models have become a cornerstone of modern generative AI, yet the mechanisms underlying their generalization remain poorly understood. In fact, if these models were perfectly minimizing their training loss, they would just generate data belonging to their training set, i.e., memorize, as empirically found in the overparameterized regime. We revisit this view by showing that, in highly overparameterized diffusion models, generalization in natural data domains is progressively achieved during training before the onset of memorization. Our results, ranging from image to language diffusion models, systematically support the empirical law that memorization time is proportional to the dataset size. Generalization vs. memorization is then best understood as a competition between time scales. We show that this phenomenology is recovered in diffusion models learning a simple probabilistic context-free grammar with random rules, where generalization corresponds to the hierarchical acquisition of deeper grammar rules as training time grows, and the generalization cost of early stopping can be characterized. We summarize these results in a phase diagram. Overall, our results support that a principled early-stopping criterion - scaling with dataset size - can effectively optimize generalization while avoiding memorization, with direct implications for hyperparameter transfer and privacy-sensitive applications.

[1029] Tri-Accel: Curvature-Aware Precision-Adaptive and Memory-Elastic Optimization for Efficient GPU Usage

Mohsen Sheibanian, Pouya Shaeri, Alimohammad Beigi, Ryan T. Woo, Aryan Keluskar

Main category: cs.LG

TL;DR: Tri-Accel is a unified optimization framework that co-adapts three acceleration strategies (precision adaptation, sparse second-order signals, and memory-elastic batch scaling) to reduce training time by 9.9% and memory usage by 13.3% while improving accuracy.

DetailsMotivation: Deep neural networks are increasingly bottlenecked by optimization costs in terms of GPU memory and compute time, with existing acceleration techniques typically used in isolation rather than synergistically.

Method: Tri-Accel combines: (1) Precision-Adaptive Updates that dynamically assign mixed-precision levels based on curvature and gradient variance; (2) Sparse Second-Order Signals using Hessian/Fisher sparsity patterns; (3) Memory-Elastic Batch Scaling that adjusts batch size in real time according to VRAM availability.

Result: Achieves up to 9.9% reduction in training time, 13.3% lower memory usage, and +1.1 percentage point accuracy improvement over FP32 baselines on CIFAR-10 with ResNet-18 and EfficientNet-B0. Maintains 78.1% accuracy while reducing memory footprint from 0.35GB to 0.31GB.

Conclusion: The framework demonstrates how algorithmic adaptivity and hardware awareness can be combined to improve scalability in resource-constrained settings, enabling more efficient neural network training on edge devices and cost-sensitive cloud deployments without manual hyperparameter tuning.

Abstract: Deep neural networks are increasingly bottlenecked by the cost of optimization, both in terms of GPU memory and compute time. Existing acceleration techniques, such as mixed precision, second-order methods, and batch size scaling, are typically used in isolation. We present Tri-Accel, a unified optimization framework that co-adapts three acceleration strategies along with adaptive parameters during training: (1) Precision-Adaptive Updates that dynamically assign mixed-precision levels to layers based on curvature and gradient variance; (2) Sparse Second-Order Signals that exploit Hessian/Fisher sparsity patterns to guide precision and step size decisions; and (3) Memory-Elastic Batch Scaling that adjusts batch size in real time according to VRAM availability. On CIFAR-10 with ResNet-18 and EfficientNet-B0, Tri-Accel achieves up to 9.9% reduction in training time and 13.3% lower memory usage, while improving accuracy by +1.1 percentage points over FP32 baselines. Tested on CIFAR-10/100, our approach demonstrates adaptive learning behavior, with efficiency gradually improving over the course of training as the system learns to allocate resources more effectively. Compared to static mixed-precision training, Tri-Accel maintains 78.1% accuracy while reducing memory footprint from 0.35GB to 0.31GB on standard hardware. The framework is implemented with custom Triton kernels, whose hardware-aware adaptation enables automatic optimization without manual hyperparameter tuning, making it practical for deployment across diverse computational environments. This work demonstrates how algorithmic adaptivity and hardware awareness can be combined to improve scalability in resource-constrained settings, paving the way for more efficient neural network training on edge devices and cost-sensitive cloud deployments.

[1030] PreGenie: An Agentic Framework for High-quality Visual Presentation Generation

Xiaojie Xu, Xinli Xu, Sirui Chen, Haoyu Chen, Fan Zhang, Ying-Cong Chen

Main category: cs.LG

TL;DR: PreGenie is an agentic framework using multimodal LLMs to generate high-quality visual presentations through iterative analysis and regeneration, addressing layout, text summarization, and image-text matching issues.

DetailsMotivation: Existing automated presentation generation methods suffer from poor layouts, inaccurate text summarization, and mismatched visuals with text, limiting their use in formal business and scientific contexts.

Method: A two-stage modular framework built on Slidev: (1) Analysis and Initial Generation summarizes multimodal input and generates initial Markdown code, (2) Review and Re-generation iteratively reviews code and slides using multiple collaborating MLLMs to produce final presentations.

Result: PreGenie outperforms existing models in multimodal understanding, achieving better aesthetics, content consistency, and alignment with human design preferences.

Conclusion: The proposed agentic framework successfully addresses key limitations in automated presentation generation, producing high-quality outputs suitable for formal applications through collaborative MLLM-based iterative refinement.

Abstract: Visual presentations are vital for effective communication. Early attempts to automate their creation using deep learning often faced issues such as poorly organized layouts, inaccurate text summarization, and a lack of image understanding, leading to mismatched visuals and text. These limitations restrict their application in formal contexts like business and scientific research. To address these challenges, we propose PreGenie, an agentic and modular framework powered by multimodal large language models (MLLMs) for generating high-quality visual presentations. PreGenie is built on the Slidev presentation framework, where slides are rendered from Markdown code. It operates in two stages: (1) Analysis and Initial Generation, which summarizes multimodal input and generates initial code, and (2) Review and Re-generation, which iteratively reviews intermediate code and rendered slides to produce final, high-quality presentations. Each stage leverages multiple MLLMs that collaborate and share information. Comprehensive experiments demonstrate that PreGenie excels in multimodal understanding, outperforming existing models in both aesthetics and content consistency, while aligning more closely with human design preferences.

[1031] ONG: Orthogonal Natural Gradient Descent

Yajat Yadav, Patrick Mendoza, Jathin Korrapati

Main category: cs.LG

TL;DR: Proposes Orthogonal Natural Gradient Descent (ONG) which combines natural gradients with orthogonal projections for continual learning, but finds naive combination has issues requiring further research.

DetailsMotivation: Orthogonal Gradient Descent (OGD) uses Euclidean projections that don't leverage information-geometric structure, leading to suboptimal convergence in continual learning tasks.

Method: Incorporates natural gradient into OGD using EKFAC approximation of inverse Fisher information matrix, with orthogonal projections onto prior tasks’ gradient complements.

Result: Preliminary results on Permuted and Rotated MNIST benchmarks show that naive combination of natural gradients and orthogonal projections can have potential issues.

Conclusion: Findings motivate continued work on robustly reconciling geometric perspectives, establishing rigorous theoretical foundation with convergence guarantees, and extending to large-scale benchmarks.

Abstract: Orthogonal Gradient Descent (OGD) has emerged as a powerful method for continual learning. However, its Euclidean projections do not leverage the underlying information-geometric structure of the problem, which can lead to suboptimal convergence in learning tasks. To address this, we propose incorporating the natural gradient into OGD and present \textbf{ONG (Orthogonal Natural Gradient Descent)}. ONG preconditions each new task-specific gradient with an efficient EKFAC approximation of the inverse Fisher information matrix, yielding updates that follow the steepest descent direction under a Riemannian metric. To preserve performance on previously learned tasks, ONG projects these natural gradients onto the orthogonal complement of prior tasks’ gradients. We provide an initial theoretical justification for this procedure, introduce the Orthogonal Natural Gradient Descent (ONG) algorithm, and present preliminary results on the Permuted and Rotated MNIST benchmarks. Our preliminary results, however, indicate that a naive combination of natural gradients and orthogonal projections can have potential issues. This finding motivates continued future work focused on robustly reconciling these geometric perspectives to develop a continual learning method, establishing a more rigorous theoretical foundation with formal convergence guarantees, and extending empirical validation to large-scale continual learning benchmarks. The anonymized version of our code can be found as the zip file here: https://drive.google.com/drive/folders/11PyU6M8pNgOUB5pwdGORtbnMtD8Shiw_?usp=sharing.

[1032] History-Aware Neural Operator: Robust Data-Driven Constitutive Modeling of Path-Dependent Materials

Binyao Guo, Zihan Lin, QiZhi He

Main category: cs.LG

TL;DR: HANO is a novel neural operator framework for path-dependent material modeling that uses hierarchical self-attention and Fourier-based architecture to overcome RNN limitations, achieving superior accuracy and robustness in simulating inelastic materials.

DetailsMotivation: To address self-consistency issues and hidden state sensitivity in recurrent neural network models for path-dependent material modeling, and to create a more robust data-driven approach that can handle complex loading conditions without relying on hidden state variables.

Method: Developed History-Aware Neural Operator (HANO) - an autoregressive model with Fourier-based neural operator backbone and hierarchical self-attention mechanism for multiscale feature extraction, enabling discretization-invariant learning from short strain-stress history segments.

Result: HANO consistently outperformed baseline models in predictive accuracy, generalization, and robustness across various conditions including irregular sampling, multi-cycle loading, noisy data, and pre-stressed states. Demonstrated effectiveness on elastoplasticity with hardening and progressive anisotropic damage problems.

Conclusion: HANO provides an effective data-driven surrogate for simulating inelastic materials with robust performance under complex conditions, making it well-suited for integration with classical numerical solvers while overcoming limitations of traditional RNN-based approaches.

Abstract: This study presents an end-to-end learning framework for data-driven modeling of path-dependent inelastic materials using neural operators. The framework is built on the premise that irreversible evolution of material responses, governed by hidden dynamics, can be inferred from observable data. We develop the History-Aware Neural Operator (HANO), an autoregressive model that predicts path-dependent material responses from short segments of recent strain-stress history without relying on hidden state variables, thereby overcoming self-consistency issues commonly encountered in recurrent neural network (RNN)-based models. Built on a Fourier-based neural operator backbone, HANO enables discretization-invariant learning. To enhance its ability to capture both global loading patterns and critical local path dependencies, we embed a hierarchical self-attention mechanism that facilitates multiscale feature extraction. Beyond ensuring self-consistency, HANO mitigates sensitivity to initial hidden states, a commonly overlooked issue that can lead to instability in recurrent models when applied to generalized loading paths. By modeling stress-strain evolution as a continuous operator rather than relying on fixed input-output mappings, HANO naturally accommodates varying path discretizations and exhibits robust performance under complex conditions, including irregular sampling, multi-cycle loading, noisy data, and pre-stressed states. We evaluate HANO on two benchmark problems: elastoplasticity with hardening and progressive anisotropic damage in brittle solids. Results show that HANO consistently outperforms baseline models in predictive accuracy, generalization, and robustness. With its demonstrated capabilities, HANO provides an effective data-driven surrogate for simulating inelastic materials and is well-suited for integration with classical numerical solvers.

[1033] Towards a Unified Textual Graph Framework for Spectral Reasoning via Physical and Chemical Information Fusion

Jiheng Liang, Ziru Yu, Zujie Xie, Yuchen Guo, Yulan Guo, Xiangyang Yu

Main category: cs.LG

TL;DR: A novel multi-modal spectral analysis framework that integrates prior knowledge graphs with LLMs to bridge physical spectral measurements and chemical structural semantics through a unified Textual Graph format.

DetailsMotivation: Current spectral analysis methods suffer from limitations including reliance on single-modality data, limited generalizability, and poor interpretability.

Method: Raw spectra are transformed into Textual Attribute Graphs (TAGs) with nodes and edges enriched with textual attributes describing spectral properties and chemical context. These are merged with prior knowledge to form Task Graphs with Prompt Nodes for LLM-based reasoning, processed by Graph Neural Networks.

Result: The framework achieves consistently high performance across multiple spectral analysis tasks (node-level, edge-level, graph-level classification) and demonstrates robust generalization in zero-shot and few-shot settings.

Conclusion: This work establishes a scalable and interpretable foundation for LLM-driven spectral analysis, unifying physical and chemical modalities for scientific applications.

Abstract: Motivated by the limitations of current spectral analysis methods-such as reliance on single-modality data, limited generalizability, and poor interpretability-we propose a novel multi-modal spectral analysis framework that integrates prior knowledge graphs with Large Language Models. Our method explicitly bridges physical spectral measurements and chemical structural semantics by representing them in a unified Textual Graph format, enabling flexible, interpretable, and generalizable spectral understanding. Raw spectra are first transformed into TAGs, where nodes and edges are enriched with textual attributes describing both spectral properties and chemical context. These are then merged with relevant prior knowledge-including functional groups and molecular graphs-to form a Task Graph that incorporates “Prompt Nodes” supporting LLM-based contextual reasoning. A Graph Neural Network further processes this structure to complete downstream tasks. This unified design enables seamless multi-modal integration and automated feature decoding with minimal manual annotation. Our framework achieves consistently high performance across multiple spectral analysis tasks, including node-level, edge-level, and graph-level classification. It demonstrates robust generalization in both zero-shot and few-shot settings, highlighting its effectiveness in learning from limited data and supporting in-context reasoning. This work establishes a scalable and interpretable foundation for LLM-driven spectral analysis, unifying physical and chemical modalities for scientific applications.

[1034] DrugReasoner: Interpretable Drug Approval Prediction with a Reasoning-augmented Language Model

Mohammadreza Ghaffarzadeh-Esfahani, Ali Motahharynia, Nahid Yousefian, Navid Mazrouei, Jafar Ghaisari, Yousof Gheisari

Main category: cs.LG

TL;DR: DrugReasoner is a reasoning-based LLM that predicts small-molecule drug approval likelihood by integrating molecular descriptors with comparative reasoning against similar compounds, achieving robust performance while providing interpretable rationales.

DetailsMotivation: Early prediction of drug approval outcomes is critical for optimizing research investments, but existing ML/DL methods have limited interpretability which constraints their impact in drug discovery.

Method: Built on LLaMA architecture and fine-tuned with group relative policy optimization (GRPO), DrugReasoner integrates molecular descriptors with comparative reasoning against structurally similar approved/unapproved compounds to generate predictions with step-by-step rationales and confidence scores.

Result: Achieved AUC of 0.732 and F1 score of 0.729 on validation set, 0.725 and 0.718 on test set, outperforming conventional baselines. On external dataset, achieved AUC of 0.728 and F1-score of 0.774, outperforming both baseline and ChemAP model.

Conclusion: DrugReasoner delivers competitive predictive accuracy while enhancing transparency through reasoning outputs, addressing a key bottleneck in AI-assisted drug discovery and demonstrating the potential of reasoning-augmented LLMs as interpretable tools for pharmaceutical decision-making.

Abstract: Drug discovery is a complex and resource-intensive process, making early prediction of approval outcomes critical for optimizing research investments. While classical machine learning and deep learning methods have shown promise in drug approval prediction, their limited interpretability constraints their impact. Here, we present DrugReasoner, a reasoning-based large language model (LLM) built on the LLaMA architecture and fine-tuned with group relative policy optimization (GRPO) to predict the likelihood of small-molecule approval. DrugReasoner integrates molecular descriptors with comparative reasoning against structurally similar approved and unapproved compounds, generating predictions alongside step-by-step rationales and confidence scores. DrugReasoner achieved robust performance with an AUC of 0.732 and an F1 score of 0.729 on the validation set and 0.725 and 0.718 on the test set, respectively. These results outperformed conventional baselines, including logistic regression, support vector machine, and k-nearest neighbors and had competitive performance relative to XGBoost. On an external independent dataset, DrugReasoner outperformed both baseline and the recently developed ChemAP model, achieving an AUC of 0.728 and an F1-score of 0.774, while maintaining high precision and balanced sensitivity, demonstrating robustness in real-world scenarios. These findings demonstrate that DrugReasoner not only delivers competitive predictive accuracy but also enhances transparency through its reasoning outputs, thereby addressing a key bottleneck in AI-assisted drug discovery. This study highlights the potential of reasoning-augmented LLMs as interpretable and effective tools for pharmaceutical decision-making.

[1035] Automating Traffic Monitoring with SHM Sensor Networks via Vision-Supervised Deep Learning

Hanshuo Wu, Xudong Jian, Christos Lataniotis, Cyprien Hoelzl, Eleni Chatzi, Yves Reuland

Main category: cs.LG

TL;DR: A deep learning pipeline using SHM sensors and GNNs achieves automated traffic monitoring with 99% accuracy for light vehicles and 94% for heavy vehicles, overcoming limitations of vision-based systems.

DetailsMotivation: Traditional traffic monitoring methods have limitations - CV-based approaches face privacy and lighting issues, while non-vision methods lack deployment flexibility. There's a need for automated, reliable monitoring to assess bridge service life.

Method: Proposes a fully automated deep-learning pipeline using SHM sensor networks with graph neural networks (GNNs). Integrates CV-assisted dataset generation with supervised training and inference to capture spatial structure and sensor interdependence.

Result: Achieves state-of-the-art performance with 99% classification accuracy for light vehicles and 94% for heavy vehicles using accelerometer and strain gauge data in real-world case study.

Conclusion: The framework successfully transfers knowledge from CV outputs to SHM sensors, enabling sensor networks to achieve vision-comparable accuracy with minimal human intervention for continuous traffic monitoring.

Abstract: Bridges, as critical components of civil infrastructure, are increasingly affected by deterioration, making reliable traffic monitoring essential for assessing their remaining service life. Among operational loads, traffic load plays a pivotal role, and recent advances in deep learning - particularly in computer vision (CV) - have enabled progress toward continuous, automated monitoring. However, CV-based approaches suffer from limitations, including privacy concerns and sensitivity to lighting conditions, while traditional non-vision-based methods often lack flexibility in deployment and validation. To bridge this gap, we propose a fully automated deep-learning pipeline for continuous traffic monitoring using structural health monitoring (SHM) sensor networks. Our approach integrates CV-assisted high-resolution dataset generation with supervised training and inference, leveraging graph neural networks (GNNs) to capture the spatial structure and interdependence of sensor data. By transferring knowledge from CV outputs to SHM sensors, the proposed framework enables sensor networks to achieve comparable accuracy of vision-based systems, with minimal human intervention. Applied to accelerometer and strain gauge data in a real-world case study, the model achieves state-of-the-art performance, with classification accuracies of 99% for light vehicles and 94% for heavy vehicles.

[1036] A Comparative Analysis of Reinforcement Learning and Conventional Deep Learning Approaches for Bearing Fault Diagnosis

Efe Çakır, Patrick Dumond

Main category: cs.LG

TL;DR: RL-based bearing fault diagnosis using DQNs shows comparable performance to traditional methods with better adaptability, though computationally intensive.

DetailsMotivation: Traditional bearing fault diagnosis methods require extensive labeled data and struggle with dynamic environments, prompting exploration of reinforcement learning for improved adaptability.

Method: Used Deep Q-Networks (DQNs) reinforcement learning approach with optimized reward structures for bearing fault classification in machine condition monitoring.

Result: RL models matched traditional supervised learning performance under controlled conditions and demonstrated superior adaptability with optimized reward structures, despite higher computational demands.

Conclusion: Reinforcement learning shows potential to complement traditional bearing fault diagnosis methods, enabling more adaptive diagnostic frameworks, though computational efficiency needs improvement.

Abstract: Bearing faults in rotating machinery can lead to significant operational disruptions and maintenance costs. Modern methods for bearing fault diagnosis rely heavily on vibration analysis and machine learning techniques, which often require extensive labeled data and may not adapt well to dynamic environments. This study explores the feasibility of reinforcement learning (RL), specifically Deep Q-Networks (DQNs), for bearing fault classification tasks in machine condition monitoring to enhance the accuracy and adaptability of bearing fault diagnosis. The results demonstrate that while RL models developed in this study can match the performance of traditional supervised learning models under controlled conditions, they excel in adaptability when equipped with optimized reward structures. However, their computational demands highlight areas for further improvement. These findings demonstrate RL’s potential to complement traditional methods, paving the way for adaptive diagnostic frameworks.

[1037] Lost in Latent Space: An Empirical Study of Latent Diffusion Models for Physics Emulation

François Rozet, Ruben Ohana, Michael McCabe, Gilles Louppe, François Lanusse, Shirley Ho

Main category: cs.LG

TL;DR: Latent-space diffusion models enable efficient physics emulation with 1000x compression while maintaining accuracy and outperforming non-generative methods.

DetailsMotivation: Reduce computational cost of diffusion models for physics emulation by using latent space instead of pixel space, similar to image/video generation approaches.

Method: Use autoencoder to compress dynamical systems data into latent space, then train diffusion models in this compressed space. Investigate various compression rates and practical design choices including architectures and optimizers.

Result: Latent-space emulation remains surprisingly accurate even at 1000x compression rates. Diffusion-based emulators are more accurate than non-generative counterparts and provide greater prediction diversity to compensate for uncertainty.

Conclusion: Latent-space diffusion models are effective for efficient physics emulation, offering significant computational savings without sacrificing accuracy, with proper architectural and optimization choices being critical for success.

Abstract: The steep computational cost of diffusion models at inference hinders their use as fast physics emulators. In the context of image and video generation, this computational drawback has been addressed by generating in the latent space of an autoencoder instead of the pixel space. In this work, we investigate whether a similar strategy can be effectively applied to the emulation of dynamical systems and at what cost. We find that the accuracy of latent-space emulation is surprisingly robust to a wide range of compression rates (up to 1000x). We also show that diffusion-based emulators are consistently more accurate than non-generative counterparts and compensate for uncertainty in their predictions with greater diversity. Finally, we cover practical design choices, spanning from architectures to optimizers, that we found critical to train latent-space emulators.

[1038] A Log-Linear Analytics Approach to Cost Model Regularization for Inpatient Stays through Diagnostic Code Merging

Chi-Ken Lu, David Alonge, Nicole Richardson, Bruno Richard

Main category: cs.LG

TL;DR: Reducing ICD-10 code granularity from 7 to 6+ characters improves OLS model stability and consistency while maintaining accuracy and interpretability in healthcare cost modeling.

DetailsMotivation: Interpretable healthcare cost models need to balance accuracy and parameter consistency, but OLS with granular ICD-10 codes suffers from coefficient instability due to infrequent codes, while regularization methods risk losing important predictors.

Method: Truncating ICD-10 codes from seven characters to six or fewer characters to reduce dimensionality while preserving diagnostic categories, leveraging the mathematical property that merging predictors increases the trace of the Hessian matrix and reduces coefficient estimation variance.

Result: The approach successfully addresses coefficient instability in OLS models while maintaining model interpretability and preserving all diagnostic code representations, explaining why broader diagnostic groupings are preferred in real-world applications.

Conclusion: Reducing ICD-10 code granularity serves as an effective regularization strategy within OLS frameworks, providing a practical solution for achieving stable, interpretable, and accurate healthcare cost models without discarding important predictors.

Abstract: Cost models in healthcare research must balance interpretability, accuracy, and parameter consistency. However, interpretable models often struggle to achieve both accuracy and consistency. Ordinary least squares (OLS) models for high-dimensional regression can be accurate but fail to produce stable regression coefficients over time when using highly granular ICD-10 diagnostic codes as predictors. This instability arises because many ICD-10 codes are infrequent in healthcare datasets. While regularization methods such as Ridge can address this issue, they risk discarding important predictors. Here, we demonstrate that reducing the granularity of ICD-10 codes is an effective regularization strategy within OLS while preserving the representation of all diagnostic code categories. By truncating ICD-10 codes from seven characters to six or fewer, we reduce the dimensionality of the regression problem while maintaining model interpretability and consistency. Mathematically, the merging of predictors in OLS leads to increased trace of the Hessian matrix, which reduces the variance of coefficient estimation. Our findings explain why broader diagnostic groupings like DRGs and HCC codes are favored over highly granular ICD-10 codes in real-world risk adjustment and cost models.

[1039] Graded Transformers

Tony Shaska Sr

Main category: cs.LG

TL;DR: Graded Transformer framework introduces algebraic inductive biases through grading transformations, offering two architectures (LGT and EGT) with rigorous theoretical guarantees and broad applications across multiple domains.

DetailsMotivation: To embed algebraic inductive biases into sequence models for improved hierarchical structure encoding, efficiency with structured data, and enabling adaptive feature prioritization through differentiable grading parameters.

Method: Proposes Linearly Graded Transformer (LGT) and Exponentially Graded Transformer (EGT) architectures that apply parameterized scaling operators with grading tuples and exponential factors to attention and representation layers.

Result: Establishes universal approximation theorems, reduced sample complexity via VC dimension bounds, Lipschitz continuity, robustness to perturbations, and gradient stability through graded loss optimization.

Conclusion: The Graded Transformer provides a mathematically principled framework for hierarchical learning and neuro-symbolic reasoning with applications spanning algebraic geometry, physics, NLP, biology, robotics, automotive AI, and cryptography.

Abstract: We introduce the Graded Transformer framework, a new class of sequence models that embeds algebraic inductive biases through grading transformations on vector spaces. Extending Graded Neural Networks (GNNs), we propose two architectures: the Linearly Graded Transformer (LGT) and the Exponentially Graded Transformer (EGT). These models apply parameterized scaling operators, governed by fixed or learnable grading tuples and in the case of EGT exponential factors, to encode hierarchical structure in attention and representation layers and to improve efficiency for structured data. We establish rigorous guarantees, including universal approximation theorems for continuous and Sobolev functions, reduced sample complexity via effective VC dimension bounds, Lipschitz continuity of graded operations, and robustness to perturbations. A graded loss ensures gradient stability and alignment with domain priors during optimization. By treating grades as differentiable parameters, the framework enables adaptive feature prioritization, overcoming limitations of fixed grades in earlier models. The Graded Transformer provides a mathematically principled approach to hierarchical learning and neuro-symbolic reasoning. Applications include algebraic geometry (moduli spaces and zeta functions), physics (multiscale systems), natural language processing (syntactic parsing), biological sequence analysis (variant prediction), robotics and autonomous systems (safety-critical prioritization), the automotive industry (certifiable AI for ADAS), and blockchain and financial cryptography (secure coding and structured prediction).

[1040] Toward Efficient Spiking Transformers: Synapse Pruning Meets Synergistic Learning-Based Compensation

Hongze Sun, Wuque Cai, Duo Chen, Shifeng Mao, Jiayi He, Zhenxing Wang, Dezhong Yao, Daqing Guo

Main category: cs.LG

TL;DR: This paper proposes lightweight spiking Transformer models using synapse pruning with synergistic learning compensation to reduce parameters and computational costs while maintaining performance.

DetailsMotivation: Existing spiking Transformer models require substantial parameters and high computational costs, limiting deployment in resource-constrained environments.

Method: Combines unstructured L1P pruning for sparse representations and structured DSP pruning for low-rank representations, plus enhanced sLIF neuron model with synergistic learning compensation.

Result: Significantly reduces model size and computational overhead while maintaining competitive performance on benchmark datasets.

Conclusion: The proposed pruning and compensation strategies effectively construct efficient and high-performing spiking Transformer-based models.

Abstract: As a foundational architecture of artificial intelligence models, Transformer has been recently adapted to spiking neural networks with promising performance across various tasks. However, existing spiking Transformer (ST)-based models require a substantial number of parameters and incur high computational costs, thus limiting their deployment in resource-constrained environments. To address these challenges, we propose combining synapse pruning with a synergistic learning-based compensation strategy to derive lightweight ST-based models. Specifically, two types of tailored pruning strategies are introduced to reduce redundancy in the weight matrices of ST blocks: an unstructured $\mathrm{L_{1}P}$ method to induce sparse representations, and a structured DSP method to induce low-rank representations. In addition, we propose an enhanced spiking neuron model, termed the synergistic leaky integrate-and-fire (sLIF) neuron, to effectively compensate for model pruning through synergistic learning between synaptic and intrinsic plasticity mechanisms. Extensive experiments on benchmark datasets demonstrate that the proposed methods significantly reduce model size and computational overhead while maintaining competitive performance. These results validate the effectiveness of the proposed pruning and compensation strategies in constructing efficient and high-performing ST-based models.

[1041] AnalogCoder-Pro: Unifying Analog Circuit Generation and Optimization via Multi-modal LLMs

Yao Lai, Souradip Poddar, Sungyoung Lee, Guojin Chen, Mengkang Hu, Bei Yu, Ping Luo, David Z. Pan

Main category: cs.LG

TL;DR: AnalogCoder-Pro is a multimodal LLM framework that automates analog circuit design through generative techniques, error diagnosis using simulation feedback, and reusable component libraries, outperforming existing methods on 13 circuit types.

DetailsMotivation: Current analog front-end design relies heavily on expert intuition and iterative simulations, limiting automation potential and requiring significant manual effort.

Method: Multimodal LLM framework integrating generative and optimization techniques with diagnosis-and-repair feedback loop using simulation error messages and waveform images, building reusable circuit tool library, and applying Bayesian optimization for device sizing.

Result: Successfully designed 28 circuits on a benchmark suite covering 13 circuit types and consistently outperformed existing LLM-based methods in figures of merit.

Conclusion: The framework enables end-to-end automation of analog circuit design from specifications to optimized implementations, demonstrating superior performance over current approaches.

Abstract: Despite recent advances, analog front-end design still relies heavily on expert intuition and iterative simulations, which limits the potential for automation. We present AnalogCoder-Pro, a multimodal large language model (LLM) framework that integrates generative and optimization techniques. The framework features a multimodal diagnosis-and-repair feedback loop that uses simulation error messages and waveform images to autonomously correct design errors. It also builds a reusable circuit tool library by archiving successful designs as modular subcircuits, accelerating the development of complex systems. Furthermore, it enables end-to-end automation by generating circuit topologies from target specifications, extracting key parameters, and applying Bayesian optimization for device sizing. On a curated benchmark suite covering 13 circuit types, AnalogCoder-Pro successfully designed 28 circuits and consistently outperformed existing LLM-based methods in figures of merit.

[1042] SolarSeer: Ultrafast and accurate 24-hour solar irradiance forecasts outperforming numerical weather prediction across the USA

Mingliang Bai, Zuliang Fang, Shengyu Tao, Siqi Xiang, Jiang Bian, Yanfei Xiang, Pengcheng Zhao, Weixin Jin, Jonathan A. Weyn, Haiyu Dong, Bin Zhang, Hongyu Sun, Kit Thambiratnam, Qi Zhang, Hongbin Sun, Xuan Zhang, Qiuwei Wu

Main category: cs.LG

TL;DR: SolarSeer is an AI model that provides 24-hour solar irradiance forecasts 1500x faster than traditional NWP methods with significantly improved accuracy.

DetailsMotivation: Traditional numerical weather prediction models are computationally expensive and rely on complex physics simulations, creating a need for more efficient solar forecasting methods for solar energy systems.

Method: End-to-end AI model that directly maps historical satellite observations to future forecasts, eliminating data assimilation and PDE solving processes used in traditional NWP.

Result: 27.28% reduction in RMSE in reanalysis data and 15.35% improvement across 1,800 stations compared to HRRR, with forecasts generated in under 3 seconds for CONUS at 5km resolution.

Conclusion: SolarSeer provides ultrafast, accurate solar irradiance forecasting that significantly outperforms traditional methods, supporting the transition to sustainable energy systems.

Abstract: Accurate 24-hour solar irradiance forecasting is essential for the safe and economic operation of solar photovoltaic systems. Traditional numerical weather prediction (NWP) models represent the state-of-the-art in forecasting performance but rely on computationally costly data assimilation and solving complicated partial differential equations (PDEs) that simulate atmospheric physics. Here, we introduce SolarSeer, an end-to-end large artificial intelligence (AI) model for solar irradiance forecasting across the Contiguous United States (CONUS). SolarSeer is designed to directly map the historical satellite observations to future forecasts, eliminating the computational overhead of data assimilation and PDEs solving. This efficiency allows SolarSeer to operate over 1,500 times faster than traditional NWP, generating 24-hour cloud cover and solar irradiance forecasts for the CONUS at 5-kilometer resolution in under 3 seconds. Compared with the state-of-the-art NWP in the CONUS, i.e., High-Resolution Rapid Refresh (HRRR), SolarSeer significantly reduces the root mean squared error of solar irradiance forecasting by 27.28% in reanalysis data and 15.35% across 1,800 stations. SolarSeer also effectively captures solar irradiance fluctuations and significantly enhances the first-order irradiance difference forecasting accuracy. SolarSeer’s ultrafast, accurate 24-hour solar irradiance forecasts provide strong support for the transition to sustainable, net-zero energy systems.

[1043] Multitask Learning with Stochastic Interpolants

Hugo Negrel, Florentin Coeurdoux, Michael S. Albergo, Eric Vanden-Eijnden

Main category: cs.LG

TL;DR: A framework generalizing flow and diffusion models using operator-based interpolants to bridge probability distributions across different dimensional spaces, enabling versatile generative models for multiple tasks without task-specific training.

DetailsMotivation: To create a unifying framework that generalizes the time dynamics of existing generative models (flow and diffusion models) and extends their capabilities to handle distributions across multiple dimensional spaces, enabling task-agnostic generative modeling.

Method: Generalize stochastic interpolants by replacing scalar time variables with vectors, matrices, or linear operators. This allows bridging probability distributions across different dimensional spaces and constructing versatile generative models that can perform multiple tasks without specialized training.

Result: The framework provides a unifying theoretical perspective for existing generative models and extends their capabilities. Numerical experiments demonstrate zero-shot efficacy on conditional generation, inpainting, fine-tuning, posterior sampling, and multiscale modeling.

Conclusion: The operator-based interpolant framework shows potential as a generic task-agnostic alternative to specialized models, offering versatile generative capabilities across multiple tasks without requiring task-specific training.

Abstract: We propose a framework for learning maps between probability distributions that broadly generalizes the time dynamics of flow and diffusion models. To enable this, we generalize stochastic interpolants by replacing the scalar time variable with vectors, matrices, or linear operators, allowing us to bridge probability distributions across multiple dimensional spaces. This approach enables the construction of versatile generative models capable of fulfilling multiple tasks without task-specific training. Our operator-based interpolants not only provide a unifying theoretical perspective for existing generative models but also extend their capabilities. Through numerical experiments, we demonstrate the zero-shot efficacy of our method on conditional generation and inpainting, fine-tuning and posterior sampling, and multiscale modeling, suggesting its potential as a generic task-agnostic alternative to specialized models.

[1044] Will You Be Aware? Eye Tracking-Based Modeling of Situational Awareness in Augmented Reality

Zhehan Qu, Tianyi Hu, Christian Fronk, Maria Gorlatova

Main category: cs.LG

TL;DR: AR systems can cause cognitive tunneling, compromising situational awareness in safety-critical tasks like CPR. This study developed an AR CPR app and used eye tracking to analyze SA, finding that higher SA correlates with specific gaze patterns. A novel graph neural network (FixGraphPool) achieved 83% accuracy in predicting SA from gaze data.

DetailsMotivation: Augmented Reality systems enhance task performance but risk inducing cognitive tunneling, where users hyperfocus on virtual content and lose situational awareness, particularly dangerous in safety-critical scenarios like cardiopulmonary resuscitation where responders must remain vigilant to unpredictable hazards.

Method: Developed an AR app on Magic Leap 2 providing real-time CPR feedback, conducted user study with simulated unexpected incidents, collected SA metrics via observation and questionnaires during freeze-probe events, and analyzed eye tracking data to identify gaze patterns associated with SA. Proposed FixGraphPool, a graph neural network that structures gaze events into spatiotemporal graphs to predict SA.

Result: Eye tracking revealed higher SA levels correlate with greater saccadic amplitude/velocity and reduced proportion/frequency of fixations on virtual content. The FixGraphPool model achieved 83.0% accuracy (F1=81.0%) in predicting SA, outperforming feature-based ML and state-of-the-art time-series models by effectively capturing dynamic attentional patterns from eye tracking data.

Conclusion: Eye tracking shows strong potential for situational awareness modeling in AR systems. The findings can inform the design of AR systems that maintain user safety and situational awareness, particularly in critical applications like medical emergencies where balancing task performance with environmental vigilance is essential.

Abstract: Augmented Reality (AR) systems, while enhancing task performance through real-time guidance, pose risks of inducing cognitive tunneling-a hyperfocus on virtual content that compromises situational awareness (SA) in safety-critical scenarios. This paper investigates SA in AR-guided cardiopulmonary resuscitation (CPR), where responders must balance effective compressions with vigilance to unpredictable hazards (e.g., patient vomiting). We developed an AR app on a Magic Leap 2 that overlays real-time CPR feedback (compression depth and rate) and conducted a user study with simulated unexpected incidents (e.g., bleeding) to evaluate SA, in which SA metrics were collected via observation and questionnaires administered during freeze-probe events. Eye tracking analysis revealed that higher SA levels were associated with greater saccadic amplitude and velocity, and with reduced proportion and frequency of fixations on virtual content. To predict SA, we propose FixGraphPool, a graph neural network that structures gaze events (fixations, saccades) into spatiotemporal graphs, effectively capturing dynamic attentional patterns. Our model achieved 83.0% accuracy (F1=81.0%), outperforming feature-based machine learning and state-of-the-art time-series models by leveraging domain knowledge and spatial-temporal information encoded in ET data. These findings demonstrate the potential of eye tracking for SA modeling in AR and highlight its utility in designing AR systems that ensure user safety and situational awareness.

[1045] RISE: Enhancing VLM Image Annotation with Self-Supervised Reasoning

Suhang Hu, Wei Hu, Yuhang Su, Fan Zhang

Main category: cs.LG

TL;DR: RISE is a two-stage framework that generates high-quality reasoning chains for VLMs through reinforcement learning and supervised fine-tuning, improving performance on complex image annotation tasks without manual CoT annotations.

DetailsMotivation: VLMs struggle with complex reasoning tasks like emotion classification and context-driven object detection. Standard SFT ignores reasoning rationales, while Visual-RFT produces inconsistent CoTs due to lack of verified reasoning chains during pre-training.

Method: Two-stage framework: 1) RISE-CoT uses reinforcement learning in an “annotation-reasoning-annotation” closed-loop to generate visually grounded CoTs by verifying reconstruction of original annotations. 2) RISE-R1 filters high-quality CoTs for supervised fine-tuning followed by reinforcement fine-tuning.

Result: RISE-trained Qwen2-VL-2B outperforms SFT and Visual-RFT on both complex and simple image annotation tasks, achieving robust performance and enhanced explainability.

Conclusion: RISE provides a self-supervised solution for advancing VLM reasoning capabilities without requiring manually annotated reasoning chains, enabling better performance on complex visual tasks.

Abstract: Vision-Language Models (VLMs) struggle with complex image annotation tasks, such as emotion classification and context-driven object detection, which demand sophisticated reasoning. Standard Supervised Fine-Tuning (SFT) focuses solely on annotation outcomes, ignoring underlying rationales, while Visual Reinforcement Fine-Tuning (Visual-RFT) produces inconsistent Chains of Thought (CoTs) due to the absence of high-quality, verified CoTs during pre-training. We introduce RISE (Reason-Inspire-Strengthen-Expertise), a two-stage framework to overcome these limitations. In the Reason stage (RISE-CoT), a reinforcement learning-driven “annotation-reasoning-annotation” closed-loop generates visually grounded, logically consistent CoTs by verifying their ability to reconstruct original annotations without direct leakage. The Inspire and Strengthen stage (RISE-R1) leverages a high-quality CoT subset, filtered by RISE-CoT rewards, for supervised fine-tuning, followed by reinforcement fine-tuning to produce interpretable reasoning and accurate annotations, achieving Expertise in complex visual tasks. Evaluated on complex and simple image annotation tasks, RISE-trained Qwen2-VL-2B outperforms SFT and Visual-RFT, achieving robust performance and enhanced explainability. RISE offers a self-supervised solution for advancing VLM reasoning without requiring manually annotated CoTs.

[1046] Reinforcement Learning-based Adaptive Path Selection for Programmable Networks

José Eduardo Zerna Torres, Marios Avgeris, Chrysa Papagianni, Gergely Pongrácz, István Gódor, Paola Grosso

Main category: cs.LG

TL;DR: Proof-of-concept implementation of distributed in-network reinforcement learning framework for adaptive path selection using Stochastic Learning Automata and real-time telemetry data.

DetailsMotivation: To enable local, data-driven forwarding decisions that dynamically adapt to congestion conditions in programmable networks, improving network efficiency and responsiveness.

Method: Combines Stochastic Learning Automata (SLA) with real-time telemetry data collected via In-Band Network Telemetry (INT) on P4-programmable BMv2 switches in a Mininet-based testbed.

Result: The SLA-based mechanism converges to effective path selections and adapts to shifting network conditions at line rate, demonstrating successful dynamic adaptation.

Conclusion: The distributed in-network reinforcement learning framework shows promise for adaptive path selection in programmable networks using local intelligence and real-time telemetry.

Abstract: This work presents a proof-of-concept implementation of a distributed, in-network reinforcement learning (IN-RL) framework for adaptive path selection in programmable networks. By combining Stochastic Learning Automata (SLA) with real-time telemetry data collected via In-Band Network Telemetry (INT), the proposed system enables local, data-driven forwarding decisions that adapt dynamically to congestion conditions. The system is evaluated on a Mininet-based testbed using P4-programmable BMv2 switches, demonstrating how our SLA-based mechanism converges to effective path selections and adapts to shifting network conditions at line rate.

[1047] Parameter-Aware Ensemble SINDy for Interpretable Symbolic SGS Closure

Hanseul Kang, Ville Vuorinen, Shervin Karimkashi

Main category: cs.LG

TL;DR: A scalable sparse regression framework for discovering PDEs and subgrid-scale closures from multi-parameter data, with four key enhancements to SINDy: symbolic parameterization, dimensional consistency filtering, memory-efficient processing, and ensemble consensus.

DetailsMotivation: To overcome limitations in existing SINDy methods for discovering interpretable partial differential equations and subgrid-scale closures from multi-parameter simulation data, particularly for turbulence modeling applications.

Method: Builds on SINDy with four enhancements: 1) symbolic parameterization for varying physical parameters, 2) Dimensional Similarity Filter for unit consistency, 3) memory-efficient Gram-matrix accumulation for large datasets, 4) ensemble consensus with coefficient stability analysis for robust model identification.

Result: Successfully discovered governing equations across parameter ranges in 1D benchmarks. For filtered Burgers datasets, autonomously discovered SGS closure τ_SGS = 0.1604·Δ²(∂ū/∂x)² with Smagorinsky constant C_s ≈ 0.4005, achieving R² = 0.885 across filter scales with improved accuracy over classical closures.

Conclusion: The framework provides a complementary approach to existing turbulence modeling methods by identifying physically meaningful SGS forms and calibrating coefficients directly from data, contributing to data-driven turbulence closure discovery.

Abstract: This work designs a scalable, parameter-aware sparse regression framework for discovering interpretable partial differential equations and subgrid-scale closures from multi-parameter simulation data. Building on SINDy (Sparse Identification of Nonlinear Dynamics), the approach addresses key limitations through four enhancements. First, symbolic parameterisation enables physical parameters to vary within unified regression. Second, the Dimensional Similarity Filter enforces unit consistency while reducing candidate libraries. Third, memory-efficient Gram-matrix accumulation enables batch processing of large datasets. Fourth, ensemble consensus with coefficient stability analysis ensures robust model identification. Validation on canonical one-dimensional benchmarks demonstrates consistent discovery of governing equations across parameter ranges. Applied to filtered Burgers datasets, the framework autonomously discovers the SGS closure $\tau_{\mathrm{SGS}} = 0.1604\cdot\Delta^2\left(\frac{\partial \bar{u}}{\partial x}\right)^2$ with the SINDy-discovered Smagorinsky constant $C_s^{\text{SINDy}} \approx 0.4005$ without predefined closure assumptions, recovering Smagorinsky-type structure directly from data. The discovered model achieves $R^2 = 0.885$ across filter scales and demonstrates improved prediction accuracy compared to classical SGS closures. The ability of the framework to identify physically meaningful SGS forms and calibrate coefficients offers a complementary approach to existing turbulence modelling methods, contributing to the broader field of data-driven turbulence closure discovery.

[1048] EEGDM: EEG Representation Learning via Generative Diffusion Model

Jia Hong Puah, Sim Kuan Goh, Ziwei Zhang, Zixuan Ye, Chow Khuen Chan, Kheng Seang Lim, Si Lei Fong, Kok Sin Woon, Cuntai Guan

Main category: cs.LG

TL;DR: EEGDM: A generative diffusion model framework for EEG representation learning that uses structured state-space diffusion pretraining and latent fusion transformers, outperforming existing EEG foundation models with better computational efficiency.

DetailsMotivation: EEG foundation models face challenges with high computational costs and marginal performance gains despite large model sizes. There's a need for more efficient and effective EEG representation learning methods.

Method: Proposed EEGDM framework with structured state-space model for diffusion pretraining (SSMDP) using DDPM framework to capture EEG temporal dynamics, followed by latent fusion transformer (LFT) for downstream classification tasks.

Result: Outperformed state-of-the-art approaches including EEG foundation models on multi-event datasets covering interictal epileptiform discharges and seizure detection tasks.

Conclusion: EEGDM provides a promising alternative to current foundation models, offering better performance with improved computational efficiency for EEG representation learning.

Abstract: While electroencephalogram (EEG) has been a crucial tool for monitoring the brain and diagnosing neurological disorders (e.g., epilepsy), learning meaningful representations from raw EEG signals remains challenging due to limited annotations and high signal variability. Recently, EEG foundation models (FMs) have shown promising potential by adopting transformer architectures and self-supervised pre-training methods from large language models (e.g., masked prediction) to learn representations from diverse EEG data, followed by fine-tuning on specific EEG tasks. Nonetheless, these large models often incurred high computational costs during both training and inference, with only marginal performance improvements as the model size increases. In this work, we proposed an EEG representation learning framework building upon Generative Diffusion Model (EEGDM). Specifically, we developed a structured state-space model for diffusion pretraining (SSMDP) to better capture the temporal dynamics of EEG signals and trained it using Denoising Diffusion Probabilistic Model (DDPM) framework. Subsequently, the resulting latent EEG representations were then used for downstream classification tasks via our proposed latent fusion transformer (LFT). To evaluate our method, we used multi-event datasets covering both interictal epileptiform discharges (TUEV) and seizure (CHB-MIT) detection, and compared EEGDM with current state-of-the-art approaches, including EEG FMs. Empirical results showed that our method outperformed the existing methods. These findings suggested that EEGDM offered a promising alternative to current FMs. Our source code and checkpoint are available at: https://github.com/jhpuah/EEGDM.

[1049] Mutual Information Surprise: Rethinking Unexpectedness in Autonomous Systems

Yinsong Wang, Quan Zeng, Xiao Liu, Yu Ding

Main category: cs.LG

TL;DR: Introduces Mutual Information Surprise (MIS) framework that redefines surprise as epistemic growth signal rather than anomaly detection, enabling autonomous systems to detect learning progression and adapt behavior dynamically.

DetailsMotivation: Current autonomous systems lack principled mechanisms to detect and adapt to unexpectedness, relying on static heuristics that cannot capture whether the system is truly learning and adapting.

Method: Develops Mutual Information Surprise (MIS) to quantify impact of new observations on mutual information, creates statistical test sequence for detecting meaningful shifts, and proposes MIS Reaction Policy (MISRP) for dynamic behavior governance through sampling adjustment and process forking.

Result: Empirical evaluations on synthetic domains and pollution map estimation show MISRP-governed strategies significantly outperform classical surprise-based approaches in stability, responsiveness, and predictive accuracy.

Conclusion: MIS shifts surprise from reactive to reflective, providing a path toward more self-aware and adaptive autonomous systems by focusing on epistemic growth rather than anomaly detection.

Abstract: Recent breakthroughs in autonomous experimentation have demonstrated remarkable physical capabilities, yet their cognitive control remains limited–often relying on static heuristics or classical optimization. A core limitation is the absence of a principled mechanism to detect and adapt to the unexpectedness. While traditional surprise measures–such as Shannon or Bayesian Surprise–offer momentary detection of deviation, they fail to capture whether a system is truly learning and adapting. In this work, we introduce Mutual Information Surprise (MIS), a new framework that redefines surprise not as anomaly detection, but as a signal of epistemic growth. MIS quantifies the impact of new observations on mutual information, enabling autonomous systems to reflect on their learning progression. We develop a statistical test sequence to detect meaningful shifts in estimated mutual information and propose a mutual information surprise reaction policy (MISRP) that dynamically governs system behavior through sampling adjustment and process forking. Empirical evaluations–on both synthetic domains and a dynamic pollution map estimation task–show that MISRP-governed strategies significantly outperform classical surprise-based approaches in stability, responsiveness, and predictive accuracy. By shifting surprise from reactive to reflective, MIS offers a path toward more self-aware and adaptive autonomous systems.

[1050] Linear cost mutual information estimation and independence test of similar performance as HSIC

Jarek Duda, Jagoda Bracha, Adrian Przybysz

Main category: cs.LG

TL;DR: HCR (Hierarchical Correlation Reconstruction) is proposed as a linear-cost alternative to HSIC for statistical dependency evaluation, offering higher sensitivity and joint distribution modeling while reducing computational complexity from O(n^2.37) to O(n).

DetailsMotivation: HSIC (Hilbert-Schmidt Information Criterion) is state-of-the-art for dependency evaluation but has impractical O(n^2.37) computational complexity for large datasets, making it unsuitable for big data applications.

Method: HCR uses hierarchical correlation reconstruction with features being mixed moments (starting with correlation and homoscedasticity) to describe dependencies. Each feature is calculated in O(n) linear time, with the number of features scaling with dimension (O(d^2) for pairwise, O(d^3) for triplewise dependencies).

Result: HCR provides a practical linear-cost alternative that shows even higher sensitivity to dependencies than HSIC, while also offering joint distribution modeling for chosen significance levels and approximation of mutual information.

Conclusion: HCR successfully addresses HSIC’s computational limitations by providing a linear-time method for dependency evaluation that is more sensitive, practical for large datasets, and additionally offers distribution modeling and mutual information approximation capabilities.

Abstract: Evaluation of statistical dependencies between two data samples is a basic problem of data science/machine learning, and HSIC (Hilbert-Schmidt Information Criterion)\cite{HSIC} is considered the state-of-art method. However, for size $n$ data sample it requires multiplication of $n\times n$ matrices, what currently needs $\sim O(n^{2.37})$ computational complexity\cite{mult}, making it impractical for large data samples. We discuss HCR (Hierarchical Correlation Reconstruction) as its linear cost practical alternative, in tests of even higher sensitivity to dependencies, and additionally providing actual joint distribution model for chosen significance level, by description of dependencies through features being mixed moments, starting with correlation and homoscedasticity. Also allowing to approximate mutual information as just sum of squares of such nontrivial mixed moments between two data samples. Such single dependence describing feature is calculated in $O(n)$ linear time. Their number to test varies with dimension $d$ - requiring $O(d^2)$ for pairwise dependencies, $O(d^3)$ if wanting to also consider more subtle triplewise, and so on.

[1051] End to End Autoencoder MLP Framework for Sepsis Prediction

Hejiang Cai, Di Wu, Ji Xu, Xiang Liu, Yiziting Zhu, Xin Shu, Yujie Li, Bin Yi

Main category: cs.LG

TL;DR: End-to-end deep learning framework combining autoencoder feature extraction with MLP classifier outperforms traditional ML methods for sepsis detection in ICU settings, achieving 74.6-93.5% accuracy across three cohorts.

DetailsMotivation: Traditional ML approaches for sepsis detection require manual feature engineering and struggle with irregular, incomplete time-series data from electronic health records, necessitating a more robust automated solution.

Method: Unsupervised autoencoder for automatic feature extraction combined with multilayer perceptron classifier, using customized down sampling strategy and non-overlapping dynamic sliding window mechanism with explicit missingness indicators.

Result: Achieved accuracies of 74.6%, 80.6%, and 93.5% across three ICU cohorts, consistently outperforming traditional machine learning baselines including Naive Bayes, SVM, Random Forest, and XGBoost.

Conclusion: The framework demonstrates superior robustness, generalizability, and clinical utility for early sepsis detection across heterogeneous ICU environments, providing an effective end-to-end solution for time-series medical data.

Abstract: Sepsis is a life threatening condition that requires timely detection in intensive care settings. Traditional machine learning approaches, including Naive Bayes, Support Vector Machine (SVM), Random Forest, and XGBoost, often rely on manual feature engineering and struggle with irregular, incomplete time-series data commonly present in electronic health records. We introduce an end-to-end deep learning framework integrating an unsupervised autoencoder for automatic feature extraction with a multilayer perceptron classifier for binary sepsis risk prediction. To enhance clinical applicability, we implement a customized down sampling strategy that extracts high information density segments during training and a non-overlapping dynamic sliding window mechanism for real-time inference. Preprocessed time series data are represented as fixed dimension vectors with explicit missingness indicators, mitigating bias and noise. We validate our approach on three ICU cohorts. Our end-to-end model achieves accuracies of 74.6 percent, 80.6 percent, and 93.5 percent, respectively, consistently outperforming traditional machine learning baselines. These results demonstrate the framework’s superior robustness, generalizability, and clinical utility for early sepsis detection across heterogeneous ICU environments.

[1052] FORGE: Foundational Optimization Representations from Graph Embeddings

Zohair Shafi, Serdar Kadioglu

Main category: cs.LG

TL;DR: Forge is a pre-training method using vector-quantized graph autoencoders on diverse MIP instances without solution dependency, enabling both unsupervised clustering and supervised predictions for solver optimization.

DetailsMotivation: Existing learning-based optimization approaches require solving many hard instances for training data and need dedicated models per problem distribution, limiting scalability and generalization.

Method: Pre-train vector-quantized graph autoencoder on large collection of mixed-integer programming instances unsupervised, creating discrete code vocabulary to represent optimization instances.

Result: Forge embeddings effectively differentiate/cluster unseen instances unsupervised. Fine-tuned embeddings predict warm-start variables and integrality gaps across multiple problem types, improving commercial solver performance.

Conclusion: Forge enables scalable pre-training on diverse MIP instances without solution dependency, providing generalizable embeddings that enhance solver performance across multiple problem distributions.

Abstract: Combinatorial optimization problems are ubiquitous in science and engineering, yet learning-based approaches to accelerate their solution often require solving a large number of hard-to-solve optimization instances to collect training data, incurring significant computational overhead. Existing methods require training dedicated models for each problem distribution for each downstream task, severely limiting their scalability and generalization. In this work, we introduce Forge, a method of pre-training a vector-quantized graph autoencoder on a large and diverse collection of mixed-integer programming (MIP) instances in an unsupervised fashion without dependency on their solution. The vector quantization process creates discrete code assignments that act as a vocabulary to represent optimization instances. We evaluate our approach under both supervised and unsupervised settings. For the unsupervised setting, we demonstrate that Forge embeddings effectively differentiate and cluster unseen instances. For the supervised setting, we fine-tuneForge embeddings and show that a single model predicts both the variables for warm-starts and integrality gaps for cut-generation across multiple problem type distributions. Both predictions help improve performance of a state-of-the-art, commercial optimization solver. Finally, we release our code and pre-trained Forge weights to encourage further research and practical use of instance-level MIP embeddings at https://github.com/skadio/forge/.

cs.MA

[1053] KG-RAG: Enhancing GUI Agent Decision-Making via Knowledge Graph-Driven Retrieval-Augmented Generation

Ziyi Guan, Jason Chun Lok Li, Zhijian Hou, Pingping Zhang, Donglai Xu, Yuzhi Zhao, Mengyang Wu, Jinpeng Chen, Thanh-Toan Nguyen, Pengfei Xian, Wenao Ma, Shengchao Qin, Graziano Chesi, Ngai Wong

Main category: cs.MA

TL;DR: KG-RAG is a Knowledge Graph-driven RAG framework that transforms UI Transition Graphs into structured vector databases to improve mobile GUI agent performance through efficient navigation path generation.

DetailsMotivation: Current GUI agents struggle with complex mobile tasks due to limited app-specific knowledge and underutilization of UI Transition Graphs (UTGs) caused by poor extraction and inefficient integration methods.

Method: Transforms fragmented UTGs into structured vector databases for real-time retrieval, uses intent-guided LLM search to generate actionable navigation paths, and enhances agent decision-making through knowledge graph-driven retrieval.

Result: Achieves 75.8% success rate (8.9% improvement over AutoDroid), 84.6% decision accuracy (8.1% improvement), reduces average task steps from 4.5 to 4.1, and demonstrates transferability to web/desktop applications with significant performance gains.

Conclusion: KG-RAG effectively addresses UTG utilization challenges, significantly improves mobile GUI agent performance, enables practical deployment trade-offs with UTG cost optimization (~4h per complex app), and provides benchmarks for future Chinese mobile ecosystem research.

Abstract: Despite recent progress, Graphic User Interface (GUI) agents powered by Large Language Models (LLMs) struggle with complex mobile tasks due to limited app-specific knowledge. While UI Transition Graphs (UTGs) offer structured navigation representations, they are underutilized due to poor extraction and inefficient integration. We introduce KG-RAG, a Knowledge Graph-driven Retrieval-Augmented Generation framework that transforms fragmented UTGs into structured vector databases for efficient real-time retrieval. By leveraging an intent-guided LLM search method, KG-RAG generates actionable navigation paths, enhancing agent decision-making. Experiments across diverse mobile apps show that KG-RAG outperforms existing methods, achieving a 75.8% success rate (8.9% improvement over AutoDroid), 84.6% decision accuracy (8.1% improvement), and reducing average task steps from 4.5 to 4.1. Additionally, we present KG-Android-Bench and KG-Harmony-Bench, two benchmarks tailored to the Chinese mobile ecosystem for future research. Finally, KG-RAG transfers to web/desktop (+40% SR on Weibo-web; +20% on QQ Music-desktop), and a UTG cost ablation shows accuracy saturates at ~4h per complex app, enabling practical deployment trade-offs.

[1054] MobiAgent: A Systematic Framework for Customizable Mobile Agents

Cheng Zhang, Erhu Feng, Xi Zhao, Yisheng Zhao, Wangbo Gong, Jiahui Sun, Dong Du, Zhichao Hua, Yubin Xia, Haibo Chen

Main category: cs.MA

TL;DR: MobiAgent is a comprehensive mobile agent system that achieves state-of-the-art performance in real-world mobile scenarios through specialized agent models, acceleration framework, and AI-assisted data collection.

DetailsMotivation: Existing GUI-based mobile agents face significant challenges in accuracy and efficiency for real-world task execution, and current mobile agents are limited by the availability of high-quality data.

Method: Proposes MobiAgent system with three core components: MobiMind-series agent models, AgentRR acceleration framework, MobiFlow benchmarking suite, plus an AI-assisted agile data collection pipeline that reduces manual annotation costs.

Result: Achieves state-of-the-art performance compared to both general-purpose LLMs and specialized GUI agent models in real-world mobile scenarios.

Conclusion: MobiAgent successfully addresses the limitations of existing mobile agents through its comprehensive system design and AI-assisted data collection approach, delivering superior accuracy and efficiency.

Abstract: With the rapid advancement of Vision-Language Models (VLMs), GUI-based mobile agents have emerged as a key development direction for intelligent mobile systems. However, existing agent models continue to face significant challenges in real-world task execution, particularly in terms of accuracy and efficiency. To address these limitations, we propose MobiAgent, a comprehensive mobile agent system comprising three core components: the MobiMind-series agent models, the AgentRR acceleration framework, and the MobiFlow benchmarking suite. Furthermore, recognizing that the capabilities of current mobile agents are still limited by the availability of high-quality data, we have developed an AI-assisted agile data collection pipeline that significantly reduces the cost of manual annotation. Compared to both general-purpose LLMs and specialized GUI agent models, MobiAgent achieves state-of-the-art performance in real-world mobile scenarios.

[1055] Nash Q-Network for Multi-Agent Cybersecurity Simulation

Qintong Xie, Edward Koh, Xavier Cadet, Peter Chin

Main category: cs.MA

TL;DR: Novel Nash Q-learning algorithm for cybersecurity defense using MARL, combining PPO, DQN and Nash-Q to achieve Nash equilibrium in two-player zero-sum Markov games.

DetailsMotivation: Cybersecurity defense involves adversarial interactions between defenders and hackers, making MARL suitable but challenging due to complexity of simultaneous training in non-trivial environments.

Method: Proposed Nash Q-Network that incorporates proximal policy optimization (PPO), deep Q-network (DQN), and Nash-Q algorithm with distributed data collection and specialized neural architectures for agents and critics.

Result: Successful implementation in complex cyber defense simulation treated as two-player zero-sum Markov game, achieving convergence to steady equilibrium.

Conclusion: The approach addresses non-stationarity and instability challenges in multi-agent learning, producing Nash-optimal strategies for robust cybersecurity defenses.

Abstract: Cybersecurity defense involves interactions between adversarial parties (namely defenders and hackers), making multi-agent reinforcement learning (MARL) an ideal approach for modeling and learning strategies for these scenarios. This paper addresses one of the key challenges to MARL, the complexity of simultaneous training of agents in nontrivial environments, and presents a novel policy-based Nash Q-learning to directly converge onto a steady equilibrium. We demonstrate the successful implementation of this algorithm in a notable complex cyber defense simulation treated as a two-player zero-sum Markov game setting. We propose the Nash Q-Network, which aims to learn Nash-optimal strategies that translate to robust defenses in cybersecurity settings. Our approach incorporates aspects of proximal policy optimization (PPO), deep Q-network (DQN), and the Nash-Q algorithm, addressing common challenges like non-stationarity and instability in multi-agent learning. The training process employs distributed data collection and carefully designed neural architectures for both agents and critics.

[1056] Controller synthesis method for multi-agent system based on temporal logic specification

Ruohan Huang, Zining Cao

Main category: cs.MA

TL;DR: Proposes a controller synthesis method for semi-cooperative semi-competitive multi-agent probabilistic discrete event systems using temporal logic specifications and probabilistic model checking.

DetailsMotivation: Traditional controller synthesis methods are limited to single-agent and non-probabilistic systems, while modern systems require handling complex multi-agent probabilistic scenarios with sophisticated control specifications.

Method: Combines probabilistic model checking with controller synthesis algorithm for multi-agent probabilistic discrete event systems, using linear temporal logic formulas to specify control requirements.

Result: Developed a controller synthesis approach that can ensure satisfaction of temporal logic specifications to a certain extent in semi-cooperative semi-competitive multi-agent probabilistic systems.

Conclusion: The proposed method effectively addresses controller synthesis for complex multi-agent probabilistic systems and was validated through case studies, demonstrating its practical applicability.

Abstract: Controller synthesis is a theoretical approach to the systematic design of discrete event systems. It constructs a controller to provide feedback and control to the system, ensuring it meets specified control specifications. Traditional controller synthesis methods often use formal languages to describe control specifications and are mainly oriented towards single-agent and non-probabilistic systems. With the increasing complexity of systems, the control requirements that need to be satisfied also become more complex. Based on this, this paper proposes a controller synthesis method for semi-cooperative semi-competitive multi-agent probabilistic discrete event systems to solve the controller synthesis problem based on temporal logic specifications. The controller can ensure the satisfaction of specifications to a certain extent. The specification is given in the form of a linear temporal logic formula. This paper designs a controller synthesis algorithm that combines probabilistic model checking. Finally, the effectiveness of this method is verified through a case study.

[1057] ShortageSim: Simulating Drug Shortages under Information Asymmetry

Mingxuan Cui, Yilan Jiang, Duo Zhou, Cheng Qian, Yuji Zhang, Qiong Wang

Main category: cs.MA

TL;DR: ShortageSim is an LLM-based multi-agent simulation framework that models pharmaceutical supply chain interactions to address drug shortages, reducing resolution lag by 83% compared to baseline models.

DetailsMotivation: Drug shortages pose critical risks to patient care globally, but regulatory interventions are poorly understood due to information asymmetries in pharmaceutical supply chains.

Method: Uses Large Language Models to simulate bounded-rational decision-making between drug manufacturers, institutional buyers, and regulatory agencies through a sequential production game spanning multiple quarters, modeling FDA announcements and their propagation.

Result: ShortageSim reduces resolution-lag percentage for discontinued-disclosed cases by 83%, bringing simulated durations more aligned to ground truth than zero-shot baseline.

Conclusion: Provides a novel computational framework for designing and testing interventions in complex, information-scarce supply chains, with open-source code and dataset of 2,925 FDA shortage events.

Abstract: Drug shortages pose critical risks to patient care and healthcare systems worldwide, yet the effectiveness of regulatory interventions remains poorly understood due to fundamental information asymmetries in pharmaceutical supply chains. We present \textbf{ShortageSim}, the first Large Language Model (LLM)-based multi-agent simulation framework that captures the complex, strategic interactions between drug manufacturers, institutional buyers, and regulatory agencies in response to shortage alerts. Unlike traditional game-theoretic models that assume perfect rationality and complete information, \textbf{ShortageSim} leverages LLMs to simulate bounded-rational decision-making under uncertainty. Through a sequential production game spanning multiple quarters, we model how FDA announcements, both reactive alerts about existing shortages and proactive warnings about potential disruptions, propagate through the supply chain and influence capacity investment and procurement decisions. Our experiments on historical shortage events reveal that \textbf{ShortageSim} reduces the resolution-lag percentage for discontinued-disclosed cases by 83%, bringing simulated durations more aligned to ground truth than the zero-shot baseline. We open-source \textbf{ShortageSim} and a dataset of 2,925 FDA shortage events at https://github.com/Lemutisme/Sortage_Management, providing a novel computational framework for designing and testing interventions in complex, information-scarce supply chains.

[1058] Contemporary Agent Technology: LLM-Driven Advancements vs Classic Multi-Agent Systems

Costin Bădică, Amelia Bădică, Maria Ganzha, Mirjana Ivanović, Marcin Paprzycki, Dan Selişteanu, Zofia Wrona

Main category: cs.MA

TL;DR: Analysis of LLM-driven agent technology vs classic Multi-Agent Systems, examining models, approaches, and their relationship to foundational MAS concepts.

DetailsMotivation: To provide comprehensive reflection on contemporary agent technology advancements, particularly comparing LLM-driven approaches with classic Multi-Agent Systems and understanding their relationship to foundational MAS literature.

Method: Critical analysis and comprehensive reflection on models, approaches, and characteristics of LLM-driven agent systems compared to classic MAS, drawing from core academic literature.

Result: Identifies key characteristics and differences between LLM-driven agent technology and classic MAS, highlighting how recent developments relate to foundational multi-agent systems concepts.

Conclusion: The paper identifies key challenges and promising future directions in the rapidly evolving domain of agent technology, emphasizing the need to understand the relationship between modern LLM approaches and traditional MAS foundations.

Abstract: This contribution provides our comprehensive reflection on the contemporary agent technology, with a particular focus on the advancements driven by Large Language Models (LLM) vs classic Multi-Agent Systems (MAS). It delves into the models, approaches, and characteristics that define these new systems. The paper emphasizes the critical analysis of how the recent developments relate to the foundational MAS, as articulated in the core academic literature. Finally, it identifies key challenges and promising future directions in this rapidly evolving domain.

[1059] Fairness Aware Reinforcement Learning via Proximal Policy Optimization

Gabriele La Malfa, Jie M. Zhang, Michael Luck, Elizabeth Black

Main category: cs.MA

TL;DR: Fair-PPO integrates fairness penalties into Proximal Policy Optimization to achieve equitable reward distribution in multi-agent systems while maintaining performance comparable to state-of-the-art fair RL algorithms.

DetailsMotivation: Address fairness challenges in multi-agent systems where equitable reward distribution is crucial, particularly when agents have sensitive attributes like race, gender, or socioeconomic status.

Method: Extends PPO with fairness penalty terms derived from demographic parity, counterfactual fairness, or conditional statistical parity. Uses retrospective (past outcomes) and prospective (future decisions) penalty components to balance reward maximization with fairness.

Result: Fair-PPO achieves fairer policies than standard PPO across multiple fairness metrics in two test environments (Allelopathic Harvest and HospitalSim). Performance matches state-of-the-art fair RL algorithms, though with some efficiency trade-off.

Conclusion: Fair-PPO effectively addresses fairness challenges in MAS through integrated penalty components, demonstrating a spectrum of strategies to improve fairness without compromising overall population equality.

Abstract: Fairness in multi-agent systems (MAS) focuses on equitable reward distribution among agents in scenarios involving sensitive attributes such as race, gender, or socioeconomic status. This paper introduces fairness in Proximal Policy Optimization (PPO) with a penalty term derived from a fairness definition such as demographic parity, counterfactual fairness, or conditional statistical parity. The proposed method, which we call Fair-PPO, balances reward maximisation with fairness by integrating two penalty components: a retrospective component that minimises disparities in past outcomes and a prospective component that ensures fairness in future decision-making. We evaluate our approach in two games: the Allelopathic Harvest, a cooperative and competitive MAS focused on resource collection, where some agents possess a sensitive attribute, and HospitalSim, a hospital simulation, in which agents coordinate the operations of hospital patients with different mobility and priority needs. Experiments show that Fair-PPO achieves fairer policies than PPO across the fairness metrics and, through the retrospective and prospective penalty components, reveals a wide spectrum of strategies to improve fairness; at the same time, its performance pairs with that of state-of-the-art fair reinforcement-learning algorithms. Fairness comes at the cost of reduced efficiency, but does not compromise equality among the overall population (Gini index). These findings underscore the potential of Fair-PPO to address fairness challenges in MAS.

[1060] On Word-of-Mouth and Private-Prior Sequential Social Learning

Andrea Da Col, Cristian R. Rojas, Vikram Krishnamurthy

Main category: cs.MA

TL;DR: Analysis of Word-of-Mouth social learning where agents sequentially estimate a dynamic system state, with the final agent’s belief being broadcast and adopted by all agents, showing mixed performance effects.

DetailsMotivation: To study social learning interactions among rational agents who observe actions but lack direct belief access, specifically examining the Word-of-Mouth paradigm in dynamic system estimation.

Method: Theoretical analysis and numerical simulations of a sequential estimation process where the first agent gets noisy measurements and subsequent agents receive degraded versions of predecessors’ estimates, with final agent’s belief being broadcast to all.

Result: Mixed performance outcomes - some agents benefit from adopting the final agent’s broadcast belief while others experience performance deterioration.

Conclusion: Word-of-Mouth social learning in dynamic systems produces heterogeneous effects on agent performance when adopting a centralized final belief, highlighting complex trade-offs in distributed estimation systems.

Abstract: Social learning constitutes a fundamental framework for studying interactions among rational agents who observe each other’s actions but lack direct access to individual beliefs. This paper investigates a specific social learning paradigm known as Word-of-Mouth (WoM), where a series of agents seeks to estimate the state of a dynamical system. The first agent receives noisy measurements of the state, while each subsequent agent relies solely on a degraded version of her predecessor’s estimate. A defining feature of WoM is that the final agent’s belief is publicly broadcast and subsequently adopted by all agents, in place of their own. We analyze this setting theoretically and through numerical simulations, noting that some agents benefit from using the belief of the last agent, while others experience performance deterioration.

[1061] RALLY: Role-Adaptive LLM-Driven Yoked Navigation for Agentic UAV Swarms

Ziyao Wang, Rongpeng Li, Sizhao Li, Yuming Xiang, Haiping Wang, Zhifeng Zhao, Honggang Zhang

Main category: cs.MA

TL;DR: RALLY is a novel LLM-MARL hybrid approach for UAV swarm navigation that combines semantic communication, dynamic role adaptation, and semi-offline training to overcome limitations of traditional MARL and LLM-only methods.

DetailsMotivation: Traditional MARL approaches suffer from semantic communication gaps and rigid homogeneous roles, while LLM-based methods lack online learning capabilities and struggle with exploration. There's a need for a solution that combines semantic reasoning with adaptive online learning for UAV swarm navigation.

Method: Proposes RALLY with three key components: 1) LLM-driven semantic decision framework using structured natural language, 2) Dynamic role-heterogeneity mechanism for adaptive role switching, 3) RMIX-based assignment strategy integrating LLM offline priors with MARL online policies for semi-offline training.

Result: Experiments in MPE environment and SITL platform show RALLY outperforms conventional approaches in task coverage, convergence speed, and generalization for collaborative UAV navigation.

Conclusion: RALLY demonstrates strong potential for agentic multi-UAV systems by effectively combining semantic reasoning with adaptive online learning, addressing key limitations of both traditional MARL and LLM-only approaches.

Abstract: Intelligent control of Unmanned Aerial Vehicles (UAVs) swarms has emerged as a critical research focus, and it typically requires the swarm to navigate effectively while avoiding obstacles and achieving continuous coverage over multiple mission targets. Although traditional Multi-Agent Reinforcement Learning (MARL) approaches offer dynamic adaptability, they are hindered by the semantic gap in numerical communication and the rigidity of homogeneous role structures, resulting in poor generalization and limited task scalability. Recent advances in Large Language Model (LLM)-based control frameworks demonstrate strong semantic reasoning capabilities by leveraging extensive prior knowledge. However, due to the lack of online learning and over-reliance on static priors, these works often struggle with effective exploration, leading to reduced individual potential and overall system performance. To address these limitations, we propose a Role-Adaptive LLM-Driven Yoked navigation algorithm RALLY. Specifically, we first develop an LLM-driven semantic decision framework that uses structured natural language for efficient semantic communication and collaborative reasoning. Afterward, we introduce a dynamic role-heterogeneity mechanism for adaptive role switching and personalized decision-making. Furthermore, we propose a Role-value Mixing Network (RMIX)-based assignment strategy that integrates LLM offline priors with MARL online policies to enable semi-offline training of role selection strategies. Experiments in the Multi-Agent Particle Environment (MPE) environment and a Software-In-The-Loop (SITL) platform demonstrate that RALLY outperforms conventional approaches in terms of task coverage, convergence speed, and generalization, highlighting its strong potential for collaborative navigation in agentic multi-UAV systems.

[1062] Grid-Agent: An LLM-Powered Multi-Agent System for Power Grid Control

Yan Zhang

Main category: cs.MA

TL;DR: Grid-Agent is an AI framework combining LLMs and multi-agent reinforcement learning for real-time power grid violation detection and remediation, showing superior performance in standard test systems.

DetailsMotivation: Increasing DERs, EV adoption, and extreme weather events have made power grid management more complex, requiring solutions that traditional rule-based systems and numerical optimization struggle to handle.

Method: Combines LLMs with multi-agent reinforcement learning using planning and validation agents, integrates semantic reasoning with numerical precision, employs adaptive multiscale network representation, and enables coordinated violation resolution through switch configurations, battery deployment, and load curtailment.

Result: Demonstrates superior violation mitigation performance in standard IEEE and CIGRE test systems (IEEE 69-bus, CIGRE MV, and IEEE 30-bus) with built-in data collection and continuous learning capabilities.

Conclusion: The autonomous framework is particularly suitable for modern smart grid applications requiring rapid response to dynamic operating conditions, offering scalability and adaptability to diverse network topologies.

Abstract: The increasing penetration of Distributed Energy Resources (DERs), widespread adoption of Electric Vehicles (EVs), and the growing frequency of extreme weather events have significantly increased the complexity of power grid planning, operation, and management. Traditional rule-based systems and numerical optimization approaches often struggle with the scale, dynamics, and adaptability required by modern power networks. This paper introduces Grid-Agent, an autonomous, AI-driven framework that combines Large Language Models (LLMs) with multi-agent reinforcement learning to detect and remediate grid violations in real time. Grid-Agent integrates semantic reasoning with numerical precision through a modular agent architecture: a planning agent generates coordinated action sequences using numerical power flow solvers, while a validation agent evaluates system stability and action effectiveness via sandboxed execution with safety rollbacks. To ensure scalability, Grid-Agent incorporates an adaptive multiscale network representation that dynamically selects optimal encoding schemes based on network size and complexity. The framework enables coordinated violation resolution through optimizing switch configurations, battery deployment, and load curtailment strategies. Experimental results in standard IEEE and CIGRE test systems (IEEE 69-bus, CIGRE MV, and IEEE 30-bus) demonstrate superior violation mitigation performance. Additionally, the framework’s built-in data collection and learning capabilities enable continuous learning and adaptation to diverse network topologies. The autonomous nature of the framework makes it particularly suitable for modern smart grid applications requiring rapid response to dynamic operating conditions.

cs.MM

[1063] Traj-MLLM: Can Multimodal Large Language Models Reform Trajectory Data Mining?

Shuo Liu, Di Yao, Yan Lin, Gao Cong, Jingping Bi

Main category: cs.MM

TL;DR: Traj-MLLM is a general framework that uses multimodal large language models for trajectory data mining, achieving state-of-the-art performance across multiple tasks without training data or fine-tuning.

DetailsMotivation: Existing trajectory analysis methods suffer from generalization problems - they are restricted to specific regions or limited tasks. The paper explores whether MLLMs can solve trajectory data mining challenges and overcome modality gaps.

Method: Integrates multiview contexts to transform raw trajectories into interleaved image-text sequences while preserving spatial-temporal characteristics. Uses MLLMs’ reasoning ability directly and proposes a prompt optimization method for task adaptation.

Result: Outperforms SOTA baselines by 48.05% on travel time estimation, 15.52% on mobility prediction, 51.52% on anomaly detection, and 1.83% on transportation mode identification across four datasets.

Conclusion: Traj-MLLM successfully demonstrates that MLLMs can effectively reform trajectory data mining, achieving superior performance without requiring training data or fine-tuning, making it a general solution for various trajectory analysis tasks.

Abstract: Building a general model capable of analyzing human trajectories across different geographic regions and different tasks becomes an emergent yet important problem for various applications. However, existing works suffer from the generalization problem, \ie, they are either restricted to train for specific regions or only suitable for a few tasks. Given the recent advances of multimodal large language models (MLLMs), we raise the question: can MLLMs reform current trajectory data mining and solve the problem? Nevertheless, due to the modality gap of trajectory, how to generate task-independent multimodal trajectory representations and how to adapt flexibly to different tasks remain the foundational challenges. In this paper, we propose \texttt{Traj-MLLM}}, which is the first general framework using MLLMs for trajectory data mining. By integrating multiview contexts, \texttt{Traj-MLLM}} transforms raw trajectories into interleaved image-text sequences while preserving key spatial-temporal characteristics, and directly utilizes the reasoning ability of MLLMs for trajectory analysis. Additionally, a prompt optimization method is proposed to finalize data-invariant prompts for task adaptation. Extensive experiments on four publicly available datasets show that \texttt{Traj-MLLM}} outperforms state-of-the-art baselines by $48.05%$, $15.52%$, $51.52%$, $1.83%$ on travel time estimation, mobility prediction, anomaly detection and transportation mode identification, respectively. \texttt{Traj-MLLM}} achieves these superior performances without requiring any training data or fine-tuning the MLLM backbones.

[1064] LLM-Guided Semantic Relational Reasoning for Multimodal Intent Recognition

Qianrui Zhou, Hua Xu, Yifan Wang, Xinzhi Dong, Hanlei Zhang

Main category: cs.MM

TL;DR: LGSRR method uses LLMs to extract fine-grained semantics and guide relational reasoning for multimodal intent understanding, achieving state-of-the-art performance without manual priors.

DetailsMotivation: Existing methods have limitations in modality-level reliance and relational reasoning for complex intent understanding from multimodal signals.

Method: LLM-based strategy with shallow-to-deep Chain-of-Thought to extract semantics, plus formal modeling of three fundamental semantic relations based on logical principles.

Result: Superior performance over state-of-the-art methods on multimodal intent and dialogue act recognition tasks with consistent gains across scenarios.

Conclusion: LGSRR effectively leverages LLM knowledge to enhance relational reasoning for multimodal intent understanding without manual priors.

Abstract: Understanding human intents from multimodal signals is critical for analyzing human behaviors and enhancing human-machine interactions in real-world scenarios. However, existing methods exhibit limitations in their modality-level reliance, constraining relational reasoning over fine-grained semantics for complex intent understanding. This paper proposes a novel LLM-Guided Semantic Relational Reasoning (LGSRR) method, which harnesses the expansive knowledge of large language models (LLMs) to establish semantic foundations that boost smaller models’ relational reasoning performance. Specifically, an LLM-based strategy is proposed to extract fine-grained semantics as guidance for subsequent reasoning, driven by a shallow-to-deep Chain-of-Thought (CoT) that autonomously uncovers, describes, and ranks semantic cues by their importance without relying on manually defined priors. Besides, we formally model three fundamental types of semantic relations grounded in logical principles and analyze their nuanced interplay to enable more effective relational reasoning. Extensive experiments on multimodal intent and dialogue act recognition tasks demonstrate LGSRR’s superiority over state-of-the-art methods, with consistent performance gains across diverse semantic understanding scenarios. The complete data and code are available at https://github.com/thuiar/LGSRR.

[1065] Efficient Geometry Compression and Communication for 3D Gaussian Splatting Point Clouds

Liang Xie, Yanting Li, Luyang Tang, Wei Gao

Main category: cs.MM

TL;DR: Proposes AVS PCRM compression for 3D Gaussian point cloud data in i3DV platform, achieving 10-25% bitrate savings while maintaining high-quality rendering under 40 Mbps bandwidth.

DetailsMotivation: Address storage and transmission challenges from explosive growth of 3D Gaussian data volume in dynamic scene representation, reducing excessive storage space occupancy.

Method: Integrates AVS PCRM reference software for efficient compression of Gaussian point cloud geometry data, combining with existing binary hash table rate-distortion optimization for inter-frame Gaussian point transformation caching.

Result: Achieves significant 10-25% bitrate savings on universal test sets while maintaining fast rendering and high-quality synthesis capabilities within 40 Mbps bandwidth constraint.

Conclusion: Provides superior rate-distortion tradeoff solution for storage, transmission, and interaction of 3D volumetric video by combining AVS PCRM compression with existing i3DV platform capabilities.

Abstract: Storage and transmission challenges in dynamic 3D scene representation based on the i3DV platform, With increasing scene complexity, the explosive growth of 3D Gaussian data volume causes excessive storage space occupancy. To address this issue, we propose adopting the AVS PCRM reference software for efficient compression of Gaussian point cloud geometry data. The strategy deeply integrates the advanced encoding capabilities of AVS PCRM into the i3DV platform, forming technical complementarity with the original rate-distortion optimization mechanism based on binary hash tables. On one hand, the hash table efficiently caches inter-frame Gaussian point transformation relationships, which allows for high-fidelity transmission within a 40 Mbps bandwidth constraint. On the other hand, AVS PCRM performs precise compression on geometry data. Experimental results demonstrate that the joint framework maintains the advantages of fast rendering and high-quality synthesis in 3D Gaussian technology while achieving significant 10%-25% bitrate savings on universal test sets. It provides a superior rate-distortion tradeoff solution for the storage, transmission, and interaction of 3D volumetric video.

[1066] Advancing Grounded Multimodal Named Entity Recognition via LLM-Based Reformulation and Box-Based Segmentation

Jinyuan Li, Ziyan Li, Han Li, Jianfei Yu, Rui Xia, Di Sun, Gang Pan

Main category: cs.MM

TL;DR: RiVEG is a unified framework that reformulates GMNER into joint MNER-VE-VG tasks using LLMs as bridges, addressing ungroundable entities and fine-grained entity challenges while enabling segmentation capabilities.

DetailsMotivation: Address two key challenges in GMNER: 1) tenuous image-text correlation leading to ungroundable entities, and 2) distinction between coarse-grained noun phrases and fine-grained named entities.

Method: Proposes RiVEG framework that reformulates GMNER as joint MNER-VE-VG task using LLMs as connecting bridges. Includes Entity Expansion Expression and Visual Entailment modules to unify Visual Grounding and Entity Grounding.

Result: Significantly outperforms state-of-the-art methods on four datasets across MNER, GMNER, and SMNER tasks. Also constructs new SMNER task and Twitter-SMNER dataset for fine-grained segmentation masks.

Conclusion: RiVEG effectively addresses GMNER limitations, provides unlimited data and model scalability, and demonstrates feasibility of using SAM for segmentation tasks, showing strong performance across multiple benchmarks.

Abstract: Grounded Multimodal Named Entity Recognition (GMNER) task aims to identify named entities, entity types and their corresponding visual regions. GMNER task exhibits two challenging attributes: 1) The tenuous correlation between images and text on social media contributes to a notable proportion of named entities being ungroundable. 2) There exists a distinction between coarse-grained noun phrases used in similar tasks (e.g., phrase localization) and fine-grained named entities. In this paper, we propose RiVEG, a unified framework that reformulates GMNER into a joint MNER-VE-VG task by leveraging large language models (LLMs) as connecting bridges. This reformulation brings two benefits: 1) It enables us to optimize the MNER module for optimal MNER performance and eliminates the need to pre-extract region features using object detection methods, thus naturally addressing the two major limitations of existing GMNER methods. 2) The introduction of Entity Expansion Expression module and Visual Entailment (VE) module unifies Visual Grounding (VG) and Entity Grounding (EG). This endows the proposed framework with unlimited data and model scalability. Furthermore, to address the potential ambiguity stemming from the coarse-grained bounding box output in GMNER, we further construct the new Segmented Multimodal Named Entity Recognition (SMNER) task and corresponding Twitter-SMNER dataset aimed at generating fine-grained segmentation masks, and experimentally demonstrate the feasibility and effectiveness of using box prompt-based Segment Anything Model (SAM) to empower any GMNER model with the ability to accomplish the SMNER task. Extensive experiments demonstrate that RiVEG significantly outperforms SoTA methods on four datasets across the MNER, GMNER, and SMNER tasks.

[1067] Feature Coding in the Era of Large Models: Dataset, Test Conditions, and Benchmark

Changsheng Gao, Yifan Ma, Qiaoxi Chen, Yenan Xu, Dong Liu, Weisi Lin

Main category: cs.MM

TL;DR: This paper introduces feature coding for large models to reduce transmission overhead in distributed deployment, providing a comprehensive dataset, standardized evaluation framework, and baseline methods.

DetailsMotivation: Large models face computational costs and privacy concerns during training/inference. Distributed deployment requires efficient feature coding to reduce transmission overhead of intermediate information between model segments.

Method: Created a dataset with diverse features from three representative large model types. Established unified test conditions for standardized evaluation. Introduced two baseline methods derived from image coding techniques.

Result: Provides benchmark performance of baseline methods on the proposed dataset. The dataset and source code are made publicly available for future research.

Conclusion: This work establishes foundational resources (dataset, evaluation framework, baselines) for feature coding research in large models, aiming to inspire broader engagement in this under-explored area.

Abstract: Large models have achieved remarkable performance across various tasks, yet they incur significant computational costs and privacy concerns during both training and inference. Distributed deployment has emerged as a potential solution, but it necessitates the exchange of intermediate information between model segments, with feature representations serving as crucial information carriers. To optimize information exchange, feature coding is required to reduce transmission and storage overhead. Despite its importance, feature coding for large models remains an under-explored area. In this paper, we draw attention to large model feature coding and make three fundamental contributions. First, we introduce a comprehensive dataset encompassing diverse features generated by three representative types of large models. Second, we establish unified test conditions, enabling standardized evaluation pipelines and fair comparisons across future feature coding studies. Third, we introduce two baseline methods derived from widely used image coding techniques and benchmark their performance on the proposed dataset. These contributions aim to provide a foundation for future research and inspire broader engagement in this field. To support a long-term study, all source code and the dataset are made available at \href{https://github.com/chansongoal/LaMoFC}{https://github.com/chansongoal/LaMoFC}.

[1068] DT-UFC: Universal Large Model Feature Coding via Peaky-to-Balanced Distribution Transformation

Changsheng Gao, Zijie Liu, Li Li, Dong Liu, Xiaoyan Sun, Weisi Lin

Main category: cs.MM

TL;DR: Proposes a universal feature coding method that transforms diverse feature distributions from different large models into a balanced target space, enabling efficient compression and cross-model generalization.

DetailsMotivation: Feature coding is crucial for distributed deployment of large models to reduce transmission/storage burden, but existing methods are task/model-specific and cannot handle the diverse distributional characteristics of features from different models.

Method: A learned peaky-to-balanced distribution transformation that non-uniformly reshapes skewed feature distributions into a common balanced target space, enabling universal codec training without modifying downstream codecs.

Result: Validated on LLaMA3, DINOv2, and SD3 models across multiple tasks/modalities, showing notable improvements in compression efficiency and cross-model generalization over task-specific baselines.

Conclusion: The proposed distribution transformation enables effective universal feature coding that generalizes across diverse large models and tasks, addressing the challenge of distributional heterogeneity in feature compression.

Abstract: Like image coding in visual data transmission, feature coding is essential for the distributed deployment of large models by significantly reducing transmission and storage burden. However, prior studies have mostly targeted task- or model-specific scenarios, leaving the challenge of universal feature coding across diverse large models largely unexplored. In this paper, we present the first systematic study on universal feature coding for large models. The key challenge lies in the inherently diverse and distributionally incompatible nature of features extracted from different models. For example, features from DINOv2 exhibit highly peaky, concentrated distributions, while those from Stable Diffusion 3 (SD3) are more dispersed and uniform. This distributional heterogeneity severely hampers both compression efficiency and cross-model generalization. To address this, we propose a learned peaky-to-balanced distribution transformation, which reshapes highly skewed feature distributions into a common, balanced target space. This transformation is non-uniform, data-driven, and plug-and-play, enabling effective alignment of heterogeneous distributions without modifying downstream codecs. With this alignment, a universal codec trained on the balanced target distribution can effectively generalize to features from different models and tasks. We validate our approach on three representative large models (LLaMA3, DINOv2, and SD3) across multiple tasks and modalities. Extensive experiments show that our method achieves notable improvements in both compression efficiency and cross-model generalization over task-specific baselines. All source code has been made available at https://github.com/chansongoal/DT-UFC.

eess.AS

[1069] DeepEmoNet: Building Machine Learning Models for Automatic Emotion Recognition in Human Speeches

Tai Vu

Main category: eess.AS

TL;DR: Machine learning approach using SVMs, LSTMs, CNNs with transfer learning and data augmentation for speech emotion recognition, achieving 66.7% accuracy with ResNet34.

DetailsMotivation: Speech emotion recognition is challenging due to unclear connections between human emotions and sound components like pitch, loudness, and energy.

Method: Built several ML models including SVMs, LSTMs, and CNNs, leveraging transfer learning and data augmentation for efficient training on small dataset.

Result: Best model was ResNet34 network achieving 66.7% accuracy and 0.631 F1 score.

Conclusion: Machine learning with transfer learning and data augmentation can effectively address SER challenges even with limited data.

Abstract: Speech emotion recognition (SER) has been a challenging problem in spoken language processing research, because it is unclear how human emotions are connected to various components of sounds such as pitch, loudness, and energy. This paper aims to tackle this problem using machine learning. Particularly, we built several machine learning models using SVMs, LTSMs, and CNNs to classify emotions in human speeches. In addition, by leveraging transfer learning and data augmentation, we efficiently trained our models to attain decent performances on a relatively small dataset. Our best model was a ResNet34 network, which achieved an accuracy of $66.7%$ and an F1 score of $0.631$.

[1070] Amplifying Emotional Signals: Data-Efficient Deep Learning for Robust Speech Emotion Recognition

Tai Vu

Main category: eess.AS

TL;DR: This paper develops machine learning models for speech emotion recognition using transfer learning and data augmentation to overcome limited dataset challenges, achieving 66.7% accuracy with a ResNet34 model.

DetailsMotivation: Speech Emotion Recognition remains challenging despite deep learning advances, particularly due to performance limitations on small datasets which hinders human-computer interaction applications.

Method: Developed and evaluated multiple machine learning models including SVMs, LSTMs, and CNNs, employing transfer learning and innovative data augmentation techniques on RAVDESS and SAVEE datasets.

Result: ResNet34 architecture achieved the best performance with 66.7% accuracy and 0.631 F1 score, establishing a new benchmark on the combined datasets.

Conclusion: Transfer learning and data augmentation effectively overcome data scarcity in SER, enabling more robust and generalizable emotion recognition systems.

Abstract: Speech Emotion Recognition (SER) presents a significant yet persistent challenge in human-computer interaction. While deep learning has advanced spoken language processing, achieving high performance on limited datasets remains a critical hurdle. This paper confronts this issue by developing and evaluating a suite of machine learning models, including Support Vector Machines (SVMs), Long Short-Term Memory networks (LSTMs), and Convolutional Neural Networks (CNNs), for automated emotion classification in human speech. We demonstrate that by strategically employing transfer learning and innovative data augmentation techniques, our models can achieve impressive performance despite the constraints of a relatively small dataset. Our most effective model, a ResNet34 architecture, establishes a new performance benchmark on the combined RAVDESS and SAVEE datasets, attaining an accuracy of 66.7% and an F1 score of 0.631. These results underscore the substantial benefits of leveraging pre-trained models and data augmentation to overcome data scarcity, thereby paving the way for more robust and generalizable SER systems.

[1071] ChipChat: Low-Latency Cascaded Conversational Agent in MLX

Tatiana Likhomanenko, Luke Carlson, Richard He Bai, Zijin Gu, Han Tran, Zakaria Aldeneh, Yizhe Zhang, Ruixiang Zhang, Huangjie Zheng, Navdeep Jaitly

Main category: eess.AS

TL;DR: ChipChat is a novel low-latency cascaded system for real-time on-device voice agents that achieves sub-second response times through architectural innovations and streaming optimizations, outperforming end-to-end approaches while maintaining privacy through complete on-device processing.

DetailsMotivation: Large language models have transformed spoken dialog systems, but the optimal architecture for real-time on-device voice agents remains unclear. While end-to-end approaches have theoretical advantages, cascaded systems continue to outperform them in language understanding despite latency constraints from sequential processing.

Method: ChipChat integrates streaming components including: conversational speech recognition with mixture-of-experts, state-action augmented LLM, text-to-speech synthesis, neural vocoder, and speaker modeling. Implemented using MLX, it achieves low latency through architectural innovations and streaming optimizations.

Result: ChipChat achieves sub-second response latency on a Mac Studio without dedicated GPUs, demonstrating that strategically redesigned cascaded systems can overcome historical latency limitations while preserving user privacy through complete on-device processing.

Conclusion: The work shows that redesigned cascaded systems offer a promising path forward for practical voice-based AI agents, overcoming traditional bottlenecks and latency issues while maintaining the performance advantages of cascaded architectures over end-to-end approaches.

Abstract: The emergence of large language models (LLMs) has transformed spoken dialog systems, yet the optimal architecture for real-time on-device voice agents remains an open question. While end-to-end approaches promise theoretical advantages, cascaded systems (CSs) continue to outperform them in language understanding tasks, despite being constrained by sequential processing latency. In this work, we introduce ChipChat, a novel low-latency CS that overcomes traditional bottlenecks through architectural innovations and streaming optimizations. Our system integrates streaming (a) conversational speech recognition with mixture-of-experts, (b) state-action augmented LLM, (c) text-to-speech synthesis, (d) neural vocoder, and (e) speaker modeling. Implemented using MLX, ChipChat achieves sub-second response latency on a Mac Studio without dedicated GPUs, while preserving user privacy through complete on-device processing. Our work shows that strategically redesigned CSs can overcome their historical latency limitations, offering a promising path forward for practical voice-based AI agents.

[1072] Automatic Pronunciation Error Detection and Correction of the Holy Quran’s Learners Using Deep Learning

Abdullah Abdelfattah, Mahmoud I. Khalil, Hazem Abbas

Main category: eess.AS

TL;DR: Automated pipeline for creating high-quality Quranic speech datasets with 98% automation, producing 850+ hours of annotated audio and novel ASR-based pronunciation error detection using custom Quran Phonetic Script.

DetailsMotivation: Spoken language assessment is challenging, and Quran recitation has rigorous rules (tajweed) but lacks high-quality annotated data for machine learning models.

Method: Developed automated pipeline with expert recitation collection, segmentation using fine-tuned wav2vec2-BERT, transcription, verification via Tasmeea algorithm, and novel Quran Phonetic Script (QPS) with two-level encoding for phonemes and articulation characteristics.

Result: Created 850+ hours of audio (~300K annotated utterances), achieved 0.16% average Phoneme Error Rate with multi-level CTC model, and released all code/data/models as open-source.

Conclusion: Successfully bridged the data scarcity gap for Quranic speech assessment through automated pipeline and novel phonetic encoding, enabling effective pronunciation error detection for tajweed rules.

Abstract: Assessing spoken language is challenging, and quantifying pronunciation metrics for machine learning models is even harder. However, for the Holy Quran, this task is simplified by the rigorous recitation rules (tajweed) established by Muslim scholars, enabling highly effective assessment. Despite this advantage, the scarcity of high-quality annotated data remains a significant barrier. In this work, we bridge these gaps by introducing: (1) A 98% automated pipeline to produce high-quality Quranic datasets – encompassing: Collection of recitations from expert reciters, Segmentation at pause points (waqf) using our fine-tuned wav2vec2-BERT model, Transcription of segments, Transcript verification via our novel Tasmeea algorithm; (2) 850+ hours of audio (~300K annotated utterances); (3) A novel ASR-based approach for pronunciation error detection, utilizing our custom Quran Phonetic Script (QPS) to encode Tajweed rules (unlike the IPA standard for Modern Standard Arabic). QPS uses a two-level script: (Phoneme level): Encodes Arabic letters with short/long vowels. (Sifa level): Encodes articulation characteristics of every phoneme. We further include comprehensive modeling with our novel multi-level CTC Model which achieved 0.16% average Phoneme Error Rate (PER) on the testset. We release all code, data, and models as open-source: https://obadx.github.io/prepare-quran-dataset/

[1073] Quantum-Enhanced Analysis and Grading of Vocal Performance

Rohan Agarwal

Main category: eess.AS

TL;DR: QuantumMelody is a hybrid quantum-classical method for singing assessment that uses quantum circuits to encode vocal features and achieves 74.29% agreement with expert graders, showing a 12.86% improvement over classical methods.

DetailsMotivation: To develop an interpretable, objective singing assessment system that can provide technique-level feedback, moving beyond subjective human evaluation in audio signal processing.

Method: Encodes grouped vocal features (pitch stability, dynamics, timbre) into a 9-qubit simulated quantum circuit with Hadamard initialization, Rx/Ry/Rz rotations, and intra/cross-group entanglement. Fuses quantum measurement probabilities with spectrogram transformer embeddings for grading.

Result: Achieved 74.29% agreement with expert graders on 168 labeled 20-second excerpts, representing a +12.86 point gain over classical-features baseline. Processing takes sub-minute per recording on laptop-class Qiskit simulator.

Conclusion: Demonstrates feasibility of hybrid quantum-classical approaches for interpretable singing assessment, though hardware speedups are not claimed. Represents a step toward objective evaluation in applied audio processing.

Abstract: We present QuantumMelody, a hybrid quantum-classical method for objective singing assessment. Grouped vocal features (pitch stability, dynamics, timbre) are encoded into a small simulated quantum circuit; all nine qubits are initialized with a Hadamard on each qubit and then receive Rx, Ry, and Rz rotations, with intra- and cross-group entanglement. The circuit measurement probabilities are fused with spectrogram transformer embeddings to estimate a grade on labels 2-5 and to surface technique-level feedback. On 168 labeled 20 second excerpts, the hybrid reaches 74.29% agreement with expert graders, a +12.86 point gain over a classical-features baseline. Processing is sub-minute per recording on a laptop-class Qiskit simulator; we do not claim hardware speedups. This is a feasibility step toward interpretable, objective singing assessment in applied audio signal processing.

[1074] Deep Learning for Personalized Binaural Audio Reproduction

Xikun Lu, Yunda Chen, Zehua Chen, Jie Wang, Mingxing Liu, Hongmei Hu, Chengshi Zheng, Stefan Bleeck, Jinqiu Sang

Main category: eess.AS

TL;DR: Survey of deep learning approaches for personalized binaural audio reproduction, categorized into explicit personalized filtering and end-to-end rendering methods, with discussion of datasets, evaluation metrics, applications, and future directions.

DetailsMotivation: Personalized binaural audio is crucial for realistic spatial localization, sound externalization, and immersive listening experiences, directly impacting user experience and listening effort.

Method: Two main paradigms: 1) Explicit personalized filtering - predicting personalized HRTFs from sparse measurements, morphological features, or environmental cues for conventional rendering; 2) End-to-end rendering - mapping source signals directly to binaural signals with visual, textual, or parametric guidance, learning personalization within the model.

Result: Comprehensive review of recent advances in deep learning for personalized binaural audio, including organization of methods, summary of datasets and evaluation metrics for fair comparison.

Conclusion: Discussion of key applications enabled by these technologies, current technical limitations, and potential research directions for deep learning-based spatial audio systems.

Abstract: Personalized binaural audio reproduction is the basis of realistic spatial localization, sound externalization, and immersive listening, directly shaping user experience and listening effort. This survey reviews recent advances in deep learning for this task and organizes them by generation mechanism into two paradigms: explicit personalized filtering and end-to-end rendering. Explicit methods predict personalized head-related transfer functions (HRTFs) from sparse measurements, morphological features, or environmental cues, and then use them in the conventional rendering pipeline. End-to-end methods map source signals directly to binaural signals, aided by other inputs such as visual, textual, or parametric guidance, and they learn personalization within the model. We also summarize the field’s main datasets and evaluation metrics to support fair and repeatable comparison. Finally, we conclude with a discussion of key applications enabled by these technologies, current technical limitations, and potential research directions for deep learning-based spatial audio systems.

[1075] Speaker-Conditioned Phrase Break Prediction for Text-to-Speech with Phoneme-Level Pre-trained Language Model

Dong Yang, Yuki Saito, Takaaki Saeki, Tomoki Koriyama, Wataru Nakata, Detai Xin, Hiroshi Saruwatari

Main category: eess.AS

TL;DR: This paper improves phrase break prediction in multi-speaker TTS by integrating speaker embeddings and phoneme-level language models, achieving better performance through both objective and subjective evaluations.

DetailsMotivation: To enhance phrase break prediction (phrasing) in multi-speaker text-to-speech systems by leveraging speaker-specific features and advanced language modeling techniques.

Method: Integrated speaker embeddings to capture speaker characteristics, explored few-shot adaptation for unseen speakers, and applied phoneme-level pre-trained language models to the phrasing task.

Result: The methods significantly boosted phrasing model accuracy and were validated through rigorous objective and subjective evaluations.

Conclusion: The proposed approach effectively improves phrase break prediction in multi-speaker TTS systems by leveraging speaker embeddings and phoneme-level language models, demonstrating strong performance across various evaluation metrics.

Abstract: This paper advances phrase break prediction (also known as phrasing) in multi-speaker text-to-speech (TTS) systems. We integrate speaker-specific features by leveraging speaker embeddings to enhance the performance of the phrasing model. We further demonstrate that these speaker embeddings can capture speaker-related characteristics solely from the phrasing task. Besides, we explore the potential of pre-trained speaker embeddings for unseen speakers through a few-shot adaptation method. Furthermore, we pioneer the application of phoneme-level pre-trained language models to this TTS front-end task, which significantly boosts the accuracy of the phrasing model. Our methods are rigorously assessed through both objective and subjective evaluations, demonstrating their effectiveness.

[1076] MPO: Multidimensional Preference Optimization for Language Model-based Text-to-Speech

Kangxiang Xia, Xinfa Zhu, Jixun Yao, Lei Xie

Main category: eess.AS

TL;DR: Proposes Multidimensional Preference Optimization (MPO) to improve text-to-speech systems by aligning with multiple human preference dimensions while addressing overconfidence issues in reward-based approaches.

DetailsMotivation: Current TTS systems struggle with optimizing across multiple preference dimensions and suffer from performance degradation due to overconfidence in rewards from human feedback integration.

Method: Introduces MPO with a preference set for streamlined multidimensional preference data construction and incorporates regularization during training to prevent degradation issues common in DPO-based approaches.

Result: Experiments show MPO achieves significant improvements in intelligibility, speaker similarity, and prosody compared to baseline TTS systems.

Conclusion: MPO effectively addresses multidimensional preference optimization challenges in TTS, providing robust alignment with human preferences across multiple quality dimensions.

Abstract: In recent years, text-to-speech (TTS) has seen impressive advancements through large-scale language models, achieving human-level speech quality. Integrating human feedback has proven effective for enhancing robustness in these systems. However, current approaches face challenges in optimizing TTS with preference data across multiple dimensions and often suffer from performance degradation due to overconfidence in rewards. We propose Multidimensional Preference Optimization (MPO) to better align TTS systems with human preferences. MPO introduces a preference set that streamlines the construction of data for multidimensional preference optimization, enabling alignment with multiple dimensions. Additionally, we incorporate regularization during training to address the typical degradation issues in DPO-based approaches. Our experiments demonstrate MPO’s effectiveness, showing significant improvements in intelligibility, speaker similarity, and prosody compared to baseline systems.

[1077] Noisy Disentanglement with Tri-stage Training for Noise-Robust Speech Recognition

Shuangyuan Chen, Shuang Wei, Dongxing Xu, Yanhua Long

Main category: eess.AS

TL;DR: NoisyD-CT is a tri-stage training framework using Conformer-Transducer architecture with a compact noisy disentanglement module that suppresses noise while preserving speech features, achieving significant WER reductions in noisy conditions.

DetailsMotivation: To improve end-to-end speech recognition performance in noisy or low SNR conditions by effectively suppressing noise while maintaining essential acoustic and linguistic features.

Method: A tri-stage training framework with a compact NoisyD module (1.71M parameters) integrated between Conformer blocks and Transducer Decoder, using clean representation consistency loss and noisy reconstruction loss for noise suppression and feature preservation.

Result: Achieved up to 25.7% and 10.6% relative word error rate reductions on simulated and real-world noisy test sets (LibriSpeech and CHiME-4 datasets) while maintaining or improving performance on clean speech.

Conclusion: NoisyD-CT effectively enhances ASR robustness in challenging acoustic noise environments through noise suppression and feature preservation, significantly outperforming baseline models in noisy conditions.

Abstract: To enhance the performance of end-to-end (E2E) speech recognition systems in noisy or low signal-to-noise ratio (SNR) conditions, this paper introduces NoisyD-CT, a novel tri-stage training framework built on the Conformer-Transducer architecture. The core of NoisyD-CT is a especially designed compact noisy disentanglement (NoisyD) module (adding only 1.71M parameters), integrated between the Conformer blocks and Transducer Decoder to perform deep noise suppression and improve ASR robustness in challenging acoustic noise environments. To fully exploit the noise suppression capability of the NoisyD-CT, we further propose a clean representation consistency loss to align high-level representations derived from noisy speech with those obtained from corresponding clean speech. Together with a noisy reconstruction loss, this consistency alignment enables the NoisyD module to effectively suppress noise while preserving essential acoustic and linguistic features consistent across both clean and noisy conditions, thereby producing cleaner internal representations that enhance ASR performance. Moreover, our tri-stage training strategy is designed to fully leverage the functionalities of both the noisy disentanglement and speech recognition modules throughout the model training process, ultimately maximizing performance gains under noisy conditions. Our experiments are performed on the LibriSpeech and CHiME-4 datasets, extensive results demonstrate that our proposed NoisyD-CT significantly outperforms the competitive Conformer-Transducer baseline, achieving up to 25.7% and 10.6% relative word error rate reductions on simulated and real-world noisy test sets, respectively, while maintaining or even improving performance on clean speech test sets. The source code, model checkpoint and data simulation scripts will be available at https://github.com/litchimo/NoisyD-CT.

[1078] MixedG2P-T5: G2P-free Speech Synthesis for Mixed-script texts using Speech Self-Supervised Learning and Language Model

Joonyong Park, Daisuke Saito, Nobuaki Minematsu

Main category: eess.AS

TL;DR: Novel voice synthesis method replaces traditional G2P conversion with deep learning model that generates discrete tokens directly from speech using pre-trained SSL model and T5 encoder.

DetailsMotivation: Eliminate need for manual phonetic transcription to reduce costs and enhance scalability, especially for large non-transcribed audio datasets with mixed-script texts.

Method: Utilize pre-trained voice SSL model to train T5 encoder for producing pseudo-language labels directly from mixed-script texts (e.g., Kanji and Kana), bypassing traditional G2P conversion.

Result: Model matches performance of conventional G2P-based TTS systems and synthesizes speech retaining natural linguistic and paralinguistic features like accents and intonations.

Conclusion: Deep learning approach successfully substitutes traditional G2P conversion, enabling cost-effective and scalable voice synthesis while maintaining speech quality and natural features.

Abstract: This study presents a novel approach to voice synthesis that can substitute the traditional grapheme-to-phoneme (G2P) conversion by using a deep learning-based model that generates discrete tokens directly from speech. Utilizing a pre-trained voice SSL model, we train a T5 encoder to produce pseudo-language labels from mixed-script texts (e.g., containing Kanji and Kana). This method eliminates the need for manual phonetic transcription, reducing costs and enhancing scalability, especially for large non-transcribed audio datasets. Our model matches the performance of conventional G2P-based text-to-speech systems and is capable of synthesizing speech that retains natural linguistic and paralinguistic features, such as accents and intonations.

[1079] Characterization of Speech Similarity Between Australian Aboriginal and High-Resource Languages: A Case Study on Dharawal

Ting Dang, Trini Manoj Jeyaseelan, Eliathamby Ambikairajah, Vidhyasaharan Sethu

Main category: eess.AS

TL;DR: This paper analyzes speech similarity between Dharawal (an Australian Aboriginal language) and 107 high-resource languages using embedding-based methods to guide transfer learning for low-resource speech AI.

DetailsMotivation: Australian Aboriginal languages are severely underrepresented in speech AI systems, and existing models struggle with low-resource languages lacking clean annotated data. The paper aims to support endangered language communities through better speech technology.

Method: Collected and cleaned Dharawal speech dataset from public recordings, then used pre-trained multilingual speech encoder to analyze similarity through misclassification rate analysis, cosine similarity, and Fréchet Inception Distance in embedding space.

Result: Dharawal shows strong speech similarity with Latin, Māori, Korean, Thai, and Welsh, providing practical guidance for transfer learning and model adaptation.

Conclusion: The findings highlight the importance of data collection and embedding-based analysis for developing speech technologies for endangered languages, offering concrete directions for future transfer learning approaches.

Abstract: Australian Aboriginal languages are of significant cultural and linguistic value but remain severely underrepresented in modern speech AI systems. While state-of-the-art speech foundation models and automatic speech recognition excel in high-resource settings, they often struggle to generalize to low-resource languages, especially those lacking clean, annotated speech data. In this work, we collect and clean a speech dataset for Dharawal, a low-resource Australian Aboriginal language, by carefully sourcing and processing publicly available recordings. Using this dataset, we analyze the speech similarity between Dharawal and 107 high-resource languages using a pre-trained multilingual speech encoder. Our approach combines (1) misclassification rate analysis to assess language confusability, and (2) fine-grained similarity measurements using cosine similarity and Fr'echet Inception Distance (FID) in the embedding space. Experimental results reveal that Dharawal shares strong speech similarity with languages such as Latin, M=aori, Korean, Thai, and Welsh. These findings offer practical guidance for future transfer learning and model adaptation efforts, and underscore the importance of data collection and embedding-based analysis in supporting speech technologies for endangered language communities.

[1080] AHAMask: Reliable Task Specification for Large Audio Language Models without Instructions

Yiwei Guo, Bohan Li, Hankun Wang, Zhihan Li, Shuai Wang, Xie Chen, Kai Yu

Main category: eess.AS

TL;DR: AHAMask proposes masking specific attention heads in audio language models to trigger acoustic task functionalities without instructions, achieving comparable or better performance than instruction-based approaches.

DetailsMotivation: Current large audio language models suffer from instruction sensitivity where different instructions with the same intention yield drastically different outcomes, making task specification unreliable.

Method: Train masks on attention heads in the decoder-only LLM backbone of LALMs, with trainable parameters equal to the attention head count, to selectively mask heads and trigger specific acoustic task functionalities.

Result: Applying selective attention head masks achieves comparable or even better performance than using instructions, both on single and composite tasks.

Conclusion: The approach provides reliable acoustic task specification for LALMs and reveals that these models exhibit certain ‘functional pathways’ in their attention heads.

Abstract: Although current large audio language models (LALMs) extend text large language models (LLMs) with generic acoustic understanding abilities, they usually suffer from instruction sensitivity, where different instructions of the same intention can yield drastically different outcomes. In this work, we propose AHAMask, where we simply mask some of the attention heads in the decoder-only LLM backbone of LALMs, to trigger specific acoustic task functionalities without instructions. These masks are efficiently obtained by training on an LALM, with the number of trainable parameters equal to the attention head count in its LLM backbone. We show by experiments that applying such selective attention head masks achieves comparable or even better performance than using instructions, either on single or composite tasks. Besides achieving reliable acoustic task specification for LALMs, this also reveals that LALMs exhibit certain “functional pathways” in their attention heads.

[1081] From Evaluation to Optimization: Neural Speech Assessment for Downstream Applications

Yu Tsao

Main category: eess.AS

TL;DR: Review of neural network-based speech assessment models that serve as differentiable perceptual proxies for optimizing speech enhancement/synthesis and enable detection of salient speech characteristics for downstream processing.

DetailsMotivation: Traditional objective metrics have weak correlation with human perception, creating a gap between system optimization and actual user experience. Subjective tests are costly and not scalable for modern speech technology development cycles.

Method: Neural network-based speech assessment models that predict quality and intelligibility, focusing on their dual role as differentiable perceptual proxies for optimization and as tools for detecting salient speech characteristics.

Result: These models achieve promising results in bridging the perceptual gap and are increasingly integrated into downstream speech processing tasks.

Conclusion: Speech assessment models show significant potential but current limitations need addressing. Future research should focus on advancing their integration into speech processing pipelines.

Abstract: The evaluation of synthetic and processed speech has long been a cornerstone of audio engineering and speech science. Although subjective listening tests remain the gold standard for assessing perceptual quality and intelligibility, their high cost, time requirements, and limited scalability present significant challenges in the rapid development cycles of modern speech technologies. Traditional objective metrics, while computationally efficient, often exhibit weak correlation with human perception, creating a perceptual gap between system optimization and actual user experience. Bridging this gap requires speech assessment models that are more closely aligned with human perception. In recent years, numerous neural network-based speech assessment models have been developed to predict quality and intelligibility, achieving promising results. Beyond their role in evaluation, these models are increasingly integrated into downstream speech processing tasks. This review focuses on their role in two main areas: (1) serving as differentiable perceptual proxies that not only assess but also guide the optimization of speech enhancement and synthesis models; and (2) enabling the detection of salient speech characteristics to support more precise and efficient downstream processing. Finally, we discuss current limitations and outline future research directions to further advance the integration of speech assessment into speech processing pipelines.

[1082] Multilingual Speech Recognition Using Discrete Tokens with a Two-step Training Strategy

Zehan Li, Yan Yang, Xueqing Li, Jian Kang, Xiao-Lei Zhang, Jie Li

Main category: eess.AS

TL;DR: Two-stage training strategy improves discrete token performance in multilingual ASR, achieving 44% CER reduction and surpassing previous benchmarks.

DetailsMotivation: SSL models show great ASR performance but discrete units offer storage efficiency and broader applications. However, multilingual ASR faces challenges as different model layers contribute differently across languages, making discrete unit modeling unification difficult.

Method: Proposed a two-stage training strategy to enhance discrete token performance of pre-trained models and bridge the gap with continuous representation performance. Validated on XLS-R model following Interspeech2024 challenge settings.

Result: Achieved 44% relative CER reduction on ML-SUPERB dataset, surpassing WavLM’s 26% reduction. Achieved first place among all single-system results on the leaderboard.

Conclusion: The two-stage training strategy effectively improves discrete token performance in multilingual ASR, demonstrating significant performance gains and establishing new state-of-the-art results.

Abstract: Pre-trained models, especially self-supervised learning (SSL) models, have demonstrated impressive results in automatic speech recognition (ASR) task. While most applications of SSL models focus on leveraging continuous representations as features for training downstream tasks, the utilization of discrete units has gained increasing attention in recent years owing to its lower storage requirements and broader range of applications. In multilingual ASR tasks, representations at different layers of the model contribute differently to various languages, complicating the unification of discrete unit modeling. In this paper, we propose a two-stage training strategy to improve the discrete token performance of pre-trained models and narrow the gap with continuous representation performance. We validate our method on the XLS-R model following the settings of Interspeech2024 Speech Processing Using Discrete Speech Unit Challenge. Our method demonstrates a significant improvement on the ML-SUPERB dataset, achieving a 44% relative reduction on CER for the XLS-R model. This surpasses the previous baseline set by the WavLM model, which achieves a 26% relative reduction on CER. Furthermore, our method achieves the first place among all the single-system results on the leaderboard.

[1083] Binaural Unmasking in Practical Use: Perceived Level of Phase-inverted Speech in Environmental Noise

Rina Kotani, Chiaki Miyazaki, Shiro Suzuki

Main category: eess.AS

TL;DR: Phase reversal in one ear improves speech audibility in noisy environments by up to 6 dB through binaural unmasking, without increasing sound pressure.

DetailsMotivation: Develop technology to make earphone/headphone sound easier to hear without increasing volume or eliminating ambient noise, focusing on practical everyday scenarios.

Method: Conducted experiments using various speakers (including women) and real-life noises (urban sounds, cheers) to evaluate binaural unmasking through phase reversal in one ear under practical conditions.

Result: Speech in noisy environments perceived up to 6 dB louder with phase reversal; 5+ dB improvement achieved for all speakers and noises tested in Japanese language.

Conclusion: Binaural unmasking through interaural phase differences is effective in practical scenarios for improving audibility without volume increase.

Abstract: We aim to develop a technology that makes the sound from earphones and headphones easier to hear without increasing the sound pressure or eliminating ambient noise. To this end, we focus on harnessing the phenomenon of binaural unmasking through phase reversal in one ear. Specifically, we conduct experiments to evaluate the improvement of audibility caused by the phenomenon, using conditions that approximate practical scenarios. We use speech sounds by various speakers, including women, and noises that can be encountered in daily life (urban environmental sounds, cheers) to verify the effects of binaural unmasking under conditions close to practical situations. The results of experiments using the Japanese language showed that (i) speech in a noisy environment is perceived to be up to about 6 dB louder with phase reversal in one ear, and (ii) a certain effect (improvement of audibility by 5 dB or more) is obtained for all speakers and noises targeted in this study. These findings demonstrate the effectiveness of binaural unmasking attributed to interaural phase differences in practical scenarios.

[1084] Group Relative Policy Optimization for Speech Recognition

Prashanth Gurunath Shivakumar, Yile Gu, Ankur Gandhe, Ivan Bulyko

Main category: eess.AS

TL;DR: Proposes Group Relative Policy Optimization (GRPO) for ASR to address limitations of next-token prediction in LLMs, achieving significant WER improvements and reduced hallucinations.

DetailsMotivation: Current LLM-based speech recognition suffers from limitations in next-token prediction objectives, leading to performance issues and hallucinations that need addressing.

Method: Applies Group Relative Policy Optimization (GRPO) with reinforcement learning from human feedback, using simple rule-based reward functions to guide policy updates.

Result: Achieved up to 18.4% relative improvement in word error rate, reduced hallucinations, increased robustness on out-of-domain datasets, and effective domain adaptation.

Conclusion: GRPO effectively enhances ASR performance by addressing LLM limitations through reinforcement learning with human feedback and rule-based rewards.

Abstract: Speech Recognition has seen a dramatic shift towards adopting Large Language Models (LLMs). This shift is partly driven by good scalability properties demonstrated by LLMs, ability to leverage large amounts of labelled, unlabelled speech and text data, streaming capabilities with auto-regressive framework and multi-tasking with instruction following characteristics of LLMs. However, simple next-token prediction objective, typically employed with LLMs, have certain limitations in performance and challenges with hallucinations. In this paper, we propose application of Group Relative Policy Optimization (GRPO) to enable reinforcement learning from human feedback for automatic speech recognition (ASR). We design simple rule based reward functions to guide the policy updates. We demonstrate significant improvements in word error rate (upto 18.4% relative), reduction in hallucinations, increased robustness on out-of-domain datasets and effectiveness in domain adaptation.

[1085] SyncNet: correlating objective for time delay estimation in audio signals

Akshay Raina, Vipul Arora

Main category: eess.AS

TL;DR: SyncNet: Deep neural network approach for robust time-delay estimation in noisy/reverberant environments by transforming signals to sequences with high cross-correlation at actual delay.

DetailsMotivation: Address limitations of traditional signal processing methods for time-delay estimation in challenging acoustic environments with noise and reverberation.

Method: Transform input signals using deep neural network into sequences that show high cross-correlation at actual time delay, using novel correlation-based objective function while preserving temporal information.

Result: Outperforms classical approaches (GCC-PHAT) and other learning-based methods across various audio signals including pulses, speech, and musical beats.

Conclusion: Deep learning approach with correlation-based training provides robust and interpretable time-delay estimation superior to traditional methods in noisy environments.

Abstract: This study addresses the task of performing robust and reliable time-delay estimation in signals in noisy and reverberating environments. In contrast to the popular signal processing based methods, this paper proposes to transform the input signals using a deep neural network into another pair of sequences which show high cross correlation at the actual time delay. This is achieved with the help of a novel correlation function based objective function for training the network. The proposed approach is also intrinsically interpretable as it does not lose temporal information. Experimental evaluations are performed for estimating mutual time delays for different types of audio signals such as pulse, speech and musical beats. SyncNet outperforms other classical approaches, such as GCC-PHAT, and some other learning based approaches.

[1086] Voice Conversion Augmentation for Speaker Recognition on Defective Datasets

Ruijie Tao, Zhan Shi, Yidi Jiang, Tianchi Liu, Haizhou Li

Main category: eess.AS

TL;DR: VCA-NN strategy uses voice conversion to generate pseudo speech from nearest neighbors to address dataset imperfections in speaker recognition systems.

DetailsMotivation: Real-world speaker recognition often faces defective datasets (partially-labeled, small-scale, imbalanced) that previous works addressed with scenario-specific algorithmic solutions rather than tackling the root cause of dataset imperfections.

Method: Proposed Voice Conversion Augmentation (VCA) strategy to generate pseudo speech from training data, with VCA-NN selecting source speech from nearest neighbors in representation space to ensure generation quality.

Result: Experimental results on three created datasets demonstrated that VCA-NN effectively mitigates dataset problems in speaker recognition.

Conclusion: Provides a new direction for handling speaker recognition problems from the data aspect rather than algorithmic perspective, offering a unified solution for various dataset imperfections.

Abstract: Modern speaker recognition system relies on abundant and balanced datasets for classification training. However, diverse defective datasets, such as partially-labelled, small-scale, and imbalanced datasets, are common in real-world applications. Previous works usually studied specific solutions for each scenario from the algorithm perspective. However, the root cause of these problems lies in dataset imperfections. To address these challenges with a unified solution, we propose the Voice Conversion Augmentation (VCA) strategy to obtain pseudo speech from the training set. Furthermore, to guarantee generation quality, we designed the VCA-NN~(nearest neighbours) strategy to select source speech from utterances that are close to the target speech in the representation space. Our experimental results on three created datasets demonstrated that VCA-NN effectively mitigates these dataset problems, which provides a new direction for handling the speaker recognition problems from the data aspect.

[1087] Spatial Audio Signal Enhancement: A Multi-output MVDR Method in The Spherical Harmonic-domain

Huawei Zhang, Jihui Zhang, Huiyuan Sun, Prasanga Samarasinghe

Main category: eess.AS

TL;DR: SH-domain MVDR-based spatial audio enhancer using Relative Harmonic Coefficients achieves better performance than baseline methods in reverberant environments.

DetailsMotivation: Existing spatial audio enhancement methods rely on impractical assumptions or have limited applicability, requiring a more robust solution for real-world reverberant environments.

Method: Proposes a spherical harmonic (SH)-domain minimum variance distortionless response (MVDR)-based spatial signal enhancer using Relative Harmonic Coefficients (ReHCs) to extract clean SH coefficients from noisy recordings.

Result: Simulation study shows lower estimation error, higher speech-distortion-ratio (SDR), and comparable noise reduction within the sweet area compared to beamforming-and-projection baseline.

Conclusion: The proposed ReHC-based MVDR method provides effective spatial audio enhancement in reverberant environments with improved performance over existing approaches.

Abstract: Spatial audio signal enhancement aims to reduce interfering source contributions while preserving the desired sound field with its spatial cues. Existing methods generally rely on impractical assumptions (e.g. accurate estimations of impractical information) or have limited applicability. This paper presents a spherical harmonic (SH)-domain minimum variance distortionless response (MVDR)-based spatial signal enhancer using Relative Harmonic Coefficients (ReHCs) to extract clean SH coefficients from noisy recordings in reverberant environments. A simulation study shows the proposed method achieves lower estimation error, higher speech-distortion-ratio (SDR), and comparable noise reduction (NR) within the sweet area in a reverberant environment, compared to a beamforming-and-projection method as the baseline.

[1088] NeuroAMP: A Novel End-to-end General Purpose Deep Neural Amplifier for Personalized Hearing Aids

Shafique Ahmed, Ryandhimas E. Zezario, Hui-Guan Yuan, Amir Hussain, Hsin-Min Wang, Wei-Ho Chung, Yu Tsao

Main category: eess.AS

TL;DR: NeuroAMP is a deep neural network for end-to-end personalized hearing aid amplification that uses spectral features and audiogram inputs, with Transformer architecture achieving best performance. Denoising NeuroAMP extension adds noise reduction capabilities.

DetailsMotivation: Traditional hearing aid amplification methods face challenges due to complex integration of multiple modular components, requiring a more integrated end-to-end solution.

Method: Developed NeuroAMP with four architectures (CNN, LSTM, CRNN, Transformer) using spectral features and audiogram inputs. Created Denoising NeuroAMP for noise reduction. Used comprehensive data augmentation on TIMIT, TMHINT speech and Cadenza Challenge MUSIC datasets.

Result: Transformer architecture achieved best performance: SRCC scores of 0.9927 (HASQI) and 0.9905 (HASPI) on TIMIT, 0.9738 (HAAQI) on music. Denoising NeuroAMP outperformed conventional methods with 10% improvement in HASPI (0.90) and HASQI (0.59) scores.

Conclusion: NeuroAMP and Denoising NeuroAMP show significant potential for delivering improved personalized hearing aid amplification through end-to-end deep learning approaches with strong generalization capabilities.

Abstract: The prevalence of hearing aids is increasing. However, optimizing the amplification processes of hearing aids remains challenging due to the complexity of integrating multiple modular components in traditional methods. To address this challenge, we present NeuroAMP, a novel deep neural network designed for end-to-end, personalized amplification in hearing aids. NeuroAMP leverages both spectral features and the listener’s audiogram as inputs, and we investigate four architectures: Convolutional Neural Network (CNN), Long Short-Term Memory (LSTM), Convolutional Recurrent Neural Network (CRNN), and Transformer. We also introduce Denoising NeuroAMP, an extension that integrates noise reduction along with amplification capabilities for improved performance in real-world scenarios. To enhance generalization, a comprehensive data augmentation strategy was employed during training on diverse speech (TIMIT and TMHINT) and music (Cadenza Challenge MUSIC) datasets. Evaluation using the Hearing Aid Speech Perception Index (HASPI), Hearing Aid Speech Quality Index (HASQI), and Hearing Aid Audio Quality Index (HAAQI) demonstrates that the Transformer architecture within NeuroAMP achieves the best performance, with SRCC scores of 0.9927 (HASQI) and 0.9905 (HASPI) on TIMIT, and 0.9738 (HAAQI) on the Cadenza Challenge MUSIC dataset. Notably, our data augmentation strategy maintains high performance on unseen datasets (e.g., VCTK, MUSDB18-HQ). Furthermore, Denoising NeuroAMP outperforms both the conventional NAL-R+WDRC approach and a two-stage baseline on the VoiceBank+DEMAND dataset, achieving a 10% improvement in both HASPI (0.90) and HASQI (0.59) scores. These results highlight the potential of NeuroAMP and Denoising NeuroAMP to deliver notable improvements in personalized hearing aid amplification.

[1089] Speaker Targeting via Self-Speaker Adaptation for Multi-talker ASR

Weiqing Wang, Taejin Park, Ivan Medennikov, Jinhan Wang, Kunal Dhawan, He Huang, Nithin Rao Koluguri, Jagadeesh Balam, Boris Ginsburg

Main category: eess.AS

TL;DR: Self-speaker adaptation method for streaming multi-talker ASR that eliminates explicit speaker queries, using speaker-wise speech activity prediction and speaker-specific kernel injection for instantaneous adaptation to target speakers.

DetailsMotivation: Conventional multi-talker ASR approaches require explicit speaker queries, target speaker embeddings, or enrollment audio, which can be cumbersome and impractical in real-time streaming scenarios.

Method: Dynamic adaptation through speaker-wise speech activity prediction, injecting speaker-specific kernels generated via speaker supervision activations into selected ASR encoder layers.

Result: State-of-the-art performance in both offline and streaming scenarios, effectively handling fully overlapped speech even in streaming settings.

Conclusion: The self-speaker adaptation approach provides a robust solution for multi-talker ASR under severe overlapping speech conditions without requiring explicit speaker queries or enrollment data.

Abstract: We propose a self-speaker adaptation method for streaming multi-talker automatic speech recognition (ASR) that eliminates the need for explicit speaker queries. Unlike conventional approaches requiring target speaker embeddings or enrollment audio, our technique dynamically adapts individual ASR instances through speaker-wise speech activity prediction. The key innovation involves injecting speaker-specific kernels generated via speaker supervision activations into selected ASR encoder layers. This enables instantaneous speaker adaptation to target speakers while handling fully overlapped speech even in a streaming scenario. Experiments show state-of-the-art performance in both offline and streaming scenarios, demonstrating that our self-adaptive method effectively addresses severe speech overlap through streamlined speaker-focused recognition. The results validate the proposed self-speaker adaptation approach as a robust solution for multi-talker ASR under severe overlapping speech conditions.

[1090] Conan: A Chunkwise Online Network for Zero-Shot Adaptive Voice Conversion

Yu Zhang, Baotong Tian, Zhiyao Duan

Main category: eess.AS

TL;DR: Conan is a real-time zero-shot voice conversion model that uses streaming content extraction, adaptive style encoding, and causal vocoding to preserve semantic content while adapting to unseen speaker styles with low latency.

DetailsMotivation: Current voice conversion models struggle with semantic fidelity under real-time constraints, natural-sounding conversions, and adaptation to unseen speaker characteristics, limiting their practical applications in real-time communications.

Method: Three core components: 1) Stream Content Extractor using Emformer for low-latency streaming content encoding, 2) Adaptive Style Encoder for fine-grained stylistic feature extraction from reference speech, 3) Causal Shuffle Vocoder implementing fully causal HiFiGAN with pixel-shuffle mechanism.

Result: Experimental evaluations show Conan outperforms baseline models in both subjective and objective metrics, demonstrating superior performance in real-time voice conversion tasks.

Conclusion: Conan effectively addresses the challenges of zero-shot online voice conversion by combining streaming content extraction, adaptive style encoding, and causal vocoding, achieving state-of-the-art performance with practical real-time applications.

Abstract: Zero-shot online voice conversion (VC) holds significant promise for real-time communications and entertainment. However, current VC models struggle to preserve semantic fidelity under real-time constraints, deliver natural-sounding conversions, and adapt effectively to unseen speaker characteristics. To address these challenges, we introduce Conan, a chunkwise online zero-shot voice conversion model that preserves the content of the source while matching the voice timbre and styles of reference speech. Conan comprises three core components: 1) a Stream Content Extractor that leverages Emformer for low-latency streaming content encoding; 2) an Adaptive Style Encoder that extracts fine-grained stylistic features from reference speech for enhanced style adaptation; 3) a Causal Shuffle Vocoder that implements a fully causal HiFiGAN using a pixel-shuffle mechanism. Experimental evaluations demonstrate that Conan outperforms baseline models in subjective and objective metrics. Audio samples can be found at https://aaronz345.github.io/ConanDemo.

eess.IV

[1091] A Novel Method to Determine Total Oxidant Concentration Produced by Non-Thermal Plasma Based on Image Processing and Machine Learning

Mirkan Emir Sancak, Unal Sen, Ulker Diler Keris-Sen

Main category: eess.IV

TL;DR: Novel computer vision method using image processing and machine learning to accurately quantify oxidant concentration in plasma-treated water by analyzing color changes in potassium iodide solutions.

DetailsMotivation: Current titration methods for measuring total oxidant concentration in non-thermal plasma treated aqueous systems are subjective and struggle with transient reactive species, requiring an objective and accurate alternative.

Method: Developed a color-based computer analysis system that captures high-resolution video of KI solution color transitions, extracts RGB/HSV/Lab color features, and trains machine learning models (linear regression, gradient boosting) to predict oxidant concentration.

Result: Strong linear correlations found between color features and oxidant concentration, with ML models achieving R^2 > 0.990. Reduced feature set (9 to 4) maintained comparable performance with improved efficiency. Final system achieved R^2 > 0.998 accuracy compared to titration measurements.

Conclusion: The CBCA method provides an accurate, objective, and efficient alternative to conventional titration for quantifying total oxidant concentration in plasma-treated water systems, with machine learning models successfully predicting concentrations from colorimetric data.

Abstract: Accurate determination of total oxidant concentration ([Ox]{tot}) in non-thermal plasma (NTP)-treated aqueous systems remains a critical challenge due to the transient nature of reactive oxygen and nitrogen species and the subjectivity of conventional titration methods used for [Ox]{tot} determination. This study introduces a novel, color-based computer analysis (CBCA) method that integrates advanced image processing with machine learning (ML) to quantify colorimetric shifts in potassium iodide (KI) solutions during oxidation. First, a custom-built visual data acquisition system captured high-resolution video of the color transitions in a KI solution during oxidation with an NTP system. The change in [Ox]{tot} during the experiments was monitored with a standard titrimetric method. Second, the captured frames were processed using a robust image processing pipeline to extract RGB, HSV, and Lab color features. The extracted features were statistically evaluated, and the results revealed strong linear correlations with the measured [Ox]{tot} values, particularly in the saturation (HSV), a and b (Lab), and blue (RGB) channels. Subsequently, the [Ox]{tot} measurements and the extracted color features were used to train and validate five ML models. Among them, linear regression and gradient boosting models achieved the highest predictive accuracy (R^2 > 0.990). It was also found that reducing the feature set from nine to four resulted in comparable performance with improved prediction efficiency, especially for gradient boosting. Finally, comparison of the model predictions with real titration measurements revealed that the CBCA system successfully predicts the [Ox]{tot} in KI solution with high accuracy (R^2 > 0.998) even with a reduced number of features.

[1092] Promptable Longitudinal Lesion Segmentation in Whole-Body CT

Yannick Kirchhoff, Maximilian Rokuss, Fabian Isensee, Klaus H. Maier-Hein

Main category: eess.IV

TL;DR: Extending LongiSeg framework with promptable capabilities for longitudinal lesion tracking in CT scans, using large-scale pretraining on synthetic data to improve performance by up to 6 Dice points.

DetailsMotivation: Accurate longitudinal lesion segmentation is crucial for monitoring disease progression, but current automated methods struggle with consistent lesion tracking across time points.

Method: Extended LongiSeg framework with promptable capabilities (point and mask interactions) for lesion-specific tracking, plus large-scale pretraining on synthetic longitudinal CT data to overcome limited training data.

Result: Pretraining substantially improved longitudinal context exploitation, achieving up to 6 Dice points improvement compared to models trained from scratch.

Conclusion: Combining longitudinal context with interactive prompting enables robust lesion tracking, demonstrating effectiveness of the proposed approach for longitudinal medical image analysis.

Abstract: Accurate segmentation of lesions in longitudinal whole-body CT is essential for monitoring disease progression and treatment response. While automated methods benefit from incorporating longitudinal information, they remain limited in their ability to consistently track individual lesions across time. Task 2 of the autoPET/CT IV Challenge addresses this by providing lesion localizations and baseline delineations, framing the problem as longitudinal promptable segmentation. In this work, we extend the recently proposed LongiSeg framework with promptable capabilities, enabling lesion-specific tracking through point and mask interactions. To address the limited size of the provided training set, we leverage large-scale pretraining on a synthetic longitudinal CT dataset. Our experiments show that pretraining substantially improves the ability to exploit longitudinal context, yielding an improvement of up to 6 Dice points compared to models trained from scratch. These findings demonstrate the effectiveness of combining longitudinal context with interactive prompting for robust lesion tracking. Code is publicly available at https://github.com/MIC-DKFZ/LongiSeg/tree/autoPET.

[1093] Cepstrum-Based Texture Features for Melanoma Detection

Keith Miller, Tristan Crawford, Jason Hagerty, William Stoecker, Ronald J. Stanley

Main category: eess.IV

TL;DR: Novel cepstrum-based texture features combined with GLCM statistics improve melanoma classification performance on dermoscopic images.

DetailsMotivation: To develop improved texture analysis features for melanoma detection in dermoscopic images by leveraging cepstral representations, which are underutilized in image analysis.

Method: Proposed applying gray-level co-occurrence matrix (GLCM) statistics to 2D cepstral representations and combining them with established handcrafted lesion descriptors. Evaluated using XGBoost models on ISIC 2019 dataset.

Result: Incorporating cepstral features improved area under ROC curve, accuracy, and F1 score for binary melanoma vs. nevus classification.

Conclusion: Cepstral GLCM features provide complementary discriminatory information that enhances melanoma detection performance when combined with traditional features.

Abstract: This paper introduces a set of cepstrum-based texture features for melanoma classification and validates their performance on dermoscopic images from the ISIC 2019 dataset. We propose applying gray-level co-occurrence matrix (GLCM) statistics to 2D cepstral representations, a novel approach in image analysis. Combined with established handcrafted lesion descriptors, these features were evaluated using XGBoost models. Incorporating select cepstral features improved the area under the receiver operating characteristic curve, accuracy, and F1 score for binary melanoma vs. nevus classification. Results suggest that cepstral GLCM features offer complementary discriminatory information for melanoma detection.

[1094] Resting-state fMRI Analysis using Quantum Time-series Transformer

Junghoon Justin Park, Jungwoo Seo, Sangyoon Bae, Samuel Yen-Chi Chen, Huan-Hsin Tseng, Jiook Cha, Shinjae Yoo

Main category: eess.IV

TL;DR: Quantum-enhanced transformer for fMRI analysis with polylogarithmic complexity, outperforms classical transformers especially in small-sample scenarios while identifying clinically meaningful ADHD biomarkers.

DetailsMotivation: Classical transformer models struggle with quadratic complexity, large parameter counts, and substantial data requirements for fMRI analysis, limiting their practical application in neuroscience.

Method: Quantum Time-series Transformer using Linear Combination of Unitaries and Quantum Singular Value Transformation to achieve polylogarithmic computational complexity.

Result: Achieves comparable or superior predictive performance to state-of-the-art classical transformers on large-scale fMRI datasets (ABCD Study and UK Biobank), with especially pronounced gains in small-sample scenarios. Reliably identifies clinically meaningful ADHD biomarkers through SHAP interpretability analyses.

Conclusion: Quantum-enhanced transformers show promise for advancing computational neuroscience by efficiently modeling complex spatio-temporal dynamics and improving clinical interpretability with reduced computational overhead.

Abstract: Resting-state functional magnetic resonance imaging (fMRI) has emerged as a pivotal tool for revealing intrinsic brain network connectivity and identifying neural biomarkers of neuropsychiatric conditions. However, classical self-attention transformer models–despite their formidable representational power–struggle with quadratic complexity, large parameter counts, and substantial data requirements. To address these barriers, we introduce a Quantum Time-series Transformer, a novel quantum-enhanced transformer architecture leveraging Linear Combination of Unitaries and Quantum Singular Value Transformation. Unlike classical transformers, Quantum Time-series Transformer operates with polylogarithmic computational complexity, markedly reducing training overhead and enabling robust performance even with fewer parameters and limited sample sizes. Empirical evaluation on the largest-scale fMRI datasets from the Adolescent Brain Cognitive Development Study and the UK Biobank demonstrates that Quantum Time-series Transformer achieves comparable or superior predictive performance compared to state-of-the-art classical transformer models, with especially pronounced gains in small-sample scenarios. Interpretability analyses using SHapley Additive exPlanations further reveal that Quantum Time-series Transformer reliably identifies clinically meaningful neural biomarkers of attention-deficit/hyperactivity disorder (ADHD). These findings underscore the promise of quantum-enhanced transformers in advancing computational neuroscience by more efficiently modeling complex spatio-temporal dynamics and improving clinical interpretability.

[1095] Can General-Purpose Omnimodels Compete with Specialists? A Case Study in Medical Image Segmentation

Yizhe Zhang, Qiang Chen, Tao Zhou

Main category: eess.IV

TL;DR: Omnimodels vs specialist models in medical image segmentation: specialists excel on easy cases but omnimodels show better robustness on hard cases, with task-dependent performance.

DetailsMotivation: To investigate whether general-purpose omnimodels can match specialized models in knowledge-intensive domains like medical image segmentation.

Method: Comparative study analyzing zero-shot performance of Gemini 2.5 Pro omnimodel against domain-specific deep learning models on three medical segmentation tasks (polyp, retinal vessel, breast tumor), focusing on easiest and hardest case subsets.

Result: Task-dependent results: specialists excel on easy samples for polyp and breast tumor segmentation, but omnimodel shows greater robustness on hard cases. For retinal vessel segmentation, specialist maintains superiority across all cases. Omnimodels may identify subtle features missed by human annotators.

Conclusion: Current omnimodels are not universal replacements for specialists but offer complementary strengths, particularly for enhancing robustness on challenging edge cases.

Abstract: The emergence of powerful, general-purpose omnimodels capable of processing diverse data modalities has raised a critical question: can these jack-of-all-trades'' systems perform on par with highly specialized models in knowledge-intensive domains? This work investigates this question within the high-stakes field of medical image segmentation. We conduct a comparative study analyzing the zero-shot performance of a state-of-the-art omnimodel (Gemini 2.5 Pro, the Nano Banana’’ model) against domain-specific deep learning models on three distinct tasks: polyp (endoscopy), retinal vessel (fundus), and breast tumor segmentation (ultrasound). Our study focuses on performance at the extremes by curating subsets of the easiest'' and hardest’’ cases based on the specialist models’ accuracy. Our findings reveal a nuanced and task-dependent landscape. For polyp and breast tumor segmentation, specialist models excel on easy samples, but the omnimodel demonstrates greater robustness on hard samples where specialists fail catastrophically. Conversely, for the fine-grained task of retinal vessel segmentation, the specialist model maintains superior performance across both easy and hard cases. Intriguingly, qualitative analysis suggests omnimodels may possess higher sensitivity, identifying subtle anatomical features missed by human annotators. Our results indicate that while current omnimodels are not yet a universal replacement for specialists, their unique strengths suggest a potential complementary role with specialist models, particularly in enhancing robustness on challenging edge cases.

[1096] Towards Early Detection: AI-Based Five-Year Forecasting of Breast Cancer Risk Using Digital Breast Tomosynthesis Imaging

Manon A. Dorster, Felix J. Dorfner, Mason C. Cleveland, Melisa S. Guelen, Jay Patel, Dania Daye, Jean-Philippe Thiran, Albert E. Kim, Christopher P. Bridge

Main category: eess.IV

TL;DR: Deep learning model using DBT scans achieves 0.80 AUROC for 5-year breast cancer risk prediction, outperforming traditional methods.

DetailsMotivation: Current breast cancer risk prediction models have modest performance and don't incorporate digital breast tomosynthesis (DBT) imaging, which was FDA-approved in 2011. There's a need for improved screening optimization through better risk assessment tools.

Method: Deep learning framework using Meta AI DINOv2 image encoder to extract features from DBT scans, combined with a cumulative hazard layer. Trained on 161,753 DBT examinations from 50,590 patients.

Result: Best-performing model achieved AUROC of 0.80 for 5-year breast cancer risk predictions on held-out test set.

Conclusion: DBT-based deep learning approaches show high potential to complement traditional risk assessment tools and serve as a promising basis for further validation and enhancement.

Abstract: As early detection of breast cancer strongly favors successful therapeutic outcomes, there is major commercial interest in optimizing breast cancer screening. However, current risk prediction models achieve modest performance and do not incorporate digital breast tomosynthesis (DBT) imaging, which was FDA-approved for breast cancer screening in 2011. To address this unmet need, we present a deep learning (DL)-based framework capable of forecasting an individual patient’s 5-year breast cancer risk directly from screening DBT. Using an unparalleled dataset of 161,753 DBT examinations from 50,590 patients, we trained a risk predictor based on features extracted using the Meta AI DINOv2 image encoder, combined with a cumulative hazard layer, to assess a patient’s likelihood of developing breast cancer over five years. On a held-out test set, our best-performing model achieved an AUROC of 0.80 on predictions within 5 years. These findings reveal the high potential of DBT-based DL approaches to complement traditional risk assessment tools, and serve as a promising basis for additional investigation to validate and enhance our work.

[1097] Ultrasound-based detection and malignancy prediction of breast lesions eligible for biopsy: A multi-center clinical-scenario study using nomograms, large language models, and radiologist evaluation

Ali Abbasian Ardakani, Afshin Mohammadi, Taha Yusuf Kuzan, Beyza Nur Kuzan, Hamid Khorshidi, Ashkan Ghorbani, Alisa Mohebbi, Fariborz Faeghi, Sepideh Hatamikia, U Rajendra Acharya

Main category: eess.IV

TL;DR: Integrated ultrasound nomogram combining BIRADS features and morphometric characteristics outperforms radiologists and ChatGPT in breast lesion biopsy recommendation and malignancy prediction.

DetailsMotivation: To develop and validate integrated ultrasound tools that can improve biopsy decision-making and malignancy prediction for breast lesions, potentially reducing unnecessary biopsies.

Method: Retrospective multicenter study of 1747 women with pathologically confirmed breast lesions. Built BIRADS, morphometric, and fused nomograms via logistic regression. Compared performance with radiologists and ChatGPT models.

Result: Fused nomogram achieved highest accuracy (83.0% biopsy recommendation, 83.8% malignancy prediction) with AUCs of 0.901 and 0.853, outperforming radiologists and ChatGPT models. External validation confirmed generalizability.

Conclusion: Integrated BIRADS-morphometric nomogram consistently outperforms standalone models, LLMs, and radiologists, showing potential to reduce unnecessary biopsies and enhance personalized decision making in breast imaging.

Abstract: To develop and externally validate integrated ultrasound nomograms combining BIRADS features and quantitative morphometric characteristics, and to compare their performance with expert radiologists and state of the art large language models in biopsy recommendation and malignancy prediction for breast lesions. In this retrospective multicenter, multinational study, 1747 women with pathologically confirmed breast lesions underwent ultrasound across three centers in Iran and Turkey. A total of 10 BIRADS and 26 morphological features were extracted from each lesion. A BIRADS, morphometric, and fused nomogram integrating both feature sets was constructed via logistic regression. Three radiologists (one senior, two general) and two ChatGPT variants independently interpreted deidentified breast lesion images. Diagnostic performance for biopsy recommendation (BIRADS 4,5) and malignancy prediction was assessed in internal and two external validation cohorts. In pooled analysis, the fused nomogram achieved the highest accuracy for biopsy recommendation (83.0%) and malignancy prediction (83.8%), outperforming the morphometric nomogram, three radiologists and both ChatGPT models. Its AUCs were 0.901 and 0.853 for the two tasks, respectively. In addition, the performance of the BIRADS nomogram was significantly higher than the morphometric nomogram, three radiologists and both ChatGPT models for biopsy recommendation and malignancy prediction. External validation confirmed the robust generalizability across different ultrasound platforms and populations. An integrated BIRADS morphometric nomogram consistently outperforms standalone models, LLMs, and radiologists in guiding biopsy decisions and predicting malignancy. These interpretable, externally validated tools have the potential to reduce unnecessary biopsies and enhance personalized decision making in breast imaging.

[1098] DRetNet: A Novel Deep Learning Framework for Diabetic Retinopathy Diagnosis

Idowu Paul Okuwobi, Jingyuan Liu, Jifeng Wan, Jiaojiao Jiang

Main category: eess.IV

TL;DR: A novel diabetic retinopathy detection framework combining physics-informed image enhancement, hybrid feature fusion, and multi-stage classification with uncertainty quantification achieves 92.7% accuracy and high clinical relevance.

DetailsMotivation: Current automated DR detection systems struggle with poor-quality images, lack interpretability, and insufficient integration of domain-specific knowledge, necessitating improved solutions for early detection to prevent blindness.

Method: Three innovative components: (1) Physics-Informed Neural Networks for adaptive retinal image enhancement, (2) Hybrid Feature Fusion Network combining deep learning and handcrafted features, (3) Multi-stage classifier with uncertainty quantification for interpretable predictions.

Result: Achieves 92.7% accuracy, 92.5% precision, 92.6% recall, 92.5% F1-score, 97.8% AUC, 0.96 mAP, and 0.85 MCC. Ophthalmologists rated it 4.8/5 for clinical relevance. Robust performance across diverse conditions including low-quality images.

Conclusion: The framework demonstrates strong performance and high clinical trust, making it a promising tool for accurate and reliable diabetic retinopathy detection in resource-limited settings with enhanced interpretability.

Abstract: Diabetic retinopathy (DR) is a leading cause of blindness worldwide, necessitating early detection to prevent vision loss. Current automated DR detection systems often struggle with poor-quality images, lack interpretability, and insufficient integration of domain-specific knowledge. To address these challenges, we introduce a novel framework that integrates three innovative contributions: (1) Adaptive Retinal Image Enhancement Using Physics-Informed Neural Networks (PINNs): this technique dynamically enhances retinal images by incorporating physical constraints, improving the visibility of critical features such as microaneurysms, hemorrhages, and exudates; (2) Hybrid Feature Fusion Network (HFFN): by combining deep learning embeddings with handcrafted features, HFFN leverages both learned representations and domain-specific knowledge to enhance generalization and accuracy; (3) Multi-Stage Classifier with Uncertainty Quantification: this method breaks down the classification process into logical stages, providing interpretable predictions and confidence scores, thereby improving clinical trust. The proposed framework achieves an accuracy of 92.7%, a precision of 92.5%, a recall of 92.6%, an F1-score of 92.5%, an AUC of 97.8%, a mAP of 0.96, and an MCC of 0.85. Ophthalmologists rated the framework’s predictions as highly clinically relevant (4.8/5), highlighting its alignment with real-world diagnostic needs. Qualitative analyses, including Grad-CAM visualizations and uncertainty heatmaps, further enhance the interpretability and trustworthiness of the system. The framework demonstrates robust performance across diverse conditions, including low-quality images, noisy data, and unseen datasets. These features make the proposed framework a promising tool for clinical adoption, enabling more accurate and reliable DR detection in resource-limited settings.

[1099] Learn2Reg 2024: New Benchmark Datasets Driving Progress on New Challenges

Lasse Hansen, Wiebke Heyer, Christoph Großbröhmer, Frederic Madesta, Thilo Sentker, Wang Jiazheng, Yuxi Zhang, Hang Zhang, Min Liu, Junyi Wang, Xi Zhu, Yuhua Li, Liwen Wang, Daniil Morozov, Nazim Haouchine, Joel Honkamaa, Pekka Marttinen, Yichao Zhou, Zuopeng Tan, Zhuoyuan Wang, Yi Wang, Hongchao Zhou, Shunbo Hu, Yi Zhang, Qian Tao, Lukas Förner, Thomas Wendler, Bailiang Jian, Christian Wachinger, Jin Kim, Dan Ruan, Marek Wodzinski, Henning Müller, Tony C. W. Mok, Xi Jia, Mikael Brudfors, Seyed-Ahmad Ahmadi, Yunzheng Zhu, William Hsu, Tina Kapur, William M. Wells, Alexandra Golby, Aaron Carass, Harrison Bai, Yihao Liu, Perrine Paul-Gilloteaux, Joakim Lindblad, Nataša Sladoje, Andreas Walter, Junyu Chen, Reuben Dorent, Alessa Hering, Mattias P. Heinrich

Main category: eess.IV

TL;DR: Learn2Reg 2024 introduces three new registration tasks addressing modality diversity and task complexity gaps, including multi-modal registration, unsupervised brain registration, and microscopy benchmarks, with new method developments.

DetailsMotivation: Previous Learn2Reg challenges (2020-2023) lacked comprehensive coverage of medical image registration aspects, particularly in modality diversity and task complexity, limiting fair benchmarking of registration methods.

Method: The 2024 edition introduces three new tasks: large-scale multi-modal registration, unsupervised inter-subject brain registration, and microscopy-focused benchmark. New method developments include invertibility constraints, pyramid features, keypoints alignment, and instance optimization.

Result: The paper introduces expanded benchmarking capabilities through new datasets and tasks that capture previously unaddressed aspects of medical image registration, enabling more comprehensive evaluation of registration methods.

Conclusion: Learn2Reg 2024 addresses limitations of previous editions by introducing diverse new tasks and datasets that better represent the complexity of medical image registration problems, while inspiring novel method developments in the field.

Abstract: Medical image registration is critical for clinical applications, and fair benchmarking of different methods is essential for monitoring ongoing progress. To date, the Learn2Reg 2020-2023 challenges have released several complementary datasets and established metrics for evaluations. However, these editions did not capture all aspects of the registration problem, particularly in terms of modality diversity and task complexity. To address these limitations, the 2024 edition introduces three new tasks, including large-scale multi-modal registration and unsupervised inter-subject brain registration, as well as the first microscopy-focused benchmark within Learn2Reg. The new datasets also inspired new method developments, including invertibility constraints, pyramid features, keypoints alignment and instance optimisation.

[1100] Temporal Representation Learning for Real-Time Ultrasound Analysis

Yves Stebler, Thomas M. Sutter, Ece Ozkan, Julia E. Vogt

Main category: eess.IV

TL;DR: Proposes a temporal learning method for ultrasound videos using consistent masking and contrastive learning to improve ejection fraction estimation accuracy by capturing heart motion patterns.

DetailsMotivation: Current deep learning models analyze ultrasound frames independently, missing temporal dependencies crucial for assessing motion patterns in cardiac monitoring and other medical applications.

Method: Leverages temporally consistent masking and contrastive learning to enforce temporal coherence across ultrasound video frames, enhancing motion pattern representation.

Result: Achieves substantial improvement in ejection fraction prediction accuracy on the EchoNet-Dynamic dataset.

Conclusion: Demonstrates the importance of temporally-aware representation learning for real-time ultrasound analysis, particularly for cardiac function assessment.

Abstract: Ultrasound (US) imaging is a critical tool in medical diagnostics, offering real-time visualization of physiological processes. One of its major advantages is its ability to capture temporal dynamics, which is essential for assessing motion patterns in applications such as cardiac monitoring, fetal development, and vascular imaging. Despite its importance, current deep learning models often overlook the temporal continuity of ultrasound sequences, analyzing frames independently and missing key temporal dependencies. To address this gap, we propose a method for learning effective temporal representations from ultrasound videos, with a focus on echocardiography-based ejection fraction (EF) estimation. EF prediction serves as an ideal case study to demonstrate the necessity of temporal learning, as it requires capturing the rhythmic contraction and relaxation of the heart. Our approach leverages temporally consistent masking and contrastive learning to enforce temporal coherence across video frames, enhancing the model’s ability to represent motion patterns. Evaluated on the EchoNet-Dynamic dataset, our method achieves a substantial improvement in EF prediction accuracy, highlighting the importance of temporally-aware representation learning for real-time ultrasound analysis.

[1101] High-resolution single-pixel imaging in real time with iterative or deep learning-based reconstruction enhancement

Anna Pastuszczak, Rafał Stojek, Piotr Wróbel, Magdalena Cwojdzińska, Kacper Sobczak, Rafał Kotyński

Main category: eess.IV

TL;DR: A compressive single-pixel imaging framework that enables high-resolution (1024x768) image capture in fractions of a second using a dedicated sampling strategy and two-phase reconstruction approach.

DetailsMotivation: To achieve real-time, high-resolution dynamic imaging with single-pixel cameras by overcoming the traditional speed limitations of compressive imaging systems.

Method: Combines dedicated sampling strategy with two-phase reconstruction: first uses generalized inverse of measurement matrix for quick recovery, then leverages spatial sparsity with iterative methods or neural networks to enhance dense areas.

Result: Achieves 6.8 Hz image acquisition rate at 22 kHz DMD operation with 0.41% compression ratio, enabling real-time high-resolution reconstruction on mid-tier desktop GPU.

Conclusion: The framework successfully enables real-time, high-resolution compressive imaging that matches acquisition rates with reconstruction performance, making it suitable for dynamic imaging applications.

Abstract: We introduce a compressive single-pixel imaging (SPI) framework for high-resolution image capture in fractions of a second. This framework combines a dedicated sampling strategy with a tailored reconstruction method to enable high-quality imaging of spatially sparse scenes at the native 1024x768 resolution of a digital micromirror device (DMD). The reconstruction process consists of two phases: first, the measured data is processed using the generalized inverse of the measurement matrix for quick image recovery. Then, the spatial sparsity of the scene is leveraged to enhance reconstruction in dense areas, using either an iterative method or a neural network-based approach. With a compression ratio of 0.41% and an image acquisition rate of 6.8 Hz at 22 kHz DMD operation, this framework supports real-time, high-resolution dynamic imaging with the reconstruction that matches the acquisition rate on a mid-tier desktop GPU.

[1102] Optimizing Paths for Adaptive Fly-Scan Microscopy: An Extended Version

Yu Lu, Thomas F. Lynn, Ming Du, Zichao Di, Sven Leyffer

Main category: eess.IV

TL;DR: Proposes adaptive multi-iteration fly-scan framework for x-ray microscopy that optimizes scanning paths to focus on regions of interest, reducing scan time and radiation exposure while maintaining image quality.

DetailsMotivation: Traditional x-ray microscopy fly-scans waste time scanning uniform, uninteresting regions. Existing deep learning approaches for optimal path selection require large datasets and high computational costs, making them impractical.

Method: Multi-iteration framework with score function to identify ROIs and objective function to optimize anchor points. Computes shortest scanning path between optimized points, performs fly-scan, and uses image completion for reconstruction.

Result: Reduces scanning time and potentially decreases x-ray exposure dose while maintaining high-quality information in critical regions through adaptive ROI-focused scanning.

Conclusion: The proposed adaptive fly-scan framework provides an efficient alternative to traditional methods by dynamically focusing on important sample regions and using iterative optimization for improved scanning efficiency.

Abstract: In x-ray microscopy, traditional raster-scanning techniques are used to acquire a microscopic image in a series of step-scans. Alternatively, scanning the x-ray probe along a continuous path, called a fly-scan, reduces scan time and increases scan efficiency. However, not all regions of an image are equally important. Currently used fly-scan methods do not adapt to the characteristics of the sample during the scan, often wasting time in uniform, uninteresting regions. One approach to avoid unnecessary scanning in uniform regions for raster step-scans is to use deep learning techniques to select a shorter optimal scan path instead of a traditional raster scan path, followed by reconstructing the entire image from the partially scanned data. However, this approach heavily depends on the quality of the initial sampling, requires a large dataset for training, and incurs high computational costs. We propose leveraging the fly-scan method along an optimal scanning path, focusing on regions of interest (ROIs) and using image completion techniques to reconstruct details in non-scanned areas. This approach further shortens the scanning process and potentially decreases x-ray exposure dose while maintaining high-quality and detailed information in critical regions. To achieve this, we introduce a multi-iteration fly-scan framework that adapts to the scanned image. Specifically, in each iteration, we define two key functions: (1) a score function to generate initial anchor points and identify potential ROIs, and (2) an objective function to optimize the anchor points for convergence to an optimal set. Using these anchor points, we compute the shortest scanning path between optimized anchor points, perform the fly-scan, and subsequently apply image completion based on the acquired information in preparation for the next scan iteration.

[1103] autoPET IV challenge: Incorporating organ supervision and human guidance for lesion segmentation in PET/CT

Junwei Huang, Yingqi Hao, Yitong Luo, Ziyu Wang, Mingxuan Liu, Yifei Chen, Yuanhan Wang, Lei Xiang, Qiyuan Tian

Main category: eess.IV

TL;DR: autoPET IV introduces human-in-the-loop segmentation with tracer classification, organ supervision, and click guidance to improve PET/CT lesion segmentation accuracy.

DetailsMotivation: To address time-intensive manual annotation and high inter-observer variability in PET/CT lesion segmentation by incorporating human guidance into automated methods.

Method: Integrated tracer classification, organ supervision, and simulated clicks guidance into nnUNet Residual Encoder framework to create an interactive segmentation pipeline.

Result: The approach demonstrates robust performance in fully automated (zero-guidance) scenarios and efficiently leverages iterative interactions to progressively enhance segmentation accuracy.

Conclusion: The integrated pipeline successfully combines automated segmentation with human guidance, showing promise for improving oncological workflows through human-in-the-loop approaches.

Abstract: Lesion Segmentation in PET/CT scans is an essential part of modern oncological workflows. To address the challenges of time-intensive manual annotation and high inter-observer variability, the autoPET challenge series seeks to advance automated segmentation methods in complex multi-tracer and multi-center settings. Building on this foundation, autoPET IV introduces a human-in-the-loop scenario to efficiently utilize interactive human guidance in segmentation tasks. In this work, we incorporated tracer classification, organ supervision and simulated clicks guidance into the nnUNet Residual Encoder framework, forming an integrated pipeline that demonstrates robust performance in a fully automated (zero-guidance) context and efficiently leverages iterative interactions to progressively enhance segmentation accuracy.

[1104] HyDeFuse: Provably Convergent Denoiser-Driven Hyperspectral Fusion

Sagar Kumar, Unni V S, Kunal Narayan Chaudhury

Main category: eess.IV

TL;DR: HyDeFuse is a denoiser-driven HS-MS fusion algorithm that uses pseudo-linear denoisers with convergence guarantees, achieving competitive performance with state-of-the-art methods.

DetailsMotivation: HS-MS fusion combines hyperspectral and multispectral images to create high-resolution spectral-spatial data, but traditional methods struggle with convergence when using powerful denoisers for regularization.

Method: Proposes HyDeFuse algorithm using pseudo-linear denoisers for implicit regularization, applying contraction mapping theorem to ensure global linear convergence with enhanced denoiser design.

Result: HyDeFuse demonstrates stable convergence and competitive performance compared to state-of-the-art fusion techniques on public datasets.

Conclusion: The denoiser-driven approach with convergence guarantees enables high-quality HS-MS fusion, making HyDeFuse a reliable and effective solution for hyperspectral image enhancement.

Abstract: Hyperspectral (HS) images provide fine spectral resolution but have limited spatial resolution, whereas multispectral (MS) images capture finer spatial details but have fewer bands. HS-MS fusion aims to integrate HS and MS images to generate a single image with improved spatial and spectral resolution. This is commonly formulated as an inverse problem with a linear forward model. However, reconstructing high-quality images using the forward model alone is challenging, necessitating the use of regularization techniques. Over the years, numerous methods have been developed, including wavelets, total variation, low-rank models, and deep neural networks. In this work, we investigate the paradigm of denoiser-driven regularization, where a powerful off-the-shelf denoiser is used for implicit regularization within an iterative algorithm. This approach has shown promise but remains relatively underexplored in hyperspectral imaging. Our focus is on a crucial aspect of denoiser-driven algorithms: ensuring convergence of the iterations. It is known that powerful denoisers can produce high-quality reconstructions, but they are also prone to instability and can cause the iterations to diverge. The challenge is to design denoisers that come with a convergence guarantee. In this work, we propose a denoiser-driven fusion algorithm, HyDeFuse, which leverages a class of pseudo-linear denoisers for implicit regularization. We demonstrate how the contraction mapping theorem can be applied to establish global linear convergence of HyDeFuse. Additionally, we introduce enhancements to the denoiser that significantly improve the performance of HyDeFuse, making it competitive with state-of-the-art techniques. We validate our theoretical results and present fusion results on publicly available datasets to demonstrate the performance of HyDeFuse.

[1105] XVertNet: Unsupervised Contrast Enhancement of Vertebral Structures with Dynamic Self-Tuning Guidance and Multi-Stage Analysis

Ella Eidlin, Assaf Hoogi, Hila Rozen, Mohammad Badarne, Nathan S. Netanyahu

Main category: eess.IV

TL;DR: XVertNet is an unsupervised deep learning framework that enhances vertebral structure visualization in chest X-rays, outperforming state-of-the-art methods and improving detection of subtle fractures without requiring labeled training data.

DetailsMotivation: Chest X-rays have limited ability to capture fine anatomical details, leading to missed or delayed diagnoses in emergency medicine. The reliance on manually labeled training data is a persistent bottleneck in medical imaging.

Method: Unsupervised learning architecture with dynamic self-tuned internal guidance mechanism featuring an adaptive feedback loop for real-time image optimization. Eliminates need for manually labeled training data.

Result: Outperforms state-of-the-art enhancement methods across four major public datasets, with improvements in entropy scores, Tenengrad criterion values, LPC-SI, and TMQI. Clinical validation with radiologists confirmed more sensitive detection of subtle vertebral fractures and degenerative changes.

Conclusion: XVertNet represents a transformative advancement in emergency radiology, providing a scalable and time-efficient solution for enhanced diagnostic accuracy without requiring additional training overhead, facilitating immediate clinical deployment.

Abstract: Chest X-rays remain the primary diagnostic tool in emergency medicine, yet their limited ability to capture fine anatomical details can result in missed or delayed diagnoses. To address this, we introduce XVertNet, a novel deep-learning framework designed to enhance vertebral structure visualization in X-ray images significantly. Our framework introduces two key innovations: (1) An unsupervised learning architecture that eliminates reliance on manually labeled training data a persistent bottleneck in medical imaging, and (2) a dynamic self-tuned internal guidance mechanism featuring an adaptive feedback loop for real-time image optimization. Extensive validation across four major public datasets revealed that XVertNet outperforms state-of-the-art enhancement methods, as demonstrated by improvements in entropy scores, Tenengrad criterion values, the local phase coherence sharpness index (LPC-SI), and thetone mapped image quality index (TMQI). Furthermore, clinical validation conducted with two board-certified radiologists confirmed that the enhanced images enabled more sensitive detection of subtle vertebral fractures and degenerative changes. The unsupervised nature of XVertNet facilitates immediate clinical deployment without requiring additional training overhead. This innovation represents a transformative advancement in emergency radiology, providing a scalable and time-efficient solution to enhance diagnostic accuracy in high-pressure clinical environments.

[1106] Universal Vessel Segmentation for Multi-Modality Retinal Images

Bo Wen, Anna Heinke, Akshay Agnihotri, Dirk-Uwe Bartsch, William Freeman, Truong Nguyen, Cheolhong An

Main category: eess.IV

TL;DR: This paper presents UVSM, a universal vessel segmentation model for multi-modality retinal images that addresses limitations of single-modality approaches and eliminates the need for modality-specific finetuning.

DetailsMotivation: Existing retinal vessel segmentation studies are limited to single modalities (mainly Color Fundus) and require separate finetuned models for new modalities, which demands additional training data that is difficult to acquire.

Method: The authors propose a universal vessel segmentation model (UVSM) that can segment vessels across multiple retinal imaging modalities without requiring modality-specific finetuning or extra training data.

Result: The universal model demonstrates comparable performance with state-of-the-art finetuned methods while being significantly more versatile across a wider range of retinal imaging modalities.

Conclusion: This is the first work to achieve modality-agnostic retinal vessel segmentation and the first to study vessel segmentation in several novel retinal imaging modalities, representing a significant advancement in multi-modality retinal image analysis.

Abstract: We identify two major limitations in the existing studies on retinal vessel segmentation: (1) Most existing works are restricted to one modality, i.e, the Color Fundus (CF). However, multi-modality retinal images are used every day in the study of retina and diagnosis of retinal diseases, and the study of vessel segmentation on the other modalities is scarce; (2) Even though a few works extended their experiments to limited new modalities such as the Multi-Color Scanning Laser Ophthalmoscopy (MC), these works still require finetuning a separate model for the new modality. The finetuning will require extra training data, which is difficult to acquire. In this work, we present a novel universal vessel segmentation model (UVSM) for multi-modality retinal images. Not only do we perform the study on a much wider range of modalities, but we also propose a universal model to segment the vessels in all these commonly-used modalities. Despite being much more versatile comparing with existing methods, our universal model still demonstrates comparable performance with the state-of-the-art finetuned methods. To the best of our knowledge, this is the first work that achieves modality-agnostic retinal vessel segmentation and also the first work that studies retinal vessel segmentation in some novel modalities.

[1107] Bridging Synthetic-to-Real Gaps: Frequency-Aware Perturbation and Selection for Single-shot Multi-Parametric Mapping Reconstruction

Linyu Fan, Che Wang, Ming Ye, Qizhi Yang, Zejun Wu, Xinghao Ding, Yue Huang, Jianfeng Bao, Shuhui Cai, Congbo Cai

Main category: eess.IV

TL;DR: FPS method combines frequency-aware perturbation and selection to address domain gaps in medical image reconstruction, showing superior performance across multiple clinical applications.

DetailsMotivation: Address synthetic-to-real gaps in medical imaging reconstruction, particularly for ultra-fast multi-parametric methods like MOLED that suffer from domain gap issues, structural integrity problems, and mapping accuracy challenges.

Method: Proposed frequency-aware perturbation and selection (FPS) with Wasserstein distance-modulated frequency-aware perturbation (WDFP) and hierarchical frequency-aware selection network (HFSNet) that includes frequency-aware adaptive selection (FAS), compact FAS (cFAS) and feature-aware architecture integration (FAI).

Result: Extensive experiments on synthetic data and real clinical cases (5 healthy volunteers, 94 ischemic stroke patients, 46 meningioma patients) demonstrate superiority and clinical applicability. Successfully applied to diffusion tensor imaging (DTI).

Conclusion: FPS establishes a robust closed-loop learning pathway for domain adaptation in medical image reconstruction, showing versatility and potential for broader medical applications beyond the tested scenarios.

Abstract: Data-centric artificial intelligence (AI) has remarkably advanced medical imaging, with emerging methods using synthetic data to address data scarcity while introducing synthetic-to-real gaps. Unsupervised domain adaptation (UDA) shows promise in ground truth-scarce tasks, but its application in reconstruction remains underexplored. Although multiple overlapping-echo detachment (MOLED) achieves ultra-fast multi-parametric reconstruction, extending its application to various clinical scenarios, the quality suffers from deficiency in mitigating the domain gap, difficulty in maintaining structural integrity, and inadequacy in ensuring mapping accuracy. To resolve these issues, we proposed frequency-aware perturbation and selection (FPS), comprising Wasserstein distance-modulated frequency-aware perturbation (WDFP) and hierarchical frequency-aware selection network (HFSNet), which integrates frequency-aware adaptive selection (FAS), compact FAS (cFAS) and feature-aware architecture integration (FAI). Specifically, perturbation activates domain-invariant feature learning within uncertainty, while selection refines optimal solutions within perturbation, establishing a robust and closed-loop learning pathway. Extensive experiments on synthetic data, along with diverse real clinical cases from 5 healthy volunteers, 94 ischemic stroke patients, and 46 meningioma patients, demonstrate the superiority and clinical applicability of FPS. Furthermore, FPS is applied to diffusion tensor imaging (DTI), underscoring its versatility and potential for broader medical applications. The code is available at https://github.com/flyannie/FPS.

[1108] DeepNuParc: A Novel Deep Clustering Framework for Fine-scale Parcellation of Brain Nuclei Using Diffusion MRI Tractography

Haolin He, Ce Zhu, Le Zhang, Yipeng Liu, Xiao Xu, Yuqian Chen, Leo Zekelman, Jarrett Rushmore, Yogesh Rathi, Nikos Makris, Lauren J. O’Donnell, Fan Zhang

Main category: eess.IV

TL;DR: DeepNuParc is a deep learning pipeline for automated fine-scale parcellation of brain nuclei using diffusion MRI tractography, achieving consistent multi-subject results that align with established atlases.

DetailsMotivation: Brain nuclei serve as critical hubs in neural circuits, and fine-scale parcellation is essential for understanding anatomico-functional correlations. Current methods lack the precision needed for detailed subdivision analysis.

Method: Combines deep learning for accurate nuclei segmentation on dMRI data, novel streamline clustering-based structural connectivity features, and improved joint dimensionality reduction with k-means clustering for finer-scale parcellation.

Result: DeepNuParc successfully parcellated amygdala and thalamus into multiple consistent parcels across subjects, showing good correspondence with widely used coarse-scale atlases.

Conclusion: The proposed pipeline enables automated, fine-scale brain nuclei parcellation with cross-subject consistency, providing a valuable tool for detailed neuroanatomical studies.

Abstract: Brain nuclei are clusters of anatomically distinct neurons that serve as important hubs for processing and relaying information in various neural circuits. Fine-scale parcellation of the brain nuclei is vital for a comprehensive understanding of its anatomico-functional correlations. Diffusion MRI tractography is an advanced imaging technique that can estimate the brain’s white matter structural connectivity to potentially reveal the topography of the nuclei of interest for studying its subdivisions. In this work, we present a deep clustering pipeline, namely DeepNuParc, to perform automated, fine-scale parcellation of brain nuclei using diffusion MRI tractography. First, we incorporate a newly proposed deep learning approach to enable accurate segmentation of the nuclei of interest directly on the dMRI data. Next, we design a novel streamline clustering-based structural connectivity feature for a robust representation of voxels within the nuclei. Finally, we improve the popular joint dimensionality reduction and k-means clustering approach to enable nuclei parcellation at a finer scale. We demonstrate DeepNuParc on two important brain structures, i.e. the amygdala and the thalamus, that are known to have multiple anatomically and functionally distinct nuclei subdivisions. Experimental results show that DeepNuParc enables consistent parcellation of the nuclei into multiple parcels across multiple subjects and achieves good correspondence with the widely used coarse-scale atlases. Our codes are available at https://github.com/HarlandZZC/deep_nuclei_parcellation.

[1109] SALT: Parameter-Efficient Fine-Tuning via Singular Value Adaptation with Low-Rank Transformation

Abdelrahman Elsayed, Sarim Hashmi, Mohammed Elseiagy, Hu Wang, Mohammad Yaqub, Ibrahim Almakky

Main category: eess.IV

TL;DR: SALT is a hybrid parameter-efficient fine-tuning method that combines singular value adaptation with low-rank transformation for medical image segmentation, outperforming existing PEFT methods with minimal trainable parameters.

DetailsMotivation: Medical image segmentation requires domain-specific features, but fine-tuning large foundation models is costly. Existing PEFT methods like LoRA may underfit with insufficient rank, while full-rank SVD methods lack flexibility and have inconsistent performance.

Method: SALT selectively adapts the most influential singular values using trainable scale and shift parameters, complemented by a low-rank update for the remaining subspace. This hybrid approach combines advantages of both LoRA and SVD without increasing model size or depth.

Result: Evaluated on 5 challenging medical datasets (20-1000 samples), SALT outperforms state-of-the-art PEFT methods (LoRA and SVD) by 2% to 5% in Dice score with only 3.9% trainable parameters, demonstrating robust adaptation in low-resource settings.

Conclusion: SALT provides an effective parameter-efficient fine-tuning solution for medical image segmentation that balances comprehensive updates with flexibility, achieving superior performance even with limited data and minimal parameter overhead.

Abstract: The complex nature of medical image segmentation calls for models that are specifically designed to capture detailed, domain-specific features. Large foundation models offer considerable flexibility, yet the cost of fine-tuning these models remains a significant barrier. Parameter-Efficient Fine-Tuning (PEFT) methods, such as Low-Rank Adaptation (LoRA), efficiently update model weights with low-rank matrices but may suffer from underfitting when the chosen rank is insufficient to capture domain-specific nuances. Conversely, full-rank Singular Value Decomposition (SVD) based methods provide comprehensive updates by modifying all singular values, yet they often lack flexibility and exhibit variable performance across datasets. We propose SALT (Singular Value Adaptation with Low-Rank Transformation), a method that selectively adapts the most influential singular values using trainable scale and shift parameters while complementing this with a low-rank update for the remaining subspace. This hybrid approach harnesses the advantages of both LoRA and SVD, enabling effective adaptation without relying on increasing model size or depth. Evaluated on 5 challenging medical datasets, ranging from as few as 20 samples to 1000, SALT outperforms state-of-the-art PEFT (LoRA and SVD) by 2% to 5% in Dice with only 3.9% trainable parameters, demonstrating robust adaptation even in low-resource settings. The code for SALT is available at: https://github.com/BioMedIA-MBZUAI/SALT

[1110] Optimizing Breast Cancer Detection in Mammograms: A Comprehensive Study of Transfer Learning, Resolution Reduction, and Multi-View Classification

Daniel G. P. Petrini, Hae Yong Kim

Main category: eess.IV

TL;DR: Systematic comparison of single-view vs multi-view mammogram classification techniques, showing multi-view approaches outperform single-view methods and achieving state-of-the-art performance in both scenarios.

DetailsMotivation: Mammography accuracy depends on expert radiologists, and while AI methods have evolved from patch-based to multi-view systems, it remains unclear whether multi-view consistently outperforms single-view approaches.

Method: Systematic evaluation and comparison of single-view and multi-view mammogram classification techniques, introducing new models with superior performance and exploring optimal architectures and transfer learning strategies.

Result: Achieved state-of-the-art performance in both single-view and two-view classification scenarios, with multi-view systems demonstrating consistent superiority over single-view approaches.

Conclusion: Multi-view mammogram classification techniques provide enhanced accuracy over single-view methods, with valuable insights for optimal model architectures and transfer learning strategies to improve mammogram interpretation efficiency and accuracy.

Abstract: Mammography, an X-ray-based imaging technique, plays a crucial role in the early detection of breast cancer. Its accuracy heavily depends on expert radiologists, making it essential to minimize interpretation errors. To support radiologists, various computer-aided detection and diagnostic methods have been proposed, increasingly leveraging advancements in artificial intelligence and machine learning. Over recent years, mammogram analysis has evolved significantly - from early patch-based classifiers, which examine only localized regions of images, to full-image classifiers, and later towards multi-view systems that simultaneously integrate multiple perspectives of the mammographic exam for enhanced accuracy. Despite this progression, critical questions remain, such as whether multi-view systems consistently outperform single-view approaches. In this paper, we systematically evaluate and compare the effectiveness of single-view and multi-view mammogram classification techniques. Our research introduces models that achieve superior performance relative to existing state-of-the-art approaches in both single-view and two-view classification scenarios. Furthermore, our findings provide valuable insights into optimal model architectures and effective transfer learning strategies, paving the way for more accurate and efficient mammogram interpretation. The inference code and model are available at https://github.com/dpetrini/multiple-view.

[1111] Towards Interpretable Counterfactual Generation via Multimodal Autoregression

Chenglong Ma, Yuanfeng Ji, Jin Ye, Lu Zhang, Ying Chen, Tianbin Li, Mingjie Li, Junjun He, Hongming Shan

Main category: eess.IV

TL;DR: ICG introduces interpretable counterfactual medical image generation that produces both progression-aligned images and textual explanations of visual changes, addressing the critical need for traceable reasoning in clinical applications.

DetailsMotivation: Existing counterfactual medical image generation methods produce silent predictions without interpretation, creating a critical gap for medical applications that require verifiable and traceable reasoning to support clinical decision-making.

Method: Proposed ICG-CXR dataset with longitudinal medical images paired with progression prompts and textual interpretations, and developed ProgEmu - an autoregressive model that jointly generates counterfactual images and interpretation texts.

Result: ProgEmu demonstrates superiority in generating progression-aligned counterfactuals and interpretations, showing significant potential for enhancing clinical decision support and medical education.

Conclusion: The ICG framework successfully addresses the interpretability gap in medical counterfactual generation, providing both visual predictions and textual explanations that enable verifiable clinical reasoning and hypothesis testing.

Abstract: Counterfactual medical image generation enables clinicians to explore clinical hypotheses, such as predicting disease progression, facilitating their decision-making. While existing methods can generate visually plausible images from disease progression prompts, they produce silent predictions that lack interpretation to verify how the generation reflects the hypothesized progression – a critical gap for medical applications that require traceable reasoning. In this paper, we propose Interpretable Counterfactual Generation (ICG), a novel task requiring the joint generation of counterfactual images that reflect the clinical hypothesis and interpretation texts that outline the visual changes induced by the hypothesis. To enable ICG, we present ICG-CXR, the first dataset pairing longitudinal medical images with hypothetical progression prompts and textual interpretations. We further introduce ProgEmu, an autoregressive model that unifies the generation of counterfactual images and textual interpretations. We demonstrate the superiority of ProgEmu in generating progression-aligned counterfactuals and interpretations, showing significant potential in enhancing clinical decision support and medical education. Project page: https://progemu.github.io.

[1112] Fine-grained Image Quality Assessment for Perceptual Image Restoration

Xiangfei Sheng, Xiaofeng Pan, Zhichao Yang, Pengfei Chen, Leida Li

Main category: eess.IV

TL;DR: New fine-grained image quality assessment dataset FGRestore and model FGResQ for image restoration tasks, addressing limitations of existing IQA metrics in distinguishing subtle quality differences.

DetailsMotivation: Existing IQA metrics are inadequate for evaluating image restoration tasks, particularly in distinguishing fine-grained quality differences among restored images, creating a need for more accurate assessment methods.

Method: Created FGRestore dataset with 18,408 restored images across 6 IR tasks and 30,886 pairwise preferences. Proposed FGResQ model combining coarse-grained score regression and fine-grained quality ranking.

Result: FGResQ significantly outperforms state-of-the-art IQA metrics in extensive experiments and comparisons, demonstrating better alignment with fine-grained restoration quality.

Conclusion: The proposed FGResQ model effectively addresses the limitations of existing IQA metrics for image restoration tasks, providing both accurate quality scoring and fine-grained ranking capabilities.

Abstract: Recent years have witnessed remarkable achievements in perceptual image restoration (IR), creating an urgent demand for accurate image quality assessment (IQA), which is essential for both performance comparison and algorithm optimization. Unfortunately, the existing IQA metrics exhibit inherent weakness for IR task, particularly when distinguishing fine-grained quality differences among restored images. To address this dilemma, we contribute the first-of-its-kind fine-grained image quality assessment dataset for image restoration, termed FGRestore, comprising 18,408 restored images across six common IR tasks. Beyond conventional scalar quality scores, FGRestore was also annotated with 30,886 fine-grained pairwise preferences. Based on FGRestore, a comprehensive benchmark was conducted on the existing IQA metrics, which reveal significant inconsistencies between score-based IQA evaluations and the fine-grained restoration quality. Motivated by these findings, we further propose FGResQ, a new IQA model specifically designed for image restoration, which features both coarse-grained score regression and fine-grained quality ranking. Extensive experiments and comparisons demonstrate that FGResQ significantly outperforms state-of-the-art IQA metrics. Codes and model weights have been released in https://pxf0429.github.io/FGResQ/

[1113] FreqSelect: Frequency-Aware fMRI-to-Image Reconstruction

Junliang Ye, Lei Wang, Md Zakir Hossain

Main category: eess.IV

TL;DR: FreqSelect is a lightweight adaptive module that selectively filters spatial-frequency bands before encoding in fMRI image reconstruction, improving both reconstruction quality and providing interpretable insights into brain’s visual frequency representation.

DetailsMotivation: Current two-stage models for fMRI image reconstruction treat all spatial-frequency components equally, forcing simultaneous feature extraction and noise suppression, which limits effectiveness. There's a need for selective frequency filtering to improve decoding accuracy.

Method: Introduces FreqSelect module that dynamically emphasizes predictive frequencies and suppresses uninformative ones before encoding. It integrates into standard VAE-diffusion pipelines without additional supervision, acting as a content-aware gate between image features and neural data.

Result: Consistently improves reconstruction quality across both low- and high-level metrics on Natural Scenes dataset. Provides interpretable insights into how different visual frequencies are represented in the brain. Generalizes across subjects and scenes.

Conclusion: FreqSelect offers a principled approach to enhance both decoding accuracy and neuroscientific interpretability in fMRI image reconstruction, with promise for extension to other neuroimaging modalities.

Abstract: Reconstructing natural images from functional magnetic resonance imaging (fMRI) data remains a core challenge in natural decoding due to the mismatch between the richness of visual stimuli and the noisy, low resolution nature of fMRI signals. While recent two-stage models, combining deep variational autoencoders (VAEs) with diffusion models, have advanced this task, they treat all spatial-frequency components of the input equally. This uniform treatment forces the model to extract meaning features and suppress irrelevant noise simultaneously, limiting its effectiveness. We introduce FreqSelect, a lightweight, adaptive module that selectively filters spatial-frequency bands before encoding. By dynamically emphasizing frequencies that are most predictive of brain activity and suppressing those that are uninformative, FreqSelect acts as a content-aware gate between image features and natural data. It integrates seamlessly into standard very deep VAE-diffusion pipelines and requires no additional supervision. Evaluated on the Natural Scenes dataset, FreqSelect consistently improves reconstruction quality across both low- and high-level metrics. Beyond performance gains, the learned frequency-selection patterns offer interpretable insights into how different visual frequencies are represented in the brain. Our method generalizes across subjects and scenes, and holds promise for extension to other neuroimaging modalities, offering a principled approach to enhancing both decoding accuracy and neuroscientific interpretability.

[1114] A versatile foundation model for cine cardiac magnetic resonance image analysis tasks

Yunguan Fu, Wenjia Bai, Weixi Yi, Charlotte Manisty, Anish N Bhuva, Thomas A Treibel, James C Moon, Matthew J Clarkson, Rhodri Huw Davies, Yipeng Hu

Main category: eess.IV

TL;DR: CineMA is a multi-view convolution-transformer masked autoencoder trained on 15M cine images that outperforms CNNs in cardiac image analysis tasks including segmentation, disease detection, and mortality prediction while maintaining fairness across demographics.

DetailsMotivation: To develop a versatile foundation model for automated cardiac image analysis that can perform multiple clinically-relevant tasks with high accuracy and efficiency to support clinical workflow and cardiovascular research.

Method: A multi-view convolution-transformer masked autoencoder (CineMA) trained on 15 million cine images from 74,916 subjects, validated on >4,500 images from eight independent datasets with diverse population characteristics.

Result: CineMA consistently outperformed conventional CNNs in ventricular boundary delineation and ejection fraction estimation, maintained performance with half the fine-tuning data, surpassed CNNs in disease detection, matched CNNs in long-axis function measurement, detected cardiac changes in systemic diseases, predicted mortality, and demonstrated consistent performance across demographic subgroups.

Conclusion: CineMA shows superior accuracy, learning efficiency, adaptability, and fairness as a foundation model for automated cardiac image analysis, with all code and models made publicly available to support clinical and research applications.

Abstract: Here we present a versatile foundation model that can perform a range of clinically-relevant image analysis tasks, including segmentation, landmark localisation, diagnosis, and prognostication. A multi-view convolution-transformer masked autoencoder, named as CineMA, was trained on 15 million cine images from 74,916 subjects. The model was validated on multiple image analysis tasks and compared to existing models on >4,500 images from eight independent datasets with diverse population characteristics, representing the largest benchmark study for cine CMR so far. CineMA consistently outperformed conventional convolutional neural networks (CNNs) in delineating ventricular boundaries and estimating ejection fraction, a key measure of cardiac function. The improved performance was preserved, even when the model only used half of fine-tuning data. CineMA also surpassed CNNs in disease detection and matched their performance in long-axis function measurement. Interestingly, we found that CineMA can also detect cardiac changes in systemic diseases, such as diabetes, hypertension and cancer, and can also predict mortality. Finally, we assessed model fairness and demonstrated consistent model performance across demographic subgroups. These findings highlight CineMA’s accuracy, learning efficiency, adaptability, and fairness, underscoring its potential as a foundation model for automated cardiac image analysis to support clinical workflow and cardiovascular research. All training and inference code and models are made publicly available at https://github.com/mathpluscode/CineMA.

[1115] Prompt-based Multimodal Semantic Communication for Multi-spectral Image Segmentation

Haoshuo Zhang, Yufei Bo, Hongwei Zhang, Meixia Tao

Main category: eess.IV

TL;DR: ProMSC-MIS is a prompt-based multimodal semantic communication system for multi-spectral image segmentation that uses cross-modal prompting and efficient fusion to achieve superior performance with low complexity.

DetailsMotivation: Multimodal semantic communication can enhance downstream task performance, but effectively fusing features from different modalities and extracting diverse semantic representations remains challenging.

Method: Proposes a pre-training algorithm where features from one modality serve as prompts for another, guiding unimodal encoders to learn complementary representations. Uses a semantic fusion module combining cross-attention mechanisms and squeeze-and-excitation networks.

Result: Significantly outperforms benchmark methods across various channel-source compression levels while maintaining low computational complexity and storage overhead.

Conclusion: The system shows great potential for applications like autonomous driving and nighttime surveillance, demonstrating effective multimodal feature fusion with practical efficiency.

Abstract: Multimodal semantic communication has gained widespread attention due to its ability to enhance downstream task performance. A key challenge in such systems is the effective fusion of features from different modalities, which requires the extraction of rich and diverse semantic representations from each modality. To this end, we propose ProMSC-MIS, a Prompt-based Multimodal Semantic Communication system for Multi-spectral Image Segmentation. Specifically, we propose a pre-training algorithm where features from one modality serve as prompts for another, guiding unimodal semantic encoders to learn diverse and complementary semantic representations. We further introduce a semantic fusion module that combines cross-attention mechanisms and squeeze-and-excitation (SE) networks to effectively fuse cross-modal features. Simulation results show that ProMSC-MIS significantly outperforms benchmark methods across various channel-source compression levels, while maintaining low computational complexity and storage overhead. Our scheme has great potential for applications such as autonomous driving and nighttime surveillance.

[1116] Learning local and global prototypes with optimal transport for unsupervised anomaly detection and localization

Robin Trombetta, Carole Lartizien

Main category: eess.IV

TL;DR: Novel unsupervised anomaly detection method using prototype learning with optimal transport and a balanced feature-spatial metric, achieving competitive performance on industrial benchmarks.

DetailsMotivation: Address the need for unsupervised anomaly detection in applications like industrial inspection and medical imaging where labeling is costly and bias avoidance is important.

Method: Leverages prototype learning with optimal transport from pre-trained image encoder embeddings, using a novel metric that balances feature-based and spatial-based costs to learn local and global prototypes with structural constraints.

Result: The approach enforces structural constraints that capture the underlying organization of normal samples, improving anomaly detection. Achieves performance on par with strong baselines on two industrial image anomaly detection benchmarks.

Conclusion: The proposed prototype learning method with optimal transport and balanced metric effectively detects anomalies by learning structured representations of normal data, demonstrating competitive performance in industrial applications.

Abstract: Unsupervised anomaly detection aims to detect defective parts of a sample by having access, during training, to a set of normal, i.e. defect-free, data. It has many applications in fields, such as industrial inspection or medical imaging, where acquiring labels is costly or when we want to avoid introducing biases in the type of anomalies that can be spotted. In this work, we propose a novel UAD method based on prototype learning and introduce a metric to compare a structured set of embeddings that balances a feature-based cost and a spatial-based cost. We leverage this metric to learn local and global prototypes with optimal transport from latent representations extracted with a pre-trained image encoder. We demonstrate that our approach can enforce a structural constraint when learning the prototypes, allowing to capture the underlying organization of the normal samples, thus improving the detection of incoherencies in images. Our model achieves performance that is on par with strong baselines on two reference benchmarks for anomaly detection on industrial images.

[1117] Efficient and Privacy-Protecting Background Removal for 2D Video Streaming using iPhone 15 Pro Max LiDAR

Jessica Kinnevan, Naifa Alqahtani, Toral Chauhan

Main category: eess.IV

TL;DR: Using iPhone 15 Pro Max LiDAR for real-time background removal at 60fps, overcoming lighting limitations of traditional methods.

DetailsMotivation: Traditional background removal techniques like chroma keying and AI models are dependent on lighting conditions and perform poorly in low-light environments. LiDAR's depth-based approach provides lighting-independent background removal.

Method: Integrated iPhone 15 Pro Max LiDAR and color cameras with GPU-based image processing using SwiftUI, Swift frameworks, and Metal Shader Language for real-time enhancement at 60fps.

Result: Successfully achieved real-time background removal at 60fps, though limited by current depth map resolution of 320x240 due to streaming bandwidth constraints and some material reflection limitations.

Conclusion: LiDAR technology shows strong potential as a superior background removal method for mobile applications, with the main constraint being resolution limitations that could be overcome with future hardware improvements.

Abstract: Light Detection and Ranging (LiDAR) technology in consumer-grade mobile devices can be used as a replacement for traditional background removal and compositing techniques. Unlike approaches such as chroma keying and trained AI models, LiDAR’s depth information is independent of subject lighting, and performs equally well in low-light and well-lit environments. We integrate the LiDAR and color cameras on the iPhone 15 Pro Max with GPU-based image processing. We use Apple’s SwiftUI and Swift frameworks for user interface and backend development, and Metal Shader Language (MSL) for realtime image enhancement at the standard iPhone streaming frame rate of 60 frames per second. The only meaningful limitations of the technology are the streaming bandwidth of the depth data, which currently reduces the depth map resolution to 320x240, and any pre-existing limitations of the LiDAR IR laser to reflect accurate depth from some materials. If the LiDAR resolution on a mobile device like the iPhone can be improved to match the color image resolution, LiDAR could feasibly become the preeminent method of background removal for video applications and photography.

[1118] Mitosis detection in domain shift scenarios: a Mamba-based approach

Gennaro Percannella, Mattia Sarno, Francesco Tortorella, Mario Vento

Main category: eess.IV

TL;DR: A Mamba-based approach using VM-UNet architecture with stain augmentation for mitosis detection under domain shift, submitted to MIDOG challenge with preliminary results showing room for improvement.

DetailsMotivation: Mitosis detection is crucial for tumor assessment but current ML algorithms suffer significant performance drops when tested on images from different domains than training data.

Method: Proposes a Mamba-based approach using VM-UNet architecture combined with stain augmentation operations to improve model robustness against domain shift.

Result: Preliminary experiments on MIDOG++ dataset show large room for improvement for the proposed method.

Conclusion: The Mamba-based approach with stain augmentation shows potential for mitosis detection under domain shift but requires further development and optimization.

Abstract: Mitosis detection in histopathology images plays a key role in tumor assessment. Although machine learning algorithms could be exploited for aiding physicians in accurately performing such a task, these algorithms suffer from significative performance drop when evaluated on images coming from domains that are different from the training ones. In this work, we propose a Mamba-based approach for mitosis detection under domain shift, inspired by the promising performance demonstrated by Mamba in medical imaging segmentation tasks. Specifically, our approach exploits a VM-UNet architecture for carrying out the addressed task, as well as stain augmentation operations for further improving model robustness against domain shift. Our approach has been submitted to the track 1 of the MItosis DOmain Generalization (MIDOG) challenge. Preliminary experiments, conducted on the MIDOG++ dataset, show large room for improvement for the proposed method.

[1119] A multi-task neural network for atypical mitosis recognition under domain shift

Gennaro Percannella, Mattia Sarno, Francesco Tortorella, Mario Vento

Main category: eess.IV

TL;DR: Multi-task learning approach for domain generalization in atypical mitosis detection, using auxiliary tasks to help models focus on classification objects while ignoring domain-varying backgrounds.

DetailsMotivation: Machine learning models for atypical mitotic figure recognition suffer significant performance drops under domain shift, making domain generalization crucial for accurate tumor aggressiveness assessment.

Method: Multi-task learning approach that exploits auxiliary tasks correlated to the main classification task to help the model focus only on the classification object while ignoring domain-varying background features.

Result: Promising performance in preliminary evaluation across three distinct datasets: MIDOG 2025 Atypical Training Set, Ami-Br dataset, and preliminary test set of MIDOG25 challenge.

Conclusion: The multi-task learning approach shows potential for improving domain generalization in atypical mitosis detection by leveraging auxiliary tasks to enhance model focus on relevant features.

Abstract: Recognizing atypical mitotic figures in histopathology images allows physicians to correctly assess tumor aggressiveness. Although machine learning models could be exploited for automatically performing such a task, under domain shift these models suffer from significative performance drops. In this work, an approach based on multi-task learning is proposed for addressing this problem. By exploiting auxiliary tasks, correlated to the main classification task, the proposed approach, submitted to the track 2 of the MItosis DOmain Generalization (MIDOG) challenge, aims to aid the model to focus only on the object to classify, ignoring the domain varying background of the image. The proposed approach shows promising performance in a preliminary evaluation conducted on three distinct datasets, i.e., the MIDOG 2025 Atypical Training Set, the Ami-Br dataset, as well as the preliminary test set of the MIDOG25 challenge.

Last updated: 2025-09-15
Built with Hugo, theme modified on Stack